| { | |
| "best_metric": null, | |
| "best_model_checkpoint": null, | |
| "epoch": 5.0, | |
| "eval_steps": 500, | |
| "global_step": 5555, | |
| "is_hyper_param_search": false, | |
| "is_local_process_zero": true, | |
| "is_world_process_zero": true, | |
| "log_history": [ | |
| { | |
| "epoch": 0.0288028802880288, | |
| "grad_norm": 0.9036304950714111, | |
| "learning_rate": 5.755395683453238e-07, | |
| "loss": 1.0187, | |
| "step": 32 | |
| }, | |
| { | |
| "epoch": 0.0576057605760576, | |
| "grad_norm": 2.02474045753479, | |
| "learning_rate": 1.1510791366906476e-06, | |
| "loss": 1.0342, | |
| "step": 64 | |
| }, | |
| { | |
| "epoch": 0.08640864086408641, | |
| "grad_norm": 1.7671840190887451, | |
| "learning_rate": 1.7266187050359715e-06, | |
| "loss": 1.0534, | |
| "step": 96 | |
| }, | |
| { | |
| "epoch": 0.1152115211521152, | |
| "grad_norm": 2.703395366668701, | |
| "learning_rate": 2.302158273381295e-06, | |
| "loss": 1.0047, | |
| "step": 128 | |
| }, | |
| { | |
| "epoch": 0.14401440144014402, | |
| "grad_norm": 2.96386456489563, | |
| "learning_rate": 2.877697841726619e-06, | |
| "loss": 1.0125, | |
| "step": 160 | |
| }, | |
| { | |
| "epoch": 0.17281728172817282, | |
| "grad_norm": 0.9527461528778076, | |
| "learning_rate": 3.453237410071943e-06, | |
| "loss": 0.9816, | |
| "step": 192 | |
| }, | |
| { | |
| "epoch": 0.20162016201620162, | |
| "grad_norm": 2.197382926940918, | |
| "learning_rate": 4.028776978417267e-06, | |
| "loss": 1.05, | |
| "step": 224 | |
| }, | |
| { | |
| "epoch": 0.2304230423042304, | |
| "grad_norm": 1.8034260272979736, | |
| "learning_rate": 4.60431654676259e-06, | |
| "loss": 0.8815, | |
| "step": 256 | |
| }, | |
| { | |
| "epoch": 0.25922592259225924, | |
| "grad_norm": 2.3050363063812256, | |
| "learning_rate": 5.179856115107914e-06, | |
| "loss": 0.9027, | |
| "step": 288 | |
| }, | |
| { | |
| "epoch": 0.28802880288028804, | |
| "grad_norm": 2.022681951522827, | |
| "learning_rate": 5.755395683453238e-06, | |
| "loss": 0.865, | |
| "step": 320 | |
| }, | |
| { | |
| "epoch": 0.31683168316831684, | |
| "grad_norm": 1.3653255701065063, | |
| "learning_rate": 6.330935251798561e-06, | |
| "loss": 0.8724, | |
| "step": 352 | |
| }, | |
| { | |
| "epoch": 0.34563456345634563, | |
| "grad_norm": 1.3452287912368774, | |
| "learning_rate": 6.906474820143886e-06, | |
| "loss": 0.8093, | |
| "step": 384 | |
| }, | |
| { | |
| "epoch": 0.37443744374437443, | |
| "grad_norm": 1.9356400966644287, | |
| "learning_rate": 7.48201438848921e-06, | |
| "loss": 0.7579, | |
| "step": 416 | |
| }, | |
| { | |
| "epoch": 0.40324032403240323, | |
| "grad_norm": 1.4237608909606934, | |
| "learning_rate": 8.057553956834533e-06, | |
| "loss": 0.7364, | |
| "step": 448 | |
| }, | |
| { | |
| "epoch": 0.43204320432043203, | |
| "grad_norm": 1.9049218893051147, | |
| "learning_rate": 8.633093525179856e-06, | |
| "loss": 0.7033, | |
| "step": 480 | |
| }, | |
| { | |
| "epoch": 0.4608460846084608, | |
| "grad_norm": 3.9214580059051514, | |
| "learning_rate": 9.20863309352518e-06, | |
| "loss": 0.696, | |
| "step": 512 | |
| }, | |
| { | |
| "epoch": 0.4896489648964896, | |
| "grad_norm": 1.9122344255447388, | |
| "learning_rate": 9.784172661870505e-06, | |
| "loss": 0.697, | |
| "step": 544 | |
| }, | |
| { | |
| "epoch": 0.5184518451845185, | |
| "grad_norm": 4.641770362854004, | |
| "learning_rate": 9.999605063062183e-06, | |
| "loss": 0.6691, | |
| "step": 576 | |
| }, | |
| { | |
| "epoch": 0.5472547254725473, | |
| "grad_norm": 1.5034338235855103, | |
| "learning_rate": 9.997330428740717e-06, | |
| "loss": 0.6846, | |
| "step": 608 | |
| }, | |
| { | |
| "epoch": 0.5760576057605761, | |
| "grad_norm": 1.306266188621521, | |
| "learning_rate": 9.993034838418394e-06, | |
| "loss": 0.6663, | |
| "step": 640 | |
| }, | |
| { | |
| "epoch": 0.6048604860486049, | |
| "grad_norm": 4.800021648406982, | |
| "learning_rate": 9.986720029262515e-06, | |
| "loss": 0.6264, | |
| "step": 672 | |
| }, | |
| { | |
| "epoch": 0.6336633663366337, | |
| "grad_norm": 1.366411566734314, | |
| "learning_rate": 9.97838855502686e-06, | |
| "loss": 0.6116, | |
| "step": 704 | |
| }, | |
| { | |
| "epoch": 0.6624662466246625, | |
| "grad_norm": 1.8418484926223755, | |
| "learning_rate": 9.968043785018917e-06, | |
| "loss": 0.6244, | |
| "step": 736 | |
| }, | |
| { | |
| "epoch": 0.6912691269126913, | |
| "grad_norm": 7.662785053253174, | |
| "learning_rate": 9.955689902737327e-06, | |
| "loss": 0.5972, | |
| "step": 768 | |
| }, | |
| { | |
| "epoch": 0.7200720072007201, | |
| "grad_norm": 1.7773517370224, | |
| "learning_rate": 9.941331904180025e-06, | |
| "loss": 0.6061, | |
| "step": 800 | |
| }, | |
| { | |
| "epoch": 0.7488748874887489, | |
| "grad_norm": 1.4460841417312622, | |
| "learning_rate": 9.924975595823843e-06, | |
| "loss": 0.5821, | |
| "step": 832 | |
| }, | |
| { | |
| "epoch": 0.7776777677767777, | |
| "grad_norm": 2.224186897277832, | |
| "learning_rate": 9.906627592276318e-06, | |
| "loss": 0.6255, | |
| "step": 864 | |
| }, | |
| { | |
| "epoch": 0.8064806480648065, | |
| "grad_norm": 1.6951428651809692, | |
| "learning_rate": 9.886295313600703e-06, | |
| "loss": 0.6272, | |
| "step": 896 | |
| }, | |
| { | |
| "epoch": 0.8352835283528353, | |
| "grad_norm": 1.9330335855484009, | |
| "learning_rate": 9.86398698231524e-06, | |
| "loss": 0.5994, | |
| "step": 928 | |
| }, | |
| { | |
| "epoch": 0.8640864086408641, | |
| "grad_norm": 1.4903757572174072, | |
| "learning_rate": 9.839711620067906e-06, | |
| "loss": 0.5589, | |
| "step": 960 | |
| }, | |
| { | |
| "epoch": 0.8928892889288929, | |
| "grad_norm": 1.2906200885772705, | |
| "learning_rate": 9.813479043988013e-06, | |
| "loss": 0.5859, | |
| "step": 992 | |
| }, | |
| { | |
| "epoch": 0.9216921692169217, | |
| "grad_norm": 1.8480700254440308, | |
| "learning_rate": 9.785299862716072e-06, | |
| "loss": 0.6045, | |
| "step": 1024 | |
| }, | |
| { | |
| "epoch": 0.9504950495049505, | |
| "grad_norm": 1.5740501880645752, | |
| "learning_rate": 9.755185472113595e-06, | |
| "loss": 0.5862, | |
| "step": 1056 | |
| }, | |
| { | |
| "epoch": 0.9792979297929792, | |
| "grad_norm": 2.271888017654419, | |
| "learning_rate": 9.723148050654522e-06, | |
| "loss": 0.5833, | |
| "step": 1088 | |
| }, | |
| { | |
| "epoch": 1.008100810081008, | |
| "grad_norm": 0.8760111331939697, | |
| "learning_rate": 9.689200554500162e-06, | |
| "loss": 0.5658, | |
| "step": 1120 | |
| }, | |
| { | |
| "epoch": 1.036903690369037, | |
| "grad_norm": 1.2301350831985474, | |
| "learning_rate": 9.65335671225963e-06, | |
| "loss": 0.5839, | |
| "step": 1152 | |
| }, | |
| { | |
| "epoch": 1.0657065706570656, | |
| "grad_norm": 1.7237297296524048, | |
| "learning_rate": 9.615631019437896e-06, | |
| "loss": 0.5718, | |
| "step": 1184 | |
| }, | |
| { | |
| "epoch": 1.0945094509450946, | |
| "grad_norm": 1.146071434020996, | |
| "learning_rate": 9.576038732573702e-06, | |
| "loss": 0.6036, | |
| "step": 1216 | |
| }, | |
| { | |
| "epoch": 1.1233123312331232, | |
| "grad_norm": 1.4510979652404785, | |
| "learning_rate": 9.53459586306971e-06, | |
| "loss": 0.602, | |
| "step": 1248 | |
| }, | |
| { | |
| "epoch": 1.1521152115211521, | |
| "grad_norm": 1.3618237972259521, | |
| "learning_rate": 9.491319170717375e-06, | |
| "loss": 0.5661, | |
| "step": 1280 | |
| }, | |
| { | |
| "epoch": 1.1809180918091808, | |
| "grad_norm": 1.5930922031402588, | |
| "learning_rate": 9.446226156919164e-06, | |
| "loss": 0.5892, | |
| "step": 1312 | |
| }, | |
| { | |
| "epoch": 1.2097209720972097, | |
| "grad_norm": 1.0816315412521362, | |
| "learning_rate": 9.399335057610872e-06, | |
| "loss": 0.5816, | |
| "step": 1344 | |
| }, | |
| { | |
| "epoch": 1.2385238523852384, | |
| "grad_norm": 1.2080299854278564, | |
| "learning_rate": 9.350664835886873e-06, | |
| "loss": 0.5766, | |
| "step": 1376 | |
| }, | |
| { | |
| "epoch": 1.2673267326732673, | |
| "grad_norm": 1.655219554901123, | |
| "learning_rate": 9.300235174331309e-06, | |
| "loss": 0.5925, | |
| "step": 1408 | |
| }, | |
| { | |
| "epoch": 1.296129612961296, | |
| "grad_norm": 1.02821683883667, | |
| "learning_rate": 9.248066467058315e-06, | |
| "loss": 0.5726, | |
| "step": 1440 | |
| }, | |
| { | |
| "epoch": 1.324932493249325, | |
| "grad_norm": 1.9538413286209106, | |
| "learning_rate": 9.194179811464499e-06, | |
| "loss": 0.5405, | |
| "step": 1472 | |
| }, | |
| { | |
| "epoch": 1.3537353735373538, | |
| "grad_norm": 1.4714971780776978, | |
| "learning_rate": 9.138596999696994e-06, | |
| "loss": 0.5848, | |
| "step": 1504 | |
| }, | |
| { | |
| "epoch": 1.3825382538253825, | |
| "grad_norm": 1.3316493034362793, | |
| "learning_rate": 9.081340509840568e-06, | |
| "loss": 0.5796, | |
| "step": 1536 | |
| }, | |
| { | |
| "epoch": 1.4113411341134112, | |
| "grad_norm": 1.2864930629730225, | |
| "learning_rate": 9.022433496827323e-06, | |
| "loss": 0.5695, | |
| "step": 1568 | |
| }, | |
| { | |
| "epoch": 1.4401440144014401, | |
| "grad_norm": 1.261751651763916, | |
| "learning_rate": 8.961899783072673e-06, | |
| "loss": 0.5779, | |
| "step": 1600 | |
| }, | |
| { | |
| "epoch": 1.468946894689469, | |
| "grad_norm": 1.6343884468078613, | |
| "learning_rate": 8.899763848841396e-06, | |
| "loss": 0.6653, | |
| "step": 1632 | |
| }, | |
| { | |
| "epoch": 1.4977497749774977, | |
| "grad_norm": 1.3842564821243286, | |
| "learning_rate": 8.836050822347632e-06, | |
| "loss": 0.6163, | |
| "step": 1664 | |
| }, | |
| { | |
| "epoch": 1.5265526552655264, | |
| "grad_norm": 1.3212823867797852, | |
| "learning_rate": 8.770786469592863e-06, | |
| "loss": 0.5873, | |
| "step": 1696 | |
| }, | |
| { | |
| "epoch": 1.5553555355535553, | |
| "grad_norm": 1.8528977632522583, | |
| "learning_rate": 8.703997183945949e-06, | |
| "loss": 0.5722, | |
| "step": 1728 | |
| }, | |
| { | |
| "epoch": 1.5841584158415842, | |
| "grad_norm": 2.5258665084838867, | |
| "learning_rate": 8.635709975469479e-06, | |
| "loss": 0.5556, | |
| "step": 1760 | |
| }, | |
| { | |
| "epoch": 1.612961296129613, | |
| "grad_norm": 1.2707765102386475, | |
| "learning_rate": 8.565952459996684e-06, | |
| "loss": 0.5536, | |
| "step": 1792 | |
| }, | |
| { | |
| "epoch": 1.6417641764176416, | |
| "grad_norm": 1.631494164466858, | |
| "learning_rate": 8.49475284796342e-06, | |
| "loss": 0.5651, | |
| "step": 1824 | |
| }, | |
| { | |
| "epoch": 1.6705670567056705, | |
| "grad_norm": 2.1226248741149902, | |
| "learning_rate": 8.422139932999658e-06, | |
| "loss": 0.5679, | |
| "step": 1856 | |
| }, | |
| { | |
| "epoch": 1.6993699369936994, | |
| "grad_norm": 2.1261792182922363, | |
| "learning_rate": 8.34814308028513e-06, | |
| "loss": 0.5896, | |
| "step": 1888 | |
| }, | |
| { | |
| "epoch": 1.7281728172817283, | |
| "grad_norm": 3.6276915073394775, | |
| "learning_rate": 8.272792214673851e-06, | |
| "loss": 0.5859, | |
| "step": 1920 | |
| }, | |
| { | |
| "epoch": 1.756975697569757, | |
| "grad_norm": 1.1661372184753418, | |
| "learning_rate": 8.19611780859229e-06, | |
| "loss": 0.6355, | |
| "step": 1952 | |
| }, | |
| { | |
| "epoch": 1.7857785778577857, | |
| "grad_norm": 1.1813215017318726, | |
| "learning_rate": 8.118150869716101e-06, | |
| "loss": 0.5659, | |
| "step": 1984 | |
| }, | |
| { | |
| "epoch": 1.8145814581458146, | |
| "grad_norm": 1.6434768438339233, | |
| "learning_rate": 8.038922928430408e-06, | |
| "loss": 0.5681, | |
| "step": 2016 | |
| }, | |
| { | |
| "epoch": 1.8433843384338435, | |
| "grad_norm": 2.215708017349243, | |
| "learning_rate": 7.95846602507868e-06, | |
| "loss": 0.5849, | |
| "step": 2048 | |
| }, | |
| { | |
| "epoch": 1.8721872187218722, | |
| "grad_norm": 1.4538156986236572, | |
| "learning_rate": 7.876812697005396e-06, | |
| "loss": 0.5677, | |
| "step": 2080 | |
| }, | |
| { | |
| "epoch": 1.900990099009901, | |
| "grad_norm": 1.8781392574310303, | |
| "learning_rate": 7.79399596539771e-06, | |
| "loss": 0.5612, | |
| "step": 2112 | |
| }, | |
| { | |
| "epoch": 1.9297929792979298, | |
| "grad_norm": 2.6248340606689453, | |
| "learning_rate": 7.710049321931453e-06, | |
| "loss": 0.5578, | |
| "step": 2144 | |
| }, | |
| { | |
| "epoch": 1.9585958595859587, | |
| "grad_norm": 2.893212080001831, | |
| "learning_rate": 7.625006715226844e-06, | |
| "loss": 0.5925, | |
| "step": 2176 | |
| }, | |
| { | |
| "epoch": 1.9873987398739874, | |
| "grad_norm": 1.3693437576293945, | |
| "learning_rate": 7.538902537119445e-06, | |
| "loss": 0.5519, | |
| "step": 2208 | |
| }, | |
| { | |
| "epoch": 2.016201620162016, | |
| "grad_norm": 0.9655335545539856, | |
| "learning_rate": 7.451771608751854e-06, | |
| "loss": 0.5848, | |
| "step": 2240 | |
| }, | |
| { | |
| "epoch": 2.045004500450045, | |
| "grad_norm": 1.7546676397323608, | |
| "learning_rate": 7.3636491664917746e-06, | |
| "loss": 0.5676, | |
| "step": 2272 | |
| }, | |
| { | |
| "epoch": 2.073807380738074, | |
| "grad_norm": 1.35713791847229, | |
| "learning_rate": 7.274570847682192e-06, | |
| "loss": 0.5995, | |
| "step": 2304 | |
| }, | |
| { | |
| "epoch": 2.102610261026103, | |
| "grad_norm": 1.368743896484375, | |
| "learning_rate": 7.184572676229373e-06, | |
| "loss": 0.5532, | |
| "step": 2336 | |
| }, | |
| { | |
| "epoch": 2.1314131413141313, | |
| "grad_norm": 1.5946100950241089, | |
| "learning_rate": 7.093691048034539e-06, | |
| "loss": 0.5732, | |
| "step": 2368 | |
| }, | |
| { | |
| "epoch": 2.16021602160216, | |
| "grad_norm": 1.4962303638458252, | |
| "learning_rate": 7.001962716275111e-06, | |
| "loss": 0.6021, | |
| "step": 2400 | |
| }, | |
| { | |
| "epoch": 2.189018901890189, | |
| "grad_norm": 2.6867597103118896, | |
| "learning_rate": 6.909424776541456e-06, | |
| "loss": 0.5429, | |
| "step": 2432 | |
| }, | |
| { | |
| "epoch": 2.217821782178218, | |
| "grad_norm": 1.6566091775894165, | |
| "learning_rate": 6.816114651835171e-06, | |
| "loss": 0.5284, | |
| "step": 2464 | |
| }, | |
| { | |
| "epoch": 2.2466246624662465, | |
| "grad_norm": 1.1805468797683716, | |
| "learning_rate": 6.7220700774349524e-06, | |
| "loss": 0.5797, | |
| "step": 2496 | |
| }, | |
| { | |
| "epoch": 2.2754275427542754, | |
| "grad_norm": 1.6449726819992065, | |
| "learning_rate": 6.627329085636178e-06, | |
| "loss": 0.5624, | |
| "step": 2528 | |
| }, | |
| { | |
| "epoch": 2.3042304230423043, | |
| "grad_norm": 1.2300235033035278, | |
| "learning_rate": 6.531929990370387e-06, | |
| "loss": 0.5629, | |
| "step": 2560 | |
| }, | |
| { | |
| "epoch": 2.333033303330333, | |
| "grad_norm": 1.2485629320144653, | |
| "learning_rate": 6.43591137171084e-06, | |
| "loss": 0.561, | |
| "step": 2592 | |
| }, | |
| { | |
| "epoch": 2.3618361836183617, | |
| "grad_norm": 1.2627168893814087, | |
| "learning_rate": 6.3393120602704675e-06, | |
| "loss": 0.5707, | |
| "step": 2624 | |
| }, | |
| { | |
| "epoch": 2.3906390639063906, | |
| "grad_norm": 1.8480335474014282, | |
| "learning_rate": 6.242171121498498e-06, | |
| "loss": 0.5759, | |
| "step": 2656 | |
| }, | |
| { | |
| "epoch": 2.4194419441944195, | |
| "grad_norm": 1.5584763288497925, | |
| "learning_rate": 6.144527839882107e-06, | |
| "loss": 0.5641, | |
| "step": 2688 | |
| }, | |
| { | |
| "epoch": 2.4482448244824484, | |
| "grad_norm": 1.1628447771072388, | |
| "learning_rate": 6.046421703059493e-06, | |
| "loss": 0.5853, | |
| "step": 2720 | |
| }, | |
| { | |
| "epoch": 2.477047704770477, | |
| "grad_norm": 1.420699119567871, | |
| "learning_rate": 5.9478923858507955e-06, | |
| "loss": 0.5572, | |
| "step": 2752 | |
| }, | |
| { | |
| "epoch": 2.5058505850585058, | |
| "grad_norm": 1.5408899784088135, | |
| "learning_rate": 5.848979734213309e-06, | |
| "loss": 0.5835, | |
| "step": 2784 | |
| }, | |
| { | |
| "epoch": 2.5346534653465347, | |
| "grad_norm": 3.342973470687866, | |
| "learning_rate": 5.74972374912751e-06, | |
| "loss": 0.5648, | |
| "step": 2816 | |
| }, | |
| { | |
| "epoch": 2.5634563456345636, | |
| "grad_norm": 1.3234657049179077, | |
| "learning_rate": 5.650164570420358e-06, | |
| "loss": 0.5838, | |
| "step": 2848 | |
| }, | |
| { | |
| "epoch": 2.592259225922592, | |
| "grad_norm": 3.896916627883911, | |
| "learning_rate": 5.5503424605324715e-06, | |
| "loss": 0.5716, | |
| "step": 2880 | |
| }, | |
| { | |
| "epoch": 2.621062106210621, | |
| "grad_norm": 1.4137349128723145, | |
| "learning_rate": 5.450297788235718e-06, | |
| "loss": 0.5895, | |
| "step": 2912 | |
| }, | |
| { | |
| "epoch": 2.64986498649865, | |
| "grad_norm": 1.852388620376587, | |
| "learning_rate": 5.350071012307776e-06, | |
| "loss": 0.5525, | |
| "step": 2944 | |
| }, | |
| { | |
| "epoch": 2.678667866786679, | |
| "grad_norm": 1.226164698600769, | |
| "learning_rate": 5.24970266517034e-06, | |
| "loss": 0.5549, | |
| "step": 2976 | |
| }, | |
| { | |
| "epoch": 2.7074707470747077, | |
| "grad_norm": 1.257299542427063, | |
| "learning_rate": 5.149233336497501e-06, | |
| "loss": 0.5797, | |
| "step": 3008 | |
| }, | |
| { | |
| "epoch": 2.736273627362736, | |
| "grad_norm": 1.4466633796691895, | |
| "learning_rate": 5.048703656801004e-06, | |
| "loss": 0.5629, | |
| "step": 3040 | |
| }, | |
| { | |
| "epoch": 2.765076507650765, | |
| "grad_norm": 2.2335245609283447, | |
| "learning_rate": 4.948154280998981e-06, | |
| "loss": 0.5626, | |
| "step": 3072 | |
| }, | |
| { | |
| "epoch": 2.793879387938794, | |
| "grad_norm": 1.4836534261703491, | |
| "learning_rate": 4.847625871974807e-06, | |
| "loss": 0.5659, | |
| "step": 3104 | |
| }, | |
| { | |
| "epoch": 2.8226822682268224, | |
| "grad_norm": 1.282368540763855, | |
| "learning_rate": 4.747159084132742e-06, | |
| "loss": 0.5957, | |
| "step": 3136 | |
| }, | |
| { | |
| "epoch": 2.8514851485148514, | |
| "grad_norm": 1.2289072275161743, | |
| "learning_rate": 4.646794546957001e-06, | |
| "loss": 0.5754, | |
| "step": 3168 | |
| }, | |
| { | |
| "epoch": 2.8802880288028803, | |
| "grad_norm": 1.3047752380371094, | |
| "learning_rate": 4.546572848580907e-06, | |
| "loss": 0.6166, | |
| "step": 3200 | |
| }, | |
| { | |
| "epoch": 2.909090909090909, | |
| "grad_norm": 1.365952730178833, | |
| "learning_rate": 4.446534519372744e-06, | |
| "loss": 0.5474, | |
| "step": 3232 | |
| }, | |
| { | |
| "epoch": 2.937893789378938, | |
| "grad_norm": 1.4504116773605347, | |
| "learning_rate": 4.34672001554501e-06, | |
| "loss": 0.564, | |
| "step": 3264 | |
| }, | |
| { | |
| "epoch": 2.9666966696669665, | |
| "grad_norm": 2.4160330295562744, | |
| "learning_rate": 4.247169702793625e-06, | |
| "loss": 0.5578, | |
| "step": 3296 | |
| }, | |
| { | |
| "epoch": 2.9954995499549955, | |
| "grad_norm": 1.876542568206787, | |
| "learning_rate": 4.1479238399737485e-06, | |
| "loss": 0.5885, | |
| "step": 3328 | |
| }, | |
| { | |
| "epoch": 3.0243024302430244, | |
| "grad_norm": 2.028958320617676, | |
| "learning_rate": 4.049022562818829e-06, | |
| "loss": 0.5875, | |
| "step": 3360 | |
| }, | |
| { | |
| "epoch": 3.0531053105310533, | |
| "grad_norm": 1.76653254032135, | |
| "learning_rate": 3.950505867709418e-06, | |
| "loss": 0.5559, | |
| "step": 3392 | |
| }, | |
| { | |
| "epoch": 3.0819081908190817, | |
| "grad_norm": 1.768809199333191, | |
| "learning_rate": 3.852413595498343e-06, | |
| "loss": 0.5792, | |
| "step": 3424 | |
| }, | |
| { | |
| "epoch": 3.1107110711071106, | |
| "grad_norm": 1.5263768434524536, | |
| "learning_rate": 3.7547854153987973e-06, | |
| "loss": 0.5742, | |
| "step": 3456 | |
| }, | |
| { | |
| "epoch": 3.1395139513951396, | |
| "grad_norm": 1.5573300123214722, | |
| "learning_rate": 3.6576608089418188e-06, | |
| "loss": 0.5664, | |
| "step": 3488 | |
| }, | |
| { | |
| "epoch": 3.1683168316831685, | |
| "grad_norm": 1.0428520441055298, | |
| "learning_rate": 3.561079054009678e-06, | |
| "loss": 0.5434, | |
| "step": 3520 | |
| }, | |
| { | |
| "epoch": 3.197119711971197, | |
| "grad_norm": 1.465216875076294, | |
| "learning_rate": 3.4650792089516238e-06, | |
| "loss": 0.5603, | |
| "step": 3552 | |
| }, | |
| { | |
| "epoch": 3.225922592259226, | |
| "grad_norm": 1.1263318061828613, | |
| "learning_rate": 3.3697000967884107e-06, | |
| "loss": 0.5636, | |
| "step": 3584 | |
| }, | |
| { | |
| "epoch": 3.2547254725472547, | |
| "grad_norm": 1.6343493461608887, | |
| "learning_rate": 3.274980289511995e-06, | |
| "loss": 0.5797, | |
| "step": 3616 | |
| }, | |
| { | |
| "epoch": 3.2835283528352837, | |
| "grad_norm": 2.2945406436920166, | |
| "learning_rate": 3.1809580924867358e-06, | |
| "loss": 0.5638, | |
| "step": 3648 | |
| }, | |
| { | |
| "epoch": 3.312331233123312, | |
| "grad_norm": 1.1425927877426147, | |
| "learning_rate": 3.087671528958445e-06, | |
| "loss": 0.559, | |
| "step": 3680 | |
| }, | |
| { | |
| "epoch": 3.341134113411341, | |
| "grad_norm": 1.1738224029541016, | |
| "learning_rate": 2.9951583246775147e-06, | |
| "loss": 0.5947, | |
| "step": 3712 | |
| }, | |
| { | |
| "epoch": 3.36993699369937, | |
| "grad_norm": 1.0960267782211304, | |
| "learning_rate": 2.9034558926423427e-06, | |
| "loss": 0.5609, | |
| "step": 3744 | |
| }, | |
| { | |
| "epoch": 3.398739873987399, | |
| "grad_norm": 2.027769088745117, | |
| "learning_rate": 2.812601317969266e-06, | |
| "loss": 0.5932, | |
| "step": 3776 | |
| }, | |
| { | |
| "epoch": 3.4275427542754278, | |
| "grad_norm": 1.8806779384613037, | |
| "learning_rate": 2.7226313428950703e-06, | |
| "loss": 0.546, | |
| "step": 3808 | |
| }, | |
| { | |
| "epoch": 3.4563456345634562, | |
| "grad_norm": 1.778334617614746, | |
| "learning_rate": 2.633582351918156e-06, | |
| "loss": 0.5645, | |
| "step": 3840 | |
| }, | |
| { | |
| "epoch": 3.485148514851485, | |
| "grad_norm": 1.2001926898956299, | |
| "learning_rate": 2.545490357084409e-06, | |
| "loss": 0.5634, | |
| "step": 3872 | |
| }, | |
| { | |
| "epoch": 3.513951395139514, | |
| "grad_norm": 3.2067384719848633, | |
| "learning_rate": 2.4583909834236563e-06, | |
| "loss": 0.5409, | |
| "step": 3904 | |
| }, | |
| { | |
| "epoch": 3.5427542754275425, | |
| "grad_norm": 2.1273581981658936, | |
| "learning_rate": 2.372319454542659e-06, | |
| "loss": 0.5632, | |
| "step": 3936 | |
| }, | |
| { | |
| "epoch": 3.5715571557155714, | |
| "grad_norm": 1.4944703578948975, | |
| "learning_rate": 2.28731057838043e-06, | |
| "loss": 0.5438, | |
| "step": 3968 | |
| }, | |
| { | |
| "epoch": 3.6003600360036003, | |
| "grad_norm": 1.5785070657730103, | |
| "learning_rate": 2.203398733131661e-06, | |
| "loss": 0.5491, | |
| "step": 4000 | |
| }, | |
| { | |
| "epoch": 3.6291629162916292, | |
| "grad_norm": 1.9863430261611938, | |
| "learning_rate": 2.120617853343926e-06, | |
| "loss": 0.5681, | |
| "step": 4032 | |
| }, | |
| { | |
| "epoch": 3.657965796579658, | |
| "grad_norm": 1.9844887256622314, | |
| "learning_rate": 2.0390014161943224e-06, | |
| "loss": 0.55, | |
| "step": 4064 | |
| }, | |
| { | |
| "epoch": 3.6867686768676866, | |
| "grad_norm": 1.2234737873077393, | |
| "learning_rate": 1.958582427951051e-06, | |
| "loss": 0.5517, | |
| "step": 4096 | |
| }, | |
| { | |
| "epoch": 3.7155715571557155, | |
| "grad_norm": 1.3753498792648315, | |
| "learning_rate": 1.8793934106254557e-06, | |
| "loss": 0.5613, | |
| "step": 4128 | |
| }, | |
| { | |
| "epoch": 3.7443744374437444, | |
| "grad_norm": 1.6427245140075684, | |
| "learning_rate": 1.8014663888198763e-06, | |
| "loss": 0.5548, | |
| "step": 4160 | |
| }, | |
| { | |
| "epoch": 3.7731773177317733, | |
| "grad_norm": 2.217555522918701, | |
| "learning_rate": 1.7248328767766875e-06, | |
| "loss": 0.5803, | |
| "step": 4192 | |
| }, | |
| { | |
| "epoch": 3.801980198019802, | |
| "grad_norm": 1.24624764919281, | |
| "learning_rate": 1.6495238656337076e-06, | |
| "loss": 0.6296, | |
| "step": 4224 | |
| }, | |
| { | |
| "epoch": 3.8307830783078307, | |
| "grad_norm": 1.3684077262878418, | |
| "learning_rate": 1.575569810891156e-06, | |
| "loss": 0.5865, | |
| "step": 4256 | |
| }, | |
| { | |
| "epoch": 3.8595859585958596, | |
| "grad_norm": 2.7931735515594482, | |
| "learning_rate": 1.503000620095244e-06, | |
| "loss": 0.5758, | |
| "step": 4288 | |
| }, | |
| { | |
| "epoch": 3.8883888388838885, | |
| "grad_norm": 1.8946348428726196, | |
| "learning_rate": 1.4318456407433434e-06, | |
| "loss": 0.563, | |
| "step": 4320 | |
| }, | |
| { | |
| "epoch": 3.9171917191719174, | |
| "grad_norm": 1.750241756439209, | |
| "learning_rate": 1.3621336484156456e-06, | |
| "loss": 0.5692, | |
| "step": 4352 | |
| }, | |
| { | |
| "epoch": 3.945994599459946, | |
| "grad_norm": 1.4938685894012451, | |
| "learning_rate": 1.2938928351381224e-06, | |
| "loss": 0.5379, | |
| "step": 4384 | |
| }, | |
| { | |
| "epoch": 3.974797479747975, | |
| "grad_norm": 4.275557041168213, | |
| "learning_rate": 1.2271507979814624e-06, | |
| "loss": 0.534, | |
| "step": 4416 | |
| }, | |
| { | |
| "epoch": 4.003600360036003, | |
| "grad_norm": 1.308555245399475, | |
| "learning_rate": 1.1619345279006212e-06, | |
| "loss": 0.5957, | |
| "step": 4448 | |
| }, | |
| { | |
| "epoch": 4.032403240324032, | |
| "grad_norm": 1.5460844039916992, | |
| "learning_rate": 1.0982703988194876e-06, | |
| "loss": 0.5711, | |
| "step": 4480 | |
| }, | |
| { | |
| "epoch": 4.061206120612061, | |
| "grad_norm": 1.5028718709945679, | |
| "learning_rate": 1.0361841569650816e-06, | |
| "loss": 0.5848, | |
| "step": 4512 | |
| }, | |
| { | |
| "epoch": 4.09000900090009, | |
| "grad_norm": 1.7333652973175049, | |
| "learning_rate": 9.75700910455592e-07, | |
| "loss": 0.5748, | |
| "step": 4544 | |
| }, | |
| { | |
| "epoch": 4.118811881188119, | |
| "grad_norm": 1.941709041595459, | |
| "learning_rate": 9.168451191464822e-07, | |
| "loss": 0.5728, | |
| "step": 4576 | |
| }, | |
| { | |
| "epoch": 4.147614761476148, | |
| "grad_norm": 1.418080449104309, | |
| "learning_rate": 8.596405847387462e-07, | |
| "loss": 0.5787, | |
| "step": 4608 | |
| }, | |
| { | |
| "epoch": 4.176417641764177, | |
| "grad_norm": 1.187457799911499, | |
| "learning_rate": 8.041104411533329e-07, | |
| "loss": 0.5324, | |
| "step": 4640 | |
| }, | |
| { | |
| "epoch": 4.205220522052206, | |
| "grad_norm": 1.3761279582977295, | |
| "learning_rate": 7.502771451756197e-07, | |
| "loss": 0.5721, | |
| "step": 4672 | |
| }, | |
| { | |
| "epoch": 4.234023402340234, | |
| "grad_norm": 0.9498482942581177, | |
| "learning_rate": 6.981624673737336e-07, | |
| "loss": 0.5712, | |
| "step": 4704 | |
| }, | |
| { | |
| "epoch": 4.262826282628263, | |
| "grad_norm": 1.5619556903839111, | |
| "learning_rate": 6.477874832943781e-07, | |
| "loss": 0.5483, | |
| "step": 4736 | |
| }, | |
| { | |
| "epoch": 4.2916291629162915, | |
| "grad_norm": 1.354587435722351, | |
| "learning_rate": 5.991725649397279e-07, | |
| "loss": 0.5314, | |
| "step": 4768 | |
| }, | |
| { | |
| "epoch": 4.32043204320432, | |
| "grad_norm": 1.8367897272109985, | |
| "learning_rate": 5.523373725288506e-07, | |
| "loss": 0.5895, | |
| "step": 4800 | |
| }, | |
| { | |
| "epoch": 4.349234923492349, | |
| "grad_norm": 1.9770034551620483, | |
| "learning_rate": 5.073008465469731e-07, | |
| "loss": 0.5482, | |
| "step": 4832 | |
| }, | |
| { | |
| "epoch": 4.378037803780378, | |
| "grad_norm": 1.3692898750305176, | |
| "learning_rate": 4.640812000858108e-07, | |
| "loss": 0.5754, | |
| "step": 4864 | |
| }, | |
| { | |
| "epoch": 4.406840684068407, | |
| "grad_norm": 1.9973537921905518, | |
| "learning_rate": 4.226959114780699e-07, | |
| "loss": 0.5309, | |
| "step": 4896 | |
| }, | |
| { | |
| "epoch": 4.435643564356436, | |
| "grad_norm": 1.6791578531265259, | |
| "learning_rate": 3.831617172290808e-07, | |
| "loss": 0.5791, | |
| "step": 4928 | |
| }, | |
| { | |
| "epoch": 4.464446444644464, | |
| "grad_norm": 1.2366864681243896, | |
| "learning_rate": 3.454946052484376e-07, | |
| "loss": 0.5401, | |
| "step": 4960 | |
| }, | |
| { | |
| "epoch": 4.493249324932493, | |
| "grad_norm": 1.513241171836853, | |
| "learning_rate": 3.0970980838437416e-07, | |
| "loss": 0.5923, | |
| "step": 4992 | |
| }, | |
| { | |
| "epoch": 4.522052205220522, | |
| "grad_norm": 1.3842517137527466, | |
| "learning_rate": 2.758217982634903e-07, | |
| "loss": 0.5654, | |
| "step": 5024 | |
| }, | |
| { | |
| "epoch": 4.550855085508551, | |
| "grad_norm": 1.9963948726654053, | |
| "learning_rate": 2.4384427943832335e-07, | |
| "loss": 0.5736, | |
| "step": 5056 | |
| }, | |
| { | |
| "epoch": 4.57965796579658, | |
| "grad_norm": 1.2688953876495361, | |
| "learning_rate": 2.1379018384513116e-07, | |
| "loss": 0.5604, | |
| "step": 5088 | |
| }, | |
| { | |
| "epoch": 4.608460846084609, | |
| "grad_norm": 1.2119768857955933, | |
| "learning_rate": 1.856716655741242e-07, | |
| "loss": 0.561, | |
| "step": 5120 | |
| }, | |
| { | |
| "epoch": 4.6372637263726375, | |
| "grad_norm": 1.7336974143981934, | |
| "learning_rate": 1.5950009595426474e-07, | |
| "loss": 0.579, | |
| "step": 5152 | |
| }, | |
| { | |
| "epoch": 4.666066606660666, | |
| "grad_norm": 3.910386562347412, | |
| "learning_rate": 1.3528605895461734e-07, | |
| "loss": 0.5591, | |
| "step": 5184 | |
| }, | |
| { | |
| "epoch": 4.694869486948695, | |
| "grad_norm": 1.0510704517364502, | |
| "learning_rate": 1.130393469041241e-07, | |
| "loss": 0.5636, | |
| "step": 5216 | |
| }, | |
| { | |
| "epoch": 4.723672367236723, | |
| "grad_norm": 1.8472617864608765, | |
| "learning_rate": 9.27689565315093e-08, | |
| "loss": 0.5553, | |
| "step": 5248 | |
| }, | |
| { | |
| "epoch": 4.752475247524752, | |
| "grad_norm": 1.5597736835479736, | |
| "learning_rate": 7.448308532694237e-08, | |
| "loss": 0.5721, | |
| "step": 5280 | |
| }, | |
| { | |
| "epoch": 4.781278127812781, | |
| "grad_norm": 1.633854627609253, | |
| "learning_rate": 5.8189128226913695e-08, | |
| "loss": 0.5994, | |
| "step": 5312 | |
| }, | |
| { | |
| "epoch": 4.81008100810081, | |
| "grad_norm": 4.339478015899658, | |
| "learning_rate": 4.389367462367011e-08, | |
| "loss": 0.5648, | |
| "step": 5344 | |
| }, | |
| { | |
| "epoch": 4.838883888388839, | |
| "grad_norm": 2.3053483963012695, | |
| "learning_rate": 3.160250570041601e-08, | |
| "loss": 0.5796, | |
| "step": 5376 | |
| }, | |
| { | |
| "epoch": 4.867686768676868, | |
| "grad_norm": 1.711835503578186, | |
| "learning_rate": 2.1320592093364013e-08, | |
| "loss": 0.5467, | |
| "step": 5408 | |
| }, | |
| { | |
| "epoch": 4.896489648964897, | |
| "grad_norm": 1.162542700767517, | |
| "learning_rate": 1.305209188157186e-08, | |
| "loss": 0.5602, | |
| "step": 5440 | |
| }, | |
| { | |
| "epoch": 4.925292529252925, | |
| "grad_norm": 1.5160582065582275, | |
| "learning_rate": 6.8003489053886094e-09, | |
| "loss": 0.5406, | |
| "step": 5472 | |
| }, | |
| { | |
| "epoch": 4.954095409540954, | |
| "grad_norm": 1.8680617809295654, | |
| "learning_rate": 2.5678914141791243e-09, | |
| "loss": 0.553, | |
| "step": 5504 | |
| }, | |
| { | |
| "epoch": 4.982898289828983, | |
| "grad_norm": 1.2880302667617798, | |
| "learning_rate": 3.564310438830543e-10, | |
| "loss": 0.5609, | |
| "step": 5536 | |
| }, | |
| { | |
| "epoch": 5.0, | |
| "step": 5555, | |
| "total_flos": 5.796630098241126e+16, | |
| "train_loss": 0.6057587686258861, | |
| "train_runtime": 1046.8819, | |
| "train_samples_per_second": 10.612, | |
| "train_steps_per_second": 5.306 | |
| } | |
| ], | |
| "logging_steps": 32, | |
| "max_steps": 5555, | |
| "num_input_tokens_seen": 0, | |
| "num_train_epochs": 5, | |
| "save_steps": 500, | |
| "stateful_callbacks": { | |
| "TrainerControl": { | |
| "args": { | |
| "should_epoch_stop": false, | |
| "should_evaluate": false, | |
| "should_log": false, | |
| "should_save": true, | |
| "should_training_stop": true | |
| }, | |
| "attributes": {} | |
| } | |
| }, | |
| "total_flos": 5.796630098241126e+16, | |
| "train_batch_size": 1, | |
| "trial_name": null, | |
| "trial_params": null | |
| } | |