Vigogne_Qwen2.5-1.5B-Instruct / trainer_state.json
moussaKam's picture
upload
d235d66 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.9998107374929026,
"eval_steps": 500,
"global_step": 3962,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0025235000946312535,
"grad_norm": 3.572803497314453,
"learning_rate": 1.2594458438287156e-06,
"loss": 1.0672,
"step": 10
},
{
"epoch": 0.005047000189262507,
"grad_norm": 1.6470932960510254,
"learning_rate": 2.518891687657431e-06,
"loss": 1.0001,
"step": 20
},
{
"epoch": 0.007570500283893761,
"grad_norm": 1.1262171268463135,
"learning_rate": 3.7783375314861467e-06,
"loss": 0.9414,
"step": 30
},
{
"epoch": 0.010094000378525014,
"grad_norm": 0.8495129346847534,
"learning_rate": 5.037783375314862e-06,
"loss": 0.9321,
"step": 40
},
{
"epoch": 0.012617500473156268,
"grad_norm": 0.8612141013145447,
"learning_rate": 6.297229219143577e-06,
"loss": 0.8746,
"step": 50
},
{
"epoch": 0.015141000567787522,
"grad_norm": 0.8412306308746338,
"learning_rate": 7.556675062972293e-06,
"loss": 0.9044,
"step": 60
},
{
"epoch": 0.017664500662418776,
"grad_norm": 0.8401440978050232,
"learning_rate": 8.816120906801008e-06,
"loss": 0.9022,
"step": 70
},
{
"epoch": 0.020188000757050028,
"grad_norm": 0.86940997838974,
"learning_rate": 1.0075566750629725e-05,
"loss": 0.8887,
"step": 80
},
{
"epoch": 0.022711500851681284,
"grad_norm": 0.8858376741409302,
"learning_rate": 1.133501259445844e-05,
"loss": 0.8715,
"step": 90
},
{
"epoch": 0.025235000946312536,
"grad_norm": 0.8635324239730835,
"learning_rate": 1.2594458438287154e-05,
"loss": 0.8709,
"step": 100
},
{
"epoch": 0.027758501040943788,
"grad_norm": 0.9026916027069092,
"learning_rate": 1.385390428211587e-05,
"loss": 0.9428,
"step": 110
},
{
"epoch": 0.030282001135575044,
"grad_norm": 0.6949167847633362,
"learning_rate": 1.5113350125944587e-05,
"loss": 0.9041,
"step": 120
},
{
"epoch": 0.0328055012302063,
"grad_norm": 0.9495663046836853,
"learning_rate": 1.63727959697733e-05,
"loss": 0.8636,
"step": 130
},
{
"epoch": 0.03532900132483755,
"grad_norm": 0.8648976683616638,
"learning_rate": 1.7632241813602016e-05,
"loss": 0.8471,
"step": 140
},
{
"epoch": 0.037852501419468804,
"grad_norm": 0.8415878415107727,
"learning_rate": 1.8891687657430733e-05,
"loss": 0.9108,
"step": 150
},
{
"epoch": 0.040376001514100056,
"grad_norm": 0.8484784364700317,
"learning_rate": 2.015113350125945e-05,
"loss": 0.8883,
"step": 160
},
{
"epoch": 0.04289950160873131,
"grad_norm": 0.8620557188987732,
"learning_rate": 2.1410579345088162e-05,
"loss": 0.8657,
"step": 170
},
{
"epoch": 0.04542300170336257,
"grad_norm": 0.8222241401672363,
"learning_rate": 2.267002518891688e-05,
"loss": 0.9174,
"step": 180
},
{
"epoch": 0.04794650179799382,
"grad_norm": 0.8526296019554138,
"learning_rate": 2.392947103274559e-05,
"loss": 0.8997,
"step": 190
},
{
"epoch": 0.05047000189262507,
"grad_norm": 0.8018633723258972,
"learning_rate": 2.5188916876574308e-05,
"loss": 0.9076,
"step": 200
},
{
"epoch": 0.052993501987256324,
"grad_norm": 0.859157145023346,
"learning_rate": 2.6448362720403024e-05,
"loss": 0.8891,
"step": 210
},
{
"epoch": 0.055517002081887576,
"grad_norm": 0.7516281604766846,
"learning_rate": 2.770780856423174e-05,
"loss": 0.9026,
"step": 220
},
{
"epoch": 0.058040502176518835,
"grad_norm": 0.8353524804115295,
"learning_rate": 2.8967254408060457e-05,
"loss": 0.8393,
"step": 230
},
{
"epoch": 0.06056400227115009,
"grad_norm": 0.7622519731521606,
"learning_rate": 3.0226700251889174e-05,
"loss": 0.8524,
"step": 240
},
{
"epoch": 0.06308750236578134,
"grad_norm": 0.8780621290206909,
"learning_rate": 3.148614609571788e-05,
"loss": 0.9286,
"step": 250
},
{
"epoch": 0.0656110024604126,
"grad_norm": 0.9684115052223206,
"learning_rate": 3.27455919395466e-05,
"loss": 0.8974,
"step": 260
},
{
"epoch": 0.06813450255504384,
"grad_norm": 0.8870697617530823,
"learning_rate": 3.4005037783375316e-05,
"loss": 0.8969,
"step": 270
},
{
"epoch": 0.0706580026496751,
"grad_norm": 0.8952546119689941,
"learning_rate": 3.526448362720403e-05,
"loss": 0.8689,
"step": 280
},
{
"epoch": 0.07318150274430635,
"grad_norm": 0.9221532344818115,
"learning_rate": 3.652392947103275e-05,
"loss": 0.883,
"step": 290
},
{
"epoch": 0.07570500283893761,
"grad_norm": 0.9585578441619873,
"learning_rate": 3.7783375314861465e-05,
"loss": 0.8508,
"step": 300
},
{
"epoch": 0.07822850293356887,
"grad_norm": 0.9648734331130981,
"learning_rate": 3.904282115869018e-05,
"loss": 0.8982,
"step": 310
},
{
"epoch": 0.08075200302820011,
"grad_norm": 0.8147997260093689,
"learning_rate": 4.03022670025189e-05,
"loss": 0.8694,
"step": 320
},
{
"epoch": 0.08327550312283137,
"grad_norm": 0.8099369406700134,
"learning_rate": 4.1561712846347615e-05,
"loss": 0.8929,
"step": 330
},
{
"epoch": 0.08579900321746262,
"grad_norm": 0.8512017130851746,
"learning_rate": 4.2821158690176324e-05,
"loss": 0.86,
"step": 340
},
{
"epoch": 0.08832250331209388,
"grad_norm": 0.8499712347984314,
"learning_rate": 4.408060453400504e-05,
"loss": 0.8423,
"step": 350
},
{
"epoch": 0.09084600340672513,
"grad_norm": 0.8530069589614868,
"learning_rate": 4.534005037783376e-05,
"loss": 0.8852,
"step": 360
},
{
"epoch": 0.09336950350135638,
"grad_norm": 0.8837921023368835,
"learning_rate": 4.659949622166247e-05,
"loss": 0.8864,
"step": 370
},
{
"epoch": 0.09589300359598764,
"grad_norm": 0.8840718865394592,
"learning_rate": 4.785894206549118e-05,
"loss": 0.8999,
"step": 380
},
{
"epoch": 0.09841650369061888,
"grad_norm": 0.7395176887512207,
"learning_rate": 4.91183879093199e-05,
"loss": 0.8954,
"step": 390
},
{
"epoch": 0.10094000378525014,
"grad_norm": 0.8697903752326965,
"learning_rate": 4.999991263591223e-05,
"loss": 0.8353,
"step": 400
},
{
"epoch": 0.1034635038798814,
"grad_norm": 0.8527745008468628,
"learning_rate": 4.9998359513560176e-05,
"loss": 0.8591,
"step": 410
},
{
"epoch": 0.10598700397451265,
"grad_norm": 0.806920051574707,
"learning_rate": 4.999486510586282e-05,
"loss": 0.9076,
"step": 420
},
{
"epoch": 0.1085105040691439,
"grad_norm": 0.8428360223770142,
"learning_rate": 4.9989429684183686e-05,
"loss": 0.9032,
"step": 430
},
{
"epoch": 0.11103400416377515,
"grad_norm": 0.8436294198036194,
"learning_rate": 4.9982053670618626e-05,
"loss": 0.8894,
"step": 440
},
{
"epoch": 0.11355750425840641,
"grad_norm": 0.7664760947227478,
"learning_rate": 4.997273763796312e-05,
"loss": 0.8732,
"step": 450
},
{
"epoch": 0.11608100435303767,
"grad_norm": 0.9134059548377991,
"learning_rate": 4.996148230966775e-05,
"loss": 0.8438,
"step": 460
},
{
"epoch": 0.11860450444766892,
"grad_norm": 0.849233090877533,
"learning_rate": 4.994828855978202e-05,
"loss": 0.9281,
"step": 470
},
{
"epoch": 0.12112800454230017,
"grad_norm": 0.8473492860794067,
"learning_rate": 4.99331574128865e-05,
"loss": 0.8368,
"step": 480
},
{
"epoch": 0.12365150463693142,
"grad_norm": 0.8474506735801697,
"learning_rate": 4.991609004401324e-05,
"loss": 0.8852,
"step": 490
},
{
"epoch": 0.12617500473156268,
"grad_norm": 0.7737945318222046,
"learning_rate": 4.989708777855453e-05,
"loss": 0.8881,
"step": 500
},
{
"epoch": 0.12869850482619394,
"grad_norm": 0.6961573958396912,
"learning_rate": 4.9876152092159994e-05,
"loss": 0.9173,
"step": 510
},
{
"epoch": 0.1312220049208252,
"grad_norm": 0.7320950031280518,
"learning_rate": 4.985328461062195e-05,
"loss": 0.8899,
"step": 520
},
{
"epoch": 0.13374550501545643,
"grad_norm": 0.7261886596679688,
"learning_rate": 4.98284871097492e-05,
"loss": 0.8855,
"step": 530
},
{
"epoch": 0.1362690051100877,
"grad_norm": 0.7850842475891113,
"learning_rate": 4.98017615152291e-05,
"loss": 0.884,
"step": 540
},
{
"epoch": 0.13879250520471895,
"grad_norm": 0.8015512228012085,
"learning_rate": 4.977310990247807e-05,
"loss": 0.8767,
"step": 550
},
{
"epoch": 0.1413160052993502,
"grad_norm": 0.8083379864692688,
"learning_rate": 4.974253449648031e-05,
"loss": 0.8861,
"step": 560
},
{
"epoch": 0.14383950539398146,
"grad_norm": 0.7726438045501709,
"learning_rate": 4.971003767161516e-05,
"loss": 0.8747,
"step": 570
},
{
"epoch": 0.1463630054886127,
"grad_norm": 0.7719607949256897,
"learning_rate": 4.9675621951472584e-05,
"loss": 0.8862,
"step": 580
},
{
"epoch": 0.14888650558324396,
"grad_norm": 0.7348030209541321,
"learning_rate": 4.9639290008657304e-05,
"loss": 0.8915,
"step": 590
},
{
"epoch": 0.15141000567787521,
"grad_norm": 0.7903275489807129,
"learning_rate": 4.960104466458118e-05,
"loss": 0.8916,
"step": 600
},
{
"epoch": 0.15393350577250647,
"grad_norm": 0.778893232345581,
"learning_rate": 4.956088888924414e-05,
"loss": 0.8674,
"step": 610
},
{
"epoch": 0.15645700586713773,
"grad_norm": 0.7972787618637085,
"learning_rate": 4.951882580100353e-05,
"loss": 0.8908,
"step": 620
},
{
"epoch": 0.15898050596176896,
"grad_norm": 0.741663932800293,
"learning_rate": 4.947485866633199e-05,
"loss": 0.8876,
"step": 630
},
{
"epoch": 0.16150400605640022,
"grad_norm": 0.8059448003768921,
"learning_rate": 4.94289908995637e-05,
"loss": 0.8164,
"step": 640
},
{
"epoch": 0.16402750615103148,
"grad_norm": 0.8004572987556458,
"learning_rate": 4.938122606262936e-05,
"loss": 0.9075,
"step": 650
},
{
"epoch": 0.16655100624566274,
"grad_norm": 1.5734790563583374,
"learning_rate": 4.9331567864779457e-05,
"loss": 0.9146,
"step": 660
},
{
"epoch": 0.169074506340294,
"grad_norm": 0.7215042114257812,
"learning_rate": 4.928002016229634e-05,
"loss": 0.885,
"step": 670
},
{
"epoch": 0.17159800643492523,
"grad_norm": 0.6652220487594604,
"learning_rate": 4.9226586958194647e-05,
"loss": 0.9085,
"step": 680
},
{
"epoch": 0.1741215065295565,
"grad_norm": 0.676682710647583,
"learning_rate": 4.9171272401910504e-05,
"loss": 0.837,
"step": 690
},
{
"epoch": 0.17664500662418775,
"grad_norm": 0.7034597396850586,
"learning_rate": 4.9114080788979284e-05,
"loss": 0.8905,
"step": 700
},
{
"epoch": 0.179168506718819,
"grad_norm": 0.7657853960990906,
"learning_rate": 4.905501656070202e-05,
"loss": 0.8945,
"step": 710
},
{
"epoch": 0.18169200681345027,
"grad_norm": 0.7395844459533691,
"learning_rate": 4.8994084303800525e-05,
"loss": 0.8762,
"step": 720
},
{
"epoch": 0.1842155069080815,
"grad_norm": 0.7073786854743958,
"learning_rate": 4.89312887500612e-05,
"loss": 0.8824,
"step": 730
},
{
"epoch": 0.18673900700271276,
"grad_norm": 0.7239210605621338,
"learning_rate": 4.8866634775967544e-05,
"loss": 0.8855,
"step": 740
},
{
"epoch": 0.18926250709734402,
"grad_norm": 0.6406372785568237,
"learning_rate": 4.880012740232154e-05,
"loss": 0.8775,
"step": 750
},
{
"epoch": 0.19178600719197528,
"grad_norm": 0.76404869556427,
"learning_rate": 4.873177179385368e-05,
"loss": 0.862,
"step": 760
},
{
"epoch": 0.19430950728660654,
"grad_norm": 0.7401562929153442,
"learning_rate": 4.866157325882192e-05,
"loss": 0.8734,
"step": 770
},
{
"epoch": 0.19683300738123777,
"grad_norm": 0.7563286423683167,
"learning_rate": 4.858953724859948e-05,
"loss": 0.8652,
"step": 780
},
{
"epoch": 0.19935650747586903,
"grad_norm": 0.7244860529899597,
"learning_rate": 4.851566935725147e-05,
"loss": 0.8436,
"step": 790
},
{
"epoch": 0.20188000757050029,
"grad_norm": 0.7061064839363098,
"learning_rate": 4.843997532110051e-05,
"loss": 0.8717,
"step": 800
},
{
"epoch": 0.20440350766513155,
"grad_norm": 0.9287751913070679,
"learning_rate": 4.836246101828124e-05,
"loss": 0.884,
"step": 810
},
{
"epoch": 0.2069270077597628,
"grad_norm": 0.6689929366111755,
"learning_rate": 4.828313246828386e-05,
"loss": 0.8871,
"step": 820
},
{
"epoch": 0.20945050785439404,
"grad_norm": 0.7176743149757385,
"learning_rate": 4.820199583148667e-05,
"loss": 0.8799,
"step": 830
},
{
"epoch": 0.2119740079490253,
"grad_norm": 0.6979175209999084,
"learning_rate": 4.811905740867769e-05,
"loss": 0.8874,
"step": 840
},
{
"epoch": 0.21449750804365655,
"grad_norm": 0.780451774597168,
"learning_rate": 4.803432364056535e-05,
"loss": 0.8843,
"step": 850
},
{
"epoch": 0.2170210081382878,
"grad_norm": 0.7175182700157166,
"learning_rate": 4.794780110727832e-05,
"loss": 0.8578,
"step": 860
},
{
"epoch": 0.21954450823291907,
"grad_norm": 0.724116861820221,
"learning_rate": 4.785949652785453e-05,
"loss": 0.8869,
"step": 870
},
{
"epoch": 0.2220680083275503,
"grad_norm": 0.8724785447120667,
"learning_rate": 4.776941675971941e-05,
"loss": 0.8648,
"step": 880
},
{
"epoch": 0.22459150842218156,
"grad_norm": 0.7354777455329895,
"learning_rate": 4.767756879815334e-05,
"loss": 0.8683,
"step": 890
},
{
"epoch": 0.22711500851681282,
"grad_norm": 0.7593517899513245,
"learning_rate": 4.758395977574841e-05,
"loss": 0.9101,
"step": 900
},
{
"epoch": 0.22963850861144408,
"grad_norm": 0.7243201732635498,
"learning_rate": 4.748859696185458e-05,
"loss": 0.8945,
"step": 910
},
{
"epoch": 0.23216200870607534,
"grad_norm": 0.6870005130767822,
"learning_rate": 4.739148776201512e-05,
"loss": 0.8158,
"step": 920
},
{
"epoch": 0.23468550880070657,
"grad_norm": 0.7116649746894836,
"learning_rate": 4.729263971739154e-05,
"loss": 0.8855,
"step": 930
},
{
"epoch": 0.23720900889533783,
"grad_norm": 0.6931096911430359,
"learning_rate": 4.719206050417796e-05,
"loss": 0.8674,
"step": 940
},
{
"epoch": 0.2397325089899691,
"grad_norm": 0.7314079999923706,
"learning_rate": 4.7089757933005016e-05,
"loss": 0.8743,
"step": 950
},
{
"epoch": 0.24225600908460035,
"grad_norm": 0.7538678646087646,
"learning_rate": 4.698573994833332e-05,
"loss": 0.866,
"step": 960
},
{
"epoch": 0.2447795091792316,
"grad_norm": 0.6961751580238342,
"learning_rate": 4.688001462783648e-05,
"loss": 0.8528,
"step": 970
},
{
"epoch": 0.24730300927386284,
"grad_norm": 0.7808176875114441,
"learning_rate": 4.6772590181773866e-05,
"loss": 0.8315,
"step": 980
},
{
"epoch": 0.2498265093684941,
"grad_norm": 0.716074526309967,
"learning_rate": 4.6663474952353004e-05,
"loss": 0.8372,
"step": 990
},
{
"epoch": 0.25235000946312536,
"grad_norm": 0.8192372918128967,
"learning_rate": 4.6552677413081756e-05,
"loss": 0.902,
"step": 1000
},
{
"epoch": 0.2548735095577566,
"grad_norm": 0.7088383436203003,
"learning_rate": 4.644020616811029e-05,
"loss": 0.8847,
"step": 1010
},
{
"epoch": 0.2573970096523879,
"grad_norm": 0.8579234480857849,
"learning_rate": 4.6326069951562924e-05,
"loss": 0.9071,
"step": 1020
},
{
"epoch": 0.25992050974701914,
"grad_norm": 0.6537004709243774,
"learning_rate": 4.6210277626859856e-05,
"loss": 0.8187,
"step": 1030
},
{
"epoch": 0.2624440098416504,
"grad_norm": 0.6265996694564819,
"learning_rate": 4.609283818602884e-05,
"loss": 0.8744,
"step": 1040
},
{
"epoch": 0.2649675099362816,
"grad_norm": 0.7445203065872192,
"learning_rate": 4.5973760749006963e-05,
"loss": 0.8831,
"step": 1050
},
{
"epoch": 0.26749101003091286,
"grad_norm": 0.7054116129875183,
"learning_rate": 4.585305456293235e-05,
"loss": 0.9171,
"step": 1060
},
{
"epoch": 0.2700145101255441,
"grad_norm": 1.429075837135315,
"learning_rate": 4.5730729001426083e-05,
"loss": 0.8894,
"step": 1070
},
{
"epoch": 0.2725380102201754,
"grad_norm": 0.6793610453605652,
"learning_rate": 4.5606793563864316e-05,
"loss": 0.8629,
"step": 1080
},
{
"epoch": 0.27506151031480663,
"grad_norm": 0.6932589411735535,
"learning_rate": 4.548125787464054e-05,
"loss": 0.8564,
"step": 1090
},
{
"epoch": 0.2775850104094379,
"grad_norm": 0.6600730419158936,
"learning_rate": 4.535413168241821e-05,
"loss": 0.8685,
"step": 1100
},
{
"epoch": 0.28010851050406915,
"grad_norm": 0.6784124970436096,
"learning_rate": 4.522542485937369e-05,
"loss": 0.9024,
"step": 1110
},
{
"epoch": 0.2826320105987004,
"grad_norm": 0.6841257214546204,
"learning_rate": 4.509514740042962e-05,
"loss": 0.8698,
"step": 1120
},
{
"epoch": 0.28515551069333167,
"grad_norm": 0.7785212397575378,
"learning_rate": 4.496330942247873e-05,
"loss": 0.8731,
"step": 1130
},
{
"epoch": 0.28767901078796293,
"grad_norm": 0.730110228061676,
"learning_rate": 4.482992116359824e-05,
"loss": 0.8542,
"step": 1140
},
{
"epoch": 0.29020251088259413,
"grad_norm": 0.6644122004508972,
"learning_rate": 4.469499298225473e-05,
"loss": 0.8246,
"step": 1150
},
{
"epoch": 0.2927260109772254,
"grad_norm": 0.7170603275299072,
"learning_rate": 4.455853535649984e-05,
"loss": 0.8576,
"step": 1160
},
{
"epoch": 0.29524951107185665,
"grad_norm": 0.6883527040481567,
"learning_rate": 4.442055888315646e-05,
"loss": 0.8639,
"step": 1170
},
{
"epoch": 0.2977730111664879,
"grad_norm": 0.6971318125724792,
"learning_rate": 4.4281074276995936e-05,
"loss": 0.8218,
"step": 1180
},
{
"epoch": 0.30029651126111917,
"grad_norm": 0.7020850777626038,
"learning_rate": 4.4140092369905914e-05,
"loss": 0.8376,
"step": 1190
},
{
"epoch": 0.30282001135575043,
"grad_norm": 0.6218104362487793,
"learning_rate": 4.399762411004922e-05,
"loss": 0.8741,
"step": 1200
},
{
"epoch": 0.3053435114503817,
"grad_norm": 0.8031836152076721,
"learning_rate": 4.3853680561013647e-05,
"loss": 0.8977,
"step": 1210
},
{
"epoch": 0.30786701154501295,
"grad_norm": 0.6999651789665222,
"learning_rate": 4.370827290095277e-05,
"loss": 0.8628,
"step": 1220
},
{
"epoch": 0.3103905116396442,
"grad_norm": 0.6727817058563232,
"learning_rate": 4.356141242171795e-05,
"loss": 0.8674,
"step": 1230
},
{
"epoch": 0.31291401173427547,
"grad_norm": 0.6965411305427551,
"learning_rate": 4.3413110527981406e-05,
"loss": 0.8416,
"step": 1240
},
{
"epoch": 0.31543751182890667,
"grad_norm": 0.7655733823776245,
"learning_rate": 4.3263378736350566e-05,
"loss": 0.8662,
"step": 1250
},
{
"epoch": 0.31796101192353793,
"grad_norm": 0.7115268111228943,
"learning_rate": 4.311222867447375e-05,
"loss": 0.9022,
"step": 1260
},
{
"epoch": 0.3204845120181692,
"grad_norm": 0.7572771310806274,
"learning_rate": 4.295967208013717e-05,
"loss": 0.8649,
"step": 1270
},
{
"epoch": 0.32300801211280045,
"grad_norm": 0.6894986629486084,
"learning_rate": 4.280572080035348e-05,
"loss": 0.8659,
"step": 1280
},
{
"epoch": 0.3255315122074317,
"grad_norm": 0.6966748833656311,
"learning_rate": 4.2650386790441696e-05,
"loss": 0.8558,
"step": 1290
},
{
"epoch": 0.32805501230206296,
"grad_norm": 0.7241553664207458,
"learning_rate": 4.2493682113098855e-05,
"loss": 0.8666,
"step": 1300
},
{
"epoch": 0.3305785123966942,
"grad_norm": 0.6839144825935364,
"learning_rate": 4.233561893746323e-05,
"loss": 0.8879,
"step": 1310
},
{
"epoch": 0.3331020124913255,
"grad_norm": 0.6955851912498474,
"learning_rate": 4.217620953816935e-05,
"loss": 0.8446,
"step": 1320
},
{
"epoch": 0.33562551258595674,
"grad_norm": 0.6097539067268372,
"learning_rate": 4.2015466294394756e-05,
"loss": 0.8816,
"step": 1330
},
{
"epoch": 0.338149012680588,
"grad_norm": 0.7663230299949646,
"learning_rate": 4.185340168889868e-05,
"loss": 0.8518,
"step": 1340
},
{
"epoch": 0.3406725127752192,
"grad_norm": 0.6563027501106262,
"learning_rate": 4.169002830705274e-05,
"loss": 0.8516,
"step": 1350
},
{
"epoch": 0.34319601286985046,
"grad_norm": 0.639011025428772,
"learning_rate": 4.152535883586352e-05,
"loss": 0.8324,
"step": 1360
},
{
"epoch": 0.3457195129644817,
"grad_norm": 0.7072712779045105,
"learning_rate": 4.135940606298738e-05,
"loss": 0.8445,
"step": 1370
},
{
"epoch": 0.348243013059113,
"grad_norm": 0.6532591581344604,
"learning_rate": 4.119218287573743e-05,
"loss": 0.8293,
"step": 1380
},
{
"epoch": 0.35076651315374424,
"grad_norm": 0.6421136260032654,
"learning_rate": 4.102370226008271e-05,
"loss": 0.8809,
"step": 1390
},
{
"epoch": 0.3532900132483755,
"grad_norm": 0.6466293931007385,
"learning_rate": 4.085397729963976e-05,
"loss": 0.8478,
"step": 1400
},
{
"epoch": 0.35581351334300676,
"grad_norm": 0.7026222348213196,
"learning_rate": 4.06830211746566e-05,
"loss": 0.8855,
"step": 1410
},
{
"epoch": 0.358337013437638,
"grad_norm": 0.7792401313781738,
"learning_rate": 4.051084716098921e-05,
"loss": 0.8523,
"step": 1420
},
{
"epoch": 0.3608605135322693,
"grad_norm": 0.641736626625061,
"learning_rate": 4.0337468629070496e-05,
"loss": 0.8605,
"step": 1430
},
{
"epoch": 0.36338401362690054,
"grad_norm": 0.6911234855651855,
"learning_rate": 4.016289904287212e-05,
"loss": 0.8492,
"step": 1440
},
{
"epoch": 0.36590751372153174,
"grad_norm": 0.7274027466773987,
"learning_rate": 3.9987151958858794e-05,
"loss": 0.8623,
"step": 1450
},
{
"epoch": 0.368431013816163,
"grad_norm": 0.6672956347465515,
"learning_rate": 3.981024102493566e-05,
"loss": 0.8309,
"step": 1460
},
{
"epoch": 0.37095451391079426,
"grad_norm": 0.7280237078666687,
"learning_rate": 3.963217997938834e-05,
"loss": 0.8633,
"step": 1470
},
{
"epoch": 0.3734780140054255,
"grad_norm": 0.749769389629364,
"learning_rate": 3.945298264981614e-05,
"loss": 0.8433,
"step": 1480
},
{
"epoch": 0.3760015141000568,
"grad_norm": 0.7026387453079224,
"learning_rate": 3.927266295205818e-05,
"loss": 0.8665,
"step": 1490
},
{
"epoch": 0.37852501419468804,
"grad_norm": 0.6626182794570923,
"learning_rate": 3.9091234889112815e-05,
"loss": 0.8597,
"step": 1500
},
{
"epoch": 0.3810485142893193,
"grad_norm": 0.6502306461334229,
"learning_rate": 3.8908712550050154e-05,
"loss": 0.8652,
"step": 1510
},
{
"epoch": 0.38357201438395055,
"grad_norm": 0.6474471688270569,
"learning_rate": 3.8725110108917975e-05,
"loss": 0.8258,
"step": 1520
},
{
"epoch": 0.3860955144785818,
"grad_norm": 0.6739810109138489,
"learning_rate": 3.854044182364098e-05,
"loss": 0.8578,
"step": 1530
},
{
"epoch": 0.3886190145732131,
"grad_norm": 0.7030637264251709,
"learning_rate": 3.835472203491367e-05,
"loss": 0.8468,
"step": 1540
},
{
"epoch": 0.3911425146678443,
"grad_norm": 0.6305805444717407,
"learning_rate": 3.816796516508658e-05,
"loss": 0.8476,
"step": 1550
},
{
"epoch": 0.39366601476247554,
"grad_norm": 0.6209976077079773,
"learning_rate": 3.798018571704638e-05,
"loss": 0.8376,
"step": 1560
},
{
"epoch": 0.3961895148571068,
"grad_norm": 0.6698387265205383,
"learning_rate": 3.779139827308956e-05,
"loss": 0.8744,
"step": 1570
},
{
"epoch": 0.39871301495173805,
"grad_norm": 0.7300374507904053,
"learning_rate": 3.760161749379008e-05,
"loss": 0.8609,
"step": 1580
},
{
"epoch": 0.4012365150463693,
"grad_norm": 0.6837272047996521,
"learning_rate": 3.7410858116860836e-05,
"loss": 0.837,
"step": 1590
},
{
"epoch": 0.40376001514100057,
"grad_norm": 0.6649072170257568,
"learning_rate": 3.721913495600923e-05,
"loss": 0.8694,
"step": 1600
},
{
"epoch": 0.40628351523563183,
"grad_norm": 0.5960752367973328,
"learning_rate": 3.7026462899786726e-05,
"loss": 0.8129,
"step": 1610
},
{
"epoch": 0.4088070153302631,
"grad_norm": 0.6648868322372437,
"learning_rate": 3.683285691043272e-05,
"loss": 0.8634,
"step": 1620
},
{
"epoch": 0.41133051542489435,
"grad_norm": 0.7035058736801147,
"learning_rate": 3.663833202271257e-05,
"loss": 0.8685,
"step": 1630
},
{
"epoch": 0.4138540155195256,
"grad_norm": 0.6673656702041626,
"learning_rate": 3.6442903342750084e-05,
"loss": 0.8063,
"step": 1640
},
{
"epoch": 0.4163775156141568,
"grad_norm": 0.6990562081336975,
"learning_rate": 3.624658604685443e-05,
"loss": 0.8335,
"step": 1650
},
{
"epoch": 0.41890101570878807,
"grad_norm": 0.7190445065498352,
"learning_rate": 3.604939538034158e-05,
"loss": 0.8509,
"step": 1660
},
{
"epoch": 0.42142451580341933,
"grad_norm": 0.7450734376907349,
"learning_rate": 3.585134665635041e-05,
"loss": 0.8446,
"step": 1670
},
{
"epoch": 0.4239480158980506,
"grad_norm": 0.6475887298583984,
"learning_rate": 3.565245525465355e-05,
"loss": 0.8836,
"step": 1680
},
{
"epoch": 0.42647151599268185,
"grad_norm": 0.6419990658760071,
"learning_rate": 3.5452736620463064e-05,
"loss": 0.8428,
"step": 1690
},
{
"epoch": 0.4289950160873131,
"grad_norm": 0.7428763508796692,
"learning_rate": 3.525220626323097e-05,
"loss": 0.8247,
"step": 1700
},
{
"epoch": 0.43151851618194437,
"grad_norm": 0.6717978119850159,
"learning_rate": 3.5050879755444877e-05,
"loss": 0.881,
"step": 1710
},
{
"epoch": 0.4340420162765756,
"grad_norm": 0.6862205862998962,
"learning_rate": 3.484877273141866e-05,
"loss": 0.8511,
"step": 1720
},
{
"epoch": 0.4365655163712069,
"grad_norm": 0.6874988079071045,
"learning_rate": 3.464590088607839e-05,
"loss": 0.8649,
"step": 1730
},
{
"epoch": 0.43908901646583814,
"grad_norm": 0.6635965704917908,
"learning_rate": 3.444227997374345e-05,
"loss": 0.8719,
"step": 1740
},
{
"epoch": 0.44161251656046935,
"grad_norm": 0.7285788655281067,
"learning_rate": 3.4237925806903184e-05,
"loss": 0.8534,
"step": 1750
},
{
"epoch": 0.4441360166551006,
"grad_norm": 0.6177170872688293,
"learning_rate": 3.403285425498889e-05,
"loss": 0.8516,
"step": 1760
},
{
"epoch": 0.44665951674973187,
"grad_norm": 0.7633406519889832,
"learning_rate": 3.3827081243141534e-05,
"loss": 0.8193,
"step": 1770
},
{
"epoch": 0.4491830168443631,
"grad_norm": 0.6661052107810974,
"learning_rate": 3.362062275097496e-05,
"loss": 0.8745,
"step": 1780
},
{
"epoch": 0.4517065169389944,
"grad_norm": 0.7744668126106262,
"learning_rate": 3.341349481133507e-05,
"loss": 0.8158,
"step": 1790
},
{
"epoch": 0.45423001703362564,
"grad_norm": 0.6634140014648438,
"learning_rate": 3.320571350905466e-05,
"loss": 0.8574,
"step": 1800
},
{
"epoch": 0.4567535171282569,
"grad_norm": 0.7289906740188599,
"learning_rate": 3.299729497970444e-05,
"loss": 0.8776,
"step": 1810
},
{
"epoch": 0.45927701722288816,
"grad_norm": 0.6595107913017273,
"learning_rate": 3.278825540833995e-05,
"loss": 0.8416,
"step": 1820
},
{
"epoch": 0.4618005173175194,
"grad_norm": 0.6596432328224182,
"learning_rate": 3.2578611028244656e-05,
"loss": 0.8295,
"step": 1830
},
{
"epoch": 0.4643240174121507,
"grad_norm": 0.7007511258125305,
"learning_rate": 3.2368378119669363e-05,
"loss": 0.8075,
"step": 1840
},
{
"epoch": 0.4668475175067819,
"grad_norm": 0.5890100598335266,
"learning_rate": 3.215757300856796e-05,
"loss": 0.8331,
"step": 1850
},
{
"epoch": 0.46937101760141314,
"grad_norm": 0.670438826084137,
"learning_rate": 3.194621206532957e-05,
"loss": 0.8739,
"step": 1860
},
{
"epoch": 0.4718945176960444,
"grad_norm": 0.6237263083457947,
"learning_rate": 3.173431170350732e-05,
"loss": 0.8377,
"step": 1870
},
{
"epoch": 0.47441801779067566,
"grad_norm": 0.7160887122154236,
"learning_rate": 3.152188837854369e-05,
"loss": 0.8708,
"step": 1880
},
{
"epoch": 0.4769415178853069,
"grad_norm": 0.6525737643241882,
"learning_rate": 3.130895858649264e-05,
"loss": 0.8207,
"step": 1890
},
{
"epoch": 0.4794650179799382,
"grad_norm": 0.7249549627304077,
"learning_rate": 3.109553886273863e-05,
"loss": 0.8516,
"step": 1900
},
{
"epoch": 0.48198851807456944,
"grad_norm": 0.6668533682823181,
"learning_rate": 3.088164578071246e-05,
"loss": 0.8275,
"step": 1910
},
{
"epoch": 0.4845120181692007,
"grad_norm": 0.7262100577354431,
"learning_rate": 3.066729595060431e-05,
"loss": 0.8147,
"step": 1920
},
{
"epoch": 0.48703551826383196,
"grad_norm": 0.7166665196418762,
"learning_rate": 3.0452506018073833e-05,
"loss": 0.8514,
"step": 1930
},
{
"epoch": 0.4895590183584632,
"grad_norm": 0.6810010075569153,
"learning_rate": 3.0237292662957473e-05,
"loss": 0.8323,
"step": 1940
},
{
"epoch": 0.4920825184530944,
"grad_norm": 0.6473044157028198,
"learning_rate": 3.0021672597973207e-05,
"loss": 0.8265,
"step": 1950
},
{
"epoch": 0.4946060185477257,
"grad_norm": 0.6784878969192505,
"learning_rate": 2.9805662567422676e-05,
"loss": 0.8636,
"step": 1960
},
{
"epoch": 0.49712951864235694,
"grad_norm": 0.7378344535827637,
"learning_rate": 2.9589279345890895e-05,
"loss": 0.8483,
"step": 1970
},
{
"epoch": 0.4996530187369882,
"grad_norm": 0.5715174078941345,
"learning_rate": 2.9372539736943577e-05,
"loss": 0.8434,
"step": 1980
},
{
"epoch": 0.5021765188316195,
"grad_norm": 0.5842220783233643,
"learning_rate": 2.9155460571822245e-05,
"loss": 0.8305,
"step": 1990
},
{
"epoch": 0.5047000189262507,
"grad_norm": 0.7206842303276062,
"learning_rate": 2.893805870813717e-05,
"loss": 0.8127,
"step": 2000
},
{
"epoch": 0.5072235190208819,
"grad_norm": 0.6641551852226257,
"learning_rate": 2.872035102855826e-05,
"loss": 0.8272,
"step": 2010
},
{
"epoch": 0.5097470191155132,
"grad_norm": 0.6917135119438171,
"learning_rate": 2.850235443950402e-05,
"loss": 0.7998,
"step": 2020
},
{
"epoch": 0.5122705192101444,
"grad_norm": 0.6133066415786743,
"learning_rate": 2.8284085869828665e-05,
"loss": 0.8413,
"step": 2030
},
{
"epoch": 0.5147940193047758,
"grad_norm": 0.6827579140663147,
"learning_rate": 2.8065562269507463e-05,
"loss": 0.8452,
"step": 2040
},
{
"epoch": 0.517317519399407,
"grad_norm": 0.7090153694152832,
"learning_rate": 2.7846800608320485e-05,
"loss": 0.8293,
"step": 2050
},
{
"epoch": 0.5198410194940383,
"grad_norm": 0.6256769299507141,
"learning_rate": 2.7627817874534762e-05,
"loss": 0.8159,
"step": 2060
},
{
"epoch": 0.5223645195886695,
"grad_norm": 0.6957070231437683,
"learning_rate": 2.7408631073585068e-05,
"loss": 0.8023,
"step": 2070
},
{
"epoch": 0.5248880196833008,
"grad_norm": 0.6817536950111389,
"learning_rate": 2.7189257226753305e-05,
"loss": 0.8334,
"step": 2080
},
{
"epoch": 0.527411519777932,
"grad_norm": 0.6535147428512573,
"learning_rate": 2.696971336984672e-05,
"loss": 0.8558,
"step": 2090
},
{
"epoch": 0.5299350198725632,
"grad_norm": 0.7457418441772461,
"learning_rate": 2.6750016551874945e-05,
"loss": 0.8244,
"step": 2100
},
{
"epoch": 0.5324585199671945,
"grad_norm": 0.6570724248886108,
"learning_rate": 2.6530183833726025e-05,
"loss": 0.8283,
"step": 2110
},
{
"epoch": 0.5349820200618257,
"grad_norm": 0.7065024375915527,
"learning_rate": 2.6310232286841546e-05,
"loss": 0.8565,
"step": 2120
},
{
"epoch": 0.537505520156457,
"grad_norm": 0.671667218208313,
"learning_rate": 2.609017899189092e-05,
"loss": 0.8447,
"step": 2130
},
{
"epoch": 0.5400290202510882,
"grad_norm": 0.6672875285148621,
"learning_rate": 2.587004103744495e-05,
"loss": 0.7912,
"step": 2140
},
{
"epoch": 0.5425525203457195,
"grad_norm": 0.6282544732093811,
"learning_rate": 2.564983551864882e-05,
"loss": 0.8079,
"step": 2150
},
{
"epoch": 0.5450760204403508,
"grad_norm": 0.7435926795005798,
"learning_rate": 2.54295795358945e-05,
"loss": 0.8342,
"step": 2160
},
{
"epoch": 0.5475995205349821,
"grad_norm": 0.6785821318626404,
"learning_rate": 2.5209290193492834e-05,
"loss": 0.8281,
"step": 2170
},
{
"epoch": 0.5501230206296133,
"grad_norm": 0.6483226418495178,
"learning_rate": 2.4988984598345247e-05,
"loss": 0.79,
"step": 2180
},
{
"epoch": 0.5526465207242445,
"grad_norm": 0.6465590000152588,
"learning_rate": 2.4768679858615304e-05,
"loss": 0.841,
"step": 2190
},
{
"epoch": 0.5551700208188758,
"grad_norm": 0.7468442916870117,
"learning_rate": 2.454839308240014e-05,
"loss": 0.8717,
"step": 2200
},
{
"epoch": 0.557693520913507,
"grad_norm": 0.6535473465919495,
"learning_rate": 2.4328141376401903e-05,
"loss": 0.826,
"step": 2210
},
{
"epoch": 0.5602170210081383,
"grad_norm": 0.6404563188552856,
"learning_rate": 2.4107941844599312e-05,
"loss": 0.8062,
"step": 2220
},
{
"epoch": 0.5627405211027695,
"grad_norm": 0.6602795720100403,
"learning_rate": 2.3887811586919424e-05,
"loss": 0.8418,
"step": 2230
},
{
"epoch": 0.5652640211974008,
"grad_norm": 0.6988357305526733,
"learning_rate": 2.3667767697909694e-05,
"loss": 0.8177,
"step": 2240
},
{
"epoch": 0.567787521292032,
"grad_norm": 0.6755298376083374,
"learning_rate": 2.3447827265410517e-05,
"loss": 0.8653,
"step": 2250
},
{
"epoch": 0.5703110213866633,
"grad_norm": 0.72756028175354,
"learning_rate": 2.3228007369228178e-05,
"loss": 0.8896,
"step": 2260
},
{
"epoch": 0.5728345214812945,
"grad_norm": 0.6584864854812622,
"learning_rate": 2.3008325079808576e-05,
"loss": 0.8393,
"step": 2270
},
{
"epoch": 0.5753580215759259,
"grad_norm": 0.6699262857437134,
"learning_rate": 2.2788797456911503e-05,
"loss": 0.7976,
"step": 2280
},
{
"epoch": 0.5778815216705571,
"grad_norm": 0.7463390827178955,
"learning_rate": 2.2569441548285934e-05,
"loss": 0.8321,
"step": 2290
},
{
"epoch": 0.5804050217651883,
"grad_norm": 0.542870283126831,
"learning_rate": 2.2350274388346064e-05,
"loss": 0.786,
"step": 2300
},
{
"epoch": 0.5829285218598196,
"grad_norm": 0.652056872844696,
"learning_rate": 2.213131299684858e-05,
"loss": 0.848,
"step": 2310
},
{
"epoch": 0.5854520219544508,
"grad_norm": 0.7307469248771667,
"learning_rate": 2.191257437757086e-05,
"loss": 0.8117,
"step": 2320
},
{
"epoch": 0.5879755220490821,
"grad_norm": 0.6336262822151184,
"learning_rate": 2.16940755169906e-05,
"loss": 0.8417,
"step": 2330
},
{
"epoch": 0.5904990221437133,
"grad_norm": 0.7636166214942932,
"learning_rate": 2.1475833382966647e-05,
"loss": 0.8786,
"step": 2340
},
{
"epoch": 0.5930225222383446,
"grad_norm": 0.6622100472450256,
"learning_rate": 2.1257864923421404e-05,
"loss": 0.8629,
"step": 2350
},
{
"epoch": 0.5955460223329758,
"grad_norm": 0.602483332157135,
"learning_rate": 2.1040187065024605e-05,
"loss": 0.7786,
"step": 2360
},
{
"epoch": 0.5980695224276071,
"grad_norm": 0.6503065824508667,
"learning_rate": 2.0822816711878978e-05,
"loss": 0.8445,
"step": 2370
},
{
"epoch": 0.6005930225222383,
"grad_norm": 0.6901794672012329,
"learning_rate": 2.0605770744207413e-05,
"loss": 0.8259,
"step": 2380
},
{
"epoch": 0.6031165226168695,
"grad_norm": 0.7173271179199219,
"learning_rate": 2.0389066017042192e-05,
"loss": 0.802,
"step": 2390
},
{
"epoch": 0.6056400227115009,
"grad_norm": 0.7431663870811462,
"learning_rate": 2.0172719358916042e-05,
"loss": 0.8092,
"step": 2400
},
{
"epoch": 0.6081635228061321,
"grad_norm": 0.7227687239646912,
"learning_rate": 1.9956747570555288e-05,
"loss": 0.8563,
"step": 2410
},
{
"epoch": 0.6106870229007634,
"grad_norm": 0.6300061345100403,
"learning_rate": 1.9741167423575186e-05,
"loss": 0.7849,
"step": 2420
},
{
"epoch": 0.6132105229953946,
"grad_norm": 0.6208367347717285,
"learning_rate": 1.9525995659177484e-05,
"loss": 0.8239,
"step": 2430
},
{
"epoch": 0.6157340230900259,
"grad_norm": 0.6272019147872925,
"learning_rate": 1.9311248986850365e-05,
"loss": 0.8102,
"step": 2440
},
{
"epoch": 0.6182575231846571,
"grad_norm": 0.6594968438148499,
"learning_rate": 1.9096944083070866e-05,
"loss": 0.8266,
"step": 2450
},
{
"epoch": 0.6207810232792884,
"grad_norm": 0.673553466796875,
"learning_rate": 1.8883097590009775e-05,
"loss": 0.8375,
"step": 2460
},
{
"epoch": 0.6233045233739196,
"grad_norm": 0.7199084162712097,
"learning_rate": 1.866972611423936e-05,
"loss": 0.8188,
"step": 2470
},
{
"epoch": 0.6258280234685509,
"grad_norm": 0.697413444519043,
"learning_rate": 1.8456846225443648e-05,
"loss": 0.7709,
"step": 2480
},
{
"epoch": 0.6283515235631821,
"grad_norm": 0.6711037158966064,
"learning_rate": 1.8244474455131792e-05,
"loss": 0.8156,
"step": 2490
},
{
"epoch": 0.6308750236578133,
"grad_norm": 0.7030087113380432,
"learning_rate": 1.8032627295354183e-05,
"loss": 0.8125,
"step": 2500
},
{
"epoch": 0.6333985237524447,
"grad_norm": 0.7960418462753296,
"learning_rate": 1.7821321197421837e-05,
"loss": 0.8604,
"step": 2510
},
{
"epoch": 0.6359220238470759,
"grad_norm": 0.6948102116584778,
"learning_rate": 1.761057257062876e-05,
"loss": 0.8301,
"step": 2520
},
{
"epoch": 0.6384455239417072,
"grad_norm": 0.5919877290725708,
"learning_rate": 1.740039778097772e-05,
"loss": 0.7821,
"step": 2530
},
{
"epoch": 0.6409690240363384,
"grad_norm": 0.6569110751152039,
"learning_rate": 1.7190813149909274e-05,
"loss": 0.8213,
"step": 2540
},
{
"epoch": 0.6434925241309697,
"grad_norm": 0.677099347114563,
"learning_rate": 1.6981834953034344e-05,
"loss": 0.8278,
"step": 2550
},
{
"epoch": 0.6460160242256009,
"grad_norm": 0.7233052253723145,
"learning_rate": 1.677347941887028e-05,
"loss": 0.7919,
"step": 2560
},
{
"epoch": 0.6485395243202322,
"grad_norm": 0.7088631987571716,
"learning_rate": 1.656576272758061e-05,
"loss": 0.8444,
"step": 2570
},
{
"epoch": 0.6510630244148634,
"grad_norm": 0.6909515857696533,
"learning_rate": 1.6358701009718577e-05,
"loss": 0.8222,
"step": 2580
},
{
"epoch": 0.6535865245094946,
"grad_norm": 0.5979318618774414,
"learning_rate": 1.615231034497444e-05,
"loss": 0.8376,
"step": 2590
},
{
"epoch": 0.6561100246041259,
"grad_norm": 0.7273426055908203,
"learning_rate": 1.5946606760926865e-05,
"loss": 0.8037,
"step": 2600
},
{
"epoch": 0.6586335246987571,
"grad_norm": 0.719450056552887,
"learning_rate": 1.574160623179816e-05,
"loss": 0.8268,
"step": 2610
},
{
"epoch": 0.6611570247933884,
"grad_norm": 0.7163055539131165,
"learning_rate": 1.553732467721392e-05,
"loss": 0.7853,
"step": 2620
},
{
"epoch": 0.6636805248880197,
"grad_norm": 0.6172025799751282,
"learning_rate": 1.5333777960966616e-05,
"loss": 0.7926,
"step": 2630
},
{
"epoch": 0.666204024982651,
"grad_norm": 0.6272744536399841,
"learning_rate": 1.5130981889783795e-05,
"loss": 0.7982,
"step": 2640
},
{
"epoch": 0.6687275250772822,
"grad_norm": 0.680596649646759,
"learning_rate": 1.4928952212100483e-05,
"loss": 0.8312,
"step": 2650
},
{
"epoch": 0.6712510251719135,
"grad_norm": 0.6080834865570068,
"learning_rate": 1.4727704616836296e-05,
"loss": 0.8273,
"step": 2660
},
{
"epoch": 0.6737745252665447,
"grad_norm": 0.6613759398460388,
"learning_rate": 1.4527254732177043e-05,
"loss": 0.8141,
"step": 2670
},
{
"epoch": 0.676298025361176,
"grad_norm": 0.6180728077888489,
"learning_rate": 1.4327618124361114e-05,
"loss": 0.8231,
"step": 2680
},
{
"epoch": 0.6788215254558072,
"grad_norm": 0.6406080722808838,
"learning_rate": 1.412881029647065e-05,
"loss": 0.7876,
"step": 2690
},
{
"epoch": 0.6813450255504384,
"grad_norm": 0.6109746098518372,
"learning_rate": 1.3930846687227664e-05,
"loss": 0.7957,
"step": 2700
},
{
"epoch": 0.6838685256450697,
"grad_norm": 0.6827517747879028,
"learning_rate": 1.3733742669795049e-05,
"loss": 0.8428,
"step": 2710
},
{
"epoch": 0.6863920257397009,
"grad_norm": 0.7277110815048218,
"learning_rate": 1.3537513550582853e-05,
"loss": 0.8326,
"step": 2720
},
{
"epoch": 0.6889155258343322,
"grad_norm": 0.597568154335022,
"learning_rate": 1.3342174568059527e-05,
"loss": 0.7998,
"step": 2730
},
{
"epoch": 0.6914390259289634,
"grad_norm": 0.6378962993621826,
"learning_rate": 1.3147740891568661e-05,
"loss": 0.785,
"step": 2740
},
{
"epoch": 0.6939625260235948,
"grad_norm": 0.6579405069351196,
"learning_rate": 1.2954227620150904e-05,
"loss": 0.8332,
"step": 2750
},
{
"epoch": 0.696486026118226,
"grad_norm": 0.6977427005767822,
"learning_rate": 1.2761649781371479e-05,
"loss": 0.8095,
"step": 2760
},
{
"epoch": 0.6990095262128573,
"grad_norm": 0.6410185098648071,
"learning_rate": 1.257002233015318e-05,
"loss": 0.8341,
"step": 2770
},
{
"epoch": 0.7015330263074885,
"grad_norm": 0.6869609355926514,
"learning_rate": 1.2379360147614994e-05,
"loss": 0.8023,
"step": 2780
},
{
"epoch": 0.7040565264021197,
"grad_norm": 0.6658973097801208,
"learning_rate": 1.2189678039916532e-05,
"loss": 0.7755,
"step": 2790
},
{
"epoch": 0.706580026496751,
"grad_norm": 0.6188139915466309,
"learning_rate": 1.2000990737108225e-05,
"loss": 0.796,
"step": 2800
},
{
"epoch": 0.7091035265913822,
"grad_norm": 0.7432144284248352,
"learning_rate": 1.1813312891987392e-05,
"loss": 0.8381,
"step": 2810
},
{
"epoch": 0.7116270266860135,
"grad_norm": 0.6776263117790222,
"learning_rate": 1.1626659078960424e-05,
"loss": 0.8087,
"step": 2820
},
{
"epoch": 0.7141505267806447,
"grad_norm": 0.6468738913536072,
"learning_rate": 1.1441043792910936e-05,
"loss": 0.8032,
"step": 2830
},
{
"epoch": 0.716674026875276,
"grad_norm": 0.7177358865737915,
"learning_rate": 1.1256481448074179e-05,
"loss": 0.8039,
"step": 2840
},
{
"epoch": 0.7191975269699072,
"grad_norm": 0.6401441693305969,
"learning_rate": 1.1072986376917638e-05,
"loss": 0.8135,
"step": 2850
},
{
"epoch": 0.7217210270645386,
"grad_norm": 0.6511224508285522,
"learning_rate": 1.0890572829028087e-05,
"loss": 0.8496,
"step": 2860
},
{
"epoch": 0.7242445271591698,
"grad_norm": 0.632625162601471,
"learning_rate": 1.0709254970004937e-05,
"loss": 0.7964,
"step": 2870
},
{
"epoch": 0.7267680272538011,
"grad_norm": 0.5535660982131958,
"learning_rate": 1.0529046880360263e-05,
"loss": 0.7932,
"step": 2880
},
{
"epoch": 0.7292915273484323,
"grad_norm": 0.5996463298797607,
"learning_rate": 1.034996255442529e-05,
"loss": 0.8437,
"step": 2890
},
{
"epoch": 0.7318150274430635,
"grad_norm": 0.6257640719413757,
"learning_rate": 1.0172015899263712e-05,
"loss": 0.8069,
"step": 2900
},
{
"epoch": 0.7343385275376948,
"grad_norm": 0.6533858776092529,
"learning_rate": 9.995220733591639e-06,
"loss": 0.7921,
"step": 2910
},
{
"epoch": 0.736862027632326,
"grad_norm": 0.6002010107040405,
"learning_rate": 9.819590786704572e-06,
"loss": 0.8307,
"step": 2920
},
{
"epoch": 0.7393855277269573,
"grad_norm": 0.6418666243553162,
"learning_rate": 9.645139697411149e-06,
"loss": 0.8036,
"step": 2930
},
{
"epoch": 0.7419090278215885,
"grad_norm": 0.6554102897644043,
"learning_rate": 9.471881012974071e-06,
"loss": 0.8285,
"step": 2940
},
{
"epoch": 0.7444325279162198,
"grad_norm": 0.6879960894584656,
"learning_rate": 9.299828188058013e-06,
"loss": 0.8154,
"step": 2950
},
{
"epoch": 0.746956028010851,
"grad_norm": 0.6418633460998535,
"learning_rate": 9.128994583684838e-06,
"loss": 0.7945,
"step": 2960
},
{
"epoch": 0.7494795281054824,
"grad_norm": 0.6467211246490479,
"learning_rate": 8.959393466195972e-06,
"loss": 0.8464,
"step": 2970
},
{
"epoch": 0.7520030282001136,
"grad_norm": 0.6477042436599731,
"learning_rate": 8.791038006222233e-06,
"loss": 0.8235,
"step": 2980
},
{
"epoch": 0.7545265282947448,
"grad_norm": 0.6426742672920227,
"learning_rate": 8.623941277660994e-06,
"loss": 0.8001,
"step": 2990
},
{
"epoch": 0.7570500283893761,
"grad_norm": 0.7026243805885315,
"learning_rate": 8.458116256660981e-06,
"loss": 0.842,
"step": 3000
},
{
"epoch": 0.7595735284840073,
"grad_norm": 0.6429437398910522,
"learning_rate": 8.293575820614508e-06,
"loss": 0.8143,
"step": 3010
},
{
"epoch": 0.7620970285786386,
"grad_norm": 0.654498815536499,
"learning_rate": 8.130332747157542e-06,
"loss": 0.7697,
"step": 3020
},
{
"epoch": 0.7646205286732698,
"grad_norm": 0.8270076513290405,
"learning_rate": 7.968399713177366e-06,
"loss": 0.825,
"step": 3030
},
{
"epoch": 0.7671440287679011,
"grad_norm": 0.6423079967498779,
"learning_rate": 7.807789293828204e-06,
"loss": 0.8366,
"step": 3040
},
{
"epoch": 0.7696675288625323,
"grad_norm": 0.662451446056366,
"learning_rate": 7.648513961554607e-06,
"loss": 0.7695,
"step": 3050
},
{
"epoch": 0.7721910289571636,
"grad_norm": 0.5953843593597412,
"learning_rate": 7.4905860851229605e-06,
"loss": 0.8296,
"step": 3060
},
{
"epoch": 0.7747145290517948,
"grad_norm": 0.7210749387741089,
"learning_rate": 7.334017928660902e-06,
"loss": 0.8201,
"step": 3070
},
{
"epoch": 0.7772380291464261,
"grad_norm": 0.6214151382446289,
"learning_rate": 7.1788216507049865e-06,
"loss": 0.8034,
"step": 3080
},
{
"epoch": 0.7797615292410573,
"grad_norm": 0.6791695356369019,
"learning_rate": 7.0250093032564494e-06,
"loss": 0.7624,
"step": 3090
},
{
"epoch": 0.7822850293356886,
"grad_norm": 0.6388612985610962,
"learning_rate": 6.872592830845339e-06,
"loss": 0.8004,
"step": 3100
},
{
"epoch": 0.7848085294303199,
"grad_norm": 0.5958021283149719,
"learning_rate": 6.72158406960289e-06,
"loss": 0.8275,
"step": 3110
},
{
"epoch": 0.7873320295249511,
"grad_norm": 0.572040855884552,
"learning_rate": 6.571994746342439e-06,
"loss": 0.8078,
"step": 3120
},
{
"epoch": 0.7898555296195824,
"grad_norm": 0.6328415274620056,
"learning_rate": 6.4238364776486785e-06,
"loss": 0.7883,
"step": 3130
},
{
"epoch": 0.7923790297142136,
"grad_norm": 0.6552072763442993,
"learning_rate": 6.277120768975644e-06,
"loss": 0.8398,
"step": 3140
},
{
"epoch": 0.7949025298088449,
"grad_norm": 0.7182049751281738,
"learning_rate": 6.131859013753155e-06,
"loss": 0.7919,
"step": 3150
},
{
"epoch": 0.7974260299034761,
"grad_norm": 0.7126038074493408,
"learning_rate": 5.988062492502117e-06,
"loss": 0.7782,
"step": 3160
},
{
"epoch": 0.7999495299981074,
"grad_norm": 0.6005820631980896,
"learning_rate": 5.8457423719584435e-06,
"loss": 0.7979,
"step": 3170
},
{
"epoch": 0.8024730300927386,
"grad_norm": 0.6624283790588379,
"learning_rate": 5.704909704205949e-06,
"loss": 0.8297,
"step": 3180
},
{
"epoch": 0.8049965301873698,
"grad_norm": 0.6289507150650024,
"learning_rate": 5.565575425818054e-06,
"loss": 0.8147,
"step": 3190
},
{
"epoch": 0.8075200302820011,
"grad_norm": 0.6975149512290955,
"learning_rate": 5.427750357008468e-06,
"loss": 0.7733,
"step": 3200
},
{
"epoch": 0.8100435303766323,
"grad_norm": 0.6802620887756348,
"learning_rate": 5.291445200790982e-06,
"loss": 0.8226,
"step": 3210
},
{
"epoch": 0.8125670304712637,
"grad_norm": 0.6158818602561951,
"learning_rate": 5.156670542148267e-06,
"loss": 0.8282,
"step": 3220
},
{
"epoch": 0.8150905305658949,
"grad_norm": 0.7228125333786011,
"learning_rate": 5.023436847209887e-06,
"loss": 0.816,
"step": 3230
},
{
"epoch": 0.8176140306605262,
"grad_norm": 0.6515725255012512,
"learning_rate": 4.891754462439557e-06,
"loss": 0.775,
"step": 3240
},
{
"epoch": 0.8201375307551574,
"grad_norm": 0.6829689741134644,
"learning_rate": 4.761633613831645e-06,
"loss": 0.8156,
"step": 3250
},
{
"epoch": 0.8226610308497887,
"grad_norm": 0.7261675596237183,
"learning_rate": 4.6330844061170914e-06,
"loss": 0.7862,
"step": 3260
},
{
"epoch": 0.8251845309444199,
"grad_norm": 0.6911167502403259,
"learning_rate": 4.506116821978662e-06,
"loss": 0.8016,
"step": 3270
},
{
"epoch": 0.8277080310390512,
"grad_norm": 0.5780116319656372,
"learning_rate": 4.380740721275786e-06,
"loss": 0.824,
"step": 3280
},
{
"epoch": 0.8302315311336824,
"grad_norm": 0.6704926490783691,
"learning_rate": 4.25696584027882e-06,
"loss": 0.8037,
"step": 3290
},
{
"epoch": 0.8327550312283136,
"grad_norm": 0.7162071466445923,
"learning_rate": 4.134801790913006e-06,
"loss": 0.7651,
"step": 3300
},
{
"epoch": 0.8352785313229449,
"grad_norm": 0.7350740432739258,
"learning_rate": 4.014258060012005e-06,
"loss": 0.8278,
"step": 3310
},
{
"epoch": 0.8378020314175761,
"grad_norm": 0.6031658053398132,
"learning_rate": 3.895344008581222e-06,
"loss": 0.7945,
"step": 3320
},
{
"epoch": 0.8403255315122075,
"grad_norm": 0.6996452212333679,
"learning_rate": 3.7780688710708223e-06,
"loss": 0.7821,
"step": 3330
},
{
"epoch": 0.8428490316068387,
"grad_norm": 0.6655017733573914,
"learning_rate": 3.6624417546586574e-06,
"loss": 0.7526,
"step": 3340
},
{
"epoch": 0.84537253170147,
"grad_norm": 0.7387165427207947,
"learning_rate": 3.548471638542991e-06,
"loss": 0.8259,
"step": 3350
},
{
"epoch": 0.8478960317961012,
"grad_norm": 0.6410266757011414,
"learning_rate": 3.436167373245247e-06,
"loss": 0.8054,
"step": 3360
},
{
"epoch": 0.8504195318907325,
"grad_norm": 0.6522373557090759,
"learning_rate": 3.325537679922672e-06,
"loss": 0.8168,
"step": 3370
},
{
"epoch": 0.8529430319853637,
"grad_norm": 0.7458412647247314,
"learning_rate": 3.2165911496911173e-06,
"loss": 0.7892,
"step": 3380
},
{
"epoch": 0.8554665320799949,
"grad_norm": 0.6441506743431091,
"learning_rate": 3.1093362429578414e-06,
"loss": 0.8105,
"step": 3390
},
{
"epoch": 0.8579900321746262,
"grad_norm": 0.5970674753189087,
"learning_rate": 3.0037812887645483e-06,
"loss": 0.8326,
"step": 3400
},
{
"epoch": 0.8605135322692574,
"grad_norm": 0.6173757314682007,
"learning_rate": 2.8999344841405373e-06,
"loss": 0.7956,
"step": 3410
},
{
"epoch": 0.8630370323638887,
"grad_norm": 0.6268020868301392,
"learning_rate": 2.7978038934662024e-06,
"loss": 0.7859,
"step": 3420
},
{
"epoch": 0.8655605324585199,
"grad_norm": 0.6534834504127502,
"learning_rate": 2.697397447846725e-06,
"loss": 0.8041,
"step": 3430
},
{
"epoch": 0.8680840325531513,
"grad_norm": 0.6108519434928894,
"learning_rate": 2.5987229444962237e-06,
"loss": 0.823,
"step": 3440
},
{
"epoch": 0.8706075326477825,
"grad_norm": 0.6347935795783997,
"learning_rate": 2.501788046132203e-06,
"loss": 0.831,
"step": 3450
},
{
"epoch": 0.8731310327424138,
"grad_norm": 0.6183903813362122,
"learning_rate": 2.4066002803805386e-06,
"loss": 0.7974,
"step": 3460
},
{
"epoch": 0.875654532837045,
"grad_norm": 0.6723082065582275,
"learning_rate": 2.313167039190861e-06,
"loss": 0.8058,
"step": 3470
},
{
"epoch": 0.8781780329316763,
"grad_norm": 0.6427431702613831,
"learning_rate": 2.2214955782625752e-06,
"loss": 0.805,
"step": 3480
},
{
"epoch": 0.8807015330263075,
"grad_norm": 0.7344009280204773,
"learning_rate": 2.1315930164813507e-06,
"loss": 0.8366,
"step": 3490
},
{
"epoch": 0.8832250331209387,
"grad_norm": 0.6524431109428406,
"learning_rate": 2.0434663353663536e-06,
"loss": 0.8022,
"step": 3500
},
{
"epoch": 0.88574853321557,
"grad_norm": 0.6769471168518066,
"learning_rate": 1.9571223785280314e-06,
"loss": 0.8062,
"step": 3510
},
{
"epoch": 0.8882720333102012,
"grad_norm": 0.6867194771766663,
"learning_rate": 1.8725678511367001e-06,
"loss": 0.8171,
"step": 3520
},
{
"epoch": 0.8907955334048325,
"grad_norm": 0.6660215854644775,
"learning_rate": 1.789809319401825e-06,
"loss": 0.8169,
"step": 3530
},
{
"epoch": 0.8933190334994637,
"grad_norm": 0.6402613520622253,
"learning_rate": 1.7088532100621224e-06,
"loss": 0.7813,
"step": 3540
},
{
"epoch": 0.895842533594095,
"grad_norm": 0.6413708925247192,
"learning_rate": 1.629705809886467e-06,
"loss": 0.7837,
"step": 3550
},
{
"epoch": 0.8983660336887263,
"grad_norm": 0.6048439741134644,
"learning_rate": 1.5523732651857082e-06,
"loss": 0.7984,
"step": 3560
},
{
"epoch": 0.9008895337833576,
"grad_norm": 0.6774916052818298,
"learning_rate": 1.4768615813353398e-06,
"loss": 0.8033,
"step": 3570
},
{
"epoch": 0.9034130338779888,
"grad_norm": 0.6154995560646057,
"learning_rate": 1.4031766223091603e-06,
"loss": 0.8015,
"step": 3580
},
{
"epoch": 0.90593653397262,
"grad_norm": 0.6018934845924377,
"learning_rate": 1.3313241102239054e-06,
"loss": 0.7761,
"step": 3590
},
{
"epoch": 0.9084600340672513,
"grad_norm": 0.658366322517395,
"learning_rate": 1.261309624894863e-06,
"loss": 0.8173,
"step": 3600
},
{
"epoch": 0.9109835341618825,
"grad_norm": 0.6167306900024414,
"learning_rate": 1.1931386034025882e-06,
"loss": 0.8024,
"step": 3610
},
{
"epoch": 0.9135070342565138,
"grad_norm": 0.5509990453720093,
"learning_rate": 1.1268163396706583e-06,
"loss": 0.8128,
"step": 3620
},
{
"epoch": 0.916030534351145,
"grad_norm": 0.6154832243919373,
"learning_rate": 1.0623479840545874e-06,
"loss": 0.7569,
"step": 3630
},
{
"epoch": 0.9185540344457763,
"grad_norm": 0.679389238357544,
"learning_rate": 9.997385429418555e-07,
"loss": 0.8276,
"step": 3640
},
{
"epoch": 0.9210775345404075,
"grad_norm": 0.662276566028595,
"learning_rate": 9.389928783631207e-07,
"loss": 0.8304,
"step": 3650
},
{
"epoch": 0.9236010346350388,
"grad_norm": 0.6233845949172974,
"learning_rate": 8.801157076146705e-07,
"loss": 0.7851,
"step": 3660
},
{
"epoch": 0.92612453472967,
"grad_norm": 0.7036879658699036,
"learning_rate": 8.231116028920765e-07,
"loss": 0.793,
"step": 3670
},
{
"epoch": 0.9286480348243014,
"grad_norm": 0.6103026270866394,
"learning_rate": 7.679849909351472e-07,
"loss": 0.7818,
"step": 3680
},
{
"epoch": 0.9311715349189326,
"grad_norm": 0.6900059580802917,
"learning_rate": 7.147401526841485e-07,
"loss": 0.773,
"step": 3690
},
{
"epoch": 0.9336950350135638,
"grad_norm": 0.681058943271637,
"learning_rate": 6.633812229473791e-07,
"loss": 0.8357,
"step": 3700
},
{
"epoch": 0.9362185351081951,
"grad_norm": 0.7187952995300293,
"learning_rate": 6.139121900800515e-07,
"loss": 0.7779,
"step": 3710
},
{
"epoch": 0.9387420352028263,
"grad_norm": 0.6179840564727783,
"learning_rate": 5.663368956745963e-07,
"loss": 0.7871,
"step": 3720
},
{
"epoch": 0.9412655352974576,
"grad_norm": 0.6663089394569397,
"learning_rate": 5.206590342623164e-07,
"loss": 0.7901,
"step": 3730
},
{
"epoch": 0.9437890353920888,
"grad_norm": 0.6079100370407104,
"learning_rate": 4.768821530264977e-07,
"loss": 0.8226,
"step": 3740
},
{
"epoch": 0.9463125354867201,
"grad_norm": 0.68614262342453,
"learning_rate": 4.350096515269325e-07,
"loss": 0.8185,
"step": 3750
},
{
"epoch": 0.9488360355813513,
"grad_norm": 0.6491347551345825,
"learning_rate": 3.950447814359409e-07,
"loss": 0.817,
"step": 3760
},
{
"epoch": 0.9513595356759826,
"grad_norm": 0.6513685584068298,
"learning_rate": 3.5699064628583745e-07,
"loss": 0.7997,
"step": 3770
},
{
"epoch": 0.9538830357706138,
"grad_norm": 0.6080814003944397,
"learning_rate": 3.2085020122793186e-07,
"loss": 0.7956,
"step": 3780
},
{
"epoch": 0.956406535865245,
"grad_norm": 0.6476254463195801,
"learning_rate": 2.8662625280304613e-07,
"loss": 0.7888,
"step": 3790
},
{
"epoch": 0.9589300359598764,
"grad_norm": 0.6439909934997559,
"learning_rate": 2.5432145872355816e-07,
"loss": 0.7847,
"step": 3800
},
{
"epoch": 0.9614535360545076,
"grad_norm": 0.6744981408119202,
"learning_rate": 2.2393832766701706e-07,
"loss": 0.8093,
"step": 3810
},
{
"epoch": 0.9639770361491389,
"grad_norm": 0.5795860886573792,
"learning_rate": 1.9547921908133483e-07,
"loss": 0.8082,
"step": 3820
},
{
"epoch": 0.9665005362437701,
"grad_norm": 0.6693094968795776,
"learning_rate": 1.689463430015442e-07,
"loss": 0.7857,
"step": 3830
},
{
"epoch": 0.9690240363384014,
"grad_norm": 0.645203173160553,
"learning_rate": 1.443417598781971e-07,
"loss": 0.8056,
"step": 3840
},
{
"epoch": 0.9715475364330326,
"grad_norm": 0.6820341348648071,
"learning_rate": 1.2166738041733684e-07,
"loss": 0.802,
"step": 3850
},
{
"epoch": 0.9740710365276639,
"grad_norm": 0.6292694807052612,
"learning_rate": 1.0092496543212814e-07,
"loss": 0.7937,
"step": 3860
},
{
"epoch": 0.9765945366222951,
"grad_norm": 0.6253132224082947,
"learning_rate": 8.211612570611926e-08,
"loss": 0.7846,
"step": 3870
},
{
"epoch": 0.9791180367169264,
"grad_norm": 0.6571831107139587,
"learning_rate": 6.524232186815305e-08,
"loss": 0.785,
"step": 3880
},
{
"epoch": 0.9816415368115576,
"grad_norm": 0.6356094479560852,
"learning_rate": 5.03048642789411e-08,
"loss": 0.7789,
"step": 3890
},
{
"epoch": 0.9841650369061888,
"grad_norm": 0.8404703140258789,
"learning_rate": 3.730491292930072e-08,
"loss": 0.7954,
"step": 3900
},
{
"epoch": 0.9866885370008202,
"grad_norm": 0.7891058325767517,
"learning_rate": 2.624347735007693e-08,
"loss": 0.8129,
"step": 3910
},
{
"epoch": 0.9892120370954514,
"grad_norm": 0.6858798265457153,
"learning_rate": 1.7121416533749658e-08,
"loss": 0.8076,
"step": 3920
},
{
"epoch": 0.9917355371900827,
"grad_norm": 0.6489024758338928,
"learning_rate": 9.939438867723194e-09,
"loss": 0.8087,
"step": 3930
},
{
"epoch": 0.9942590372847139,
"grad_norm": 0.6204003691673279,
"learning_rate": 4.6981020793118725e-09,
"loss": 0.8162,
"step": 3940
},
{
"epoch": 0.9967825373793452,
"grad_norm": 0.6356140971183777,
"learning_rate": 1.3978131924385906e-09,
"loss": 0.7862,
"step": 3950
},
{
"epoch": 0.9993060374739764,
"grad_norm": 0.6472454071044922,
"learning_rate": 3.88284960184393e-11,
"loss": 0.8188,
"step": 3960
}
],
"logging_steps": 10,
"max_steps": 3962,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 1000,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 2.0380844918675866e+18,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}