LamRA-Ret-Pretrained / trainer_state.json
code-kunkun's picture
Upload 5 files
8540148 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.0,
"eval_steps": 500,
"global_step": 958,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0020876826722338203,
"grad_norm": 0.8805100452450126,
"learning_rate": 1.3793103448275863e-05,
"loss": 3.4375,
"step": 1
},
{
"epoch": 0.0041753653444676405,
"grad_norm": 0.8962056177448029,
"learning_rate": 2.7586206896551727e-05,
"loss": 3.4531,
"step": 2
},
{
"epoch": 0.006263048016701462,
"grad_norm": 0.8975533488683494,
"learning_rate": 4.1379310344827587e-05,
"loss": 3.375,
"step": 3
},
{
"epoch": 0.008350730688935281,
"grad_norm": 0.8562154080719621,
"learning_rate": 5.517241379310345e-05,
"loss": 3.0312,
"step": 4
},
{
"epoch": 0.010438413361169102,
"grad_norm": 0.7722844092130382,
"learning_rate": 6.896551724137931e-05,
"loss": 2.0938,
"step": 5
},
{
"epoch": 0.012526096033402923,
"grad_norm": 0.5469423252625637,
"learning_rate": 8.275862068965517e-05,
"loss": 1.5391,
"step": 6
},
{
"epoch": 0.014613778705636743,
"grad_norm": 0.269527877565377,
"learning_rate": 9.655172413793105e-05,
"loss": 1.3281,
"step": 7
},
{
"epoch": 0.016701461377870562,
"grad_norm": 0.2129561647095634,
"learning_rate": 0.0001103448275862069,
"loss": 1.0547,
"step": 8
},
{
"epoch": 0.018789144050104383,
"grad_norm": 0.209174835903919,
"learning_rate": 0.00012413793103448277,
"loss": 0.8672,
"step": 9
},
{
"epoch": 0.020876826722338204,
"grad_norm": 0.16723166688980518,
"learning_rate": 0.00013793103448275863,
"loss": 0.7461,
"step": 10
},
{
"epoch": 0.022964509394572025,
"grad_norm": 0.13523071310549678,
"learning_rate": 0.00015172413793103449,
"loss": 0.75,
"step": 11
},
{
"epoch": 0.025052192066805846,
"grad_norm": 0.13783786374844356,
"learning_rate": 0.00016551724137931035,
"loss": 0.7578,
"step": 12
},
{
"epoch": 0.027139874739039668,
"grad_norm": 0.11159117473743674,
"learning_rate": 0.0001793103448275862,
"loss": 0.6914,
"step": 13
},
{
"epoch": 0.029227557411273485,
"grad_norm": 0.10486826556036052,
"learning_rate": 0.0001931034482758621,
"loss": 0.6484,
"step": 14
},
{
"epoch": 0.031315240083507306,
"grad_norm": 0.1012538060823961,
"learning_rate": 0.00020689655172413795,
"loss": 0.6562,
"step": 15
},
{
"epoch": 0.033402922755741124,
"grad_norm": 0.10310591282634196,
"learning_rate": 0.0002206896551724138,
"loss": 0.6602,
"step": 16
},
{
"epoch": 0.03549060542797495,
"grad_norm": 0.07838991965173675,
"learning_rate": 0.00023448275862068965,
"loss": 0.5469,
"step": 17
},
{
"epoch": 0.037578288100208766,
"grad_norm": 0.07804122247550843,
"learning_rate": 0.00024827586206896553,
"loss": 0.543,
"step": 18
},
{
"epoch": 0.03966597077244259,
"grad_norm": 0.08565464666867292,
"learning_rate": 0.00026206896551724137,
"loss": 0.5625,
"step": 19
},
{
"epoch": 0.04175365344467641,
"grad_norm": 0.07879295016594862,
"learning_rate": 0.00027586206896551725,
"loss": 0.4551,
"step": 20
},
{
"epoch": 0.04384133611691023,
"grad_norm": 0.07549627256843491,
"learning_rate": 0.00028965517241379314,
"loss": 0.4863,
"step": 21
},
{
"epoch": 0.04592901878914405,
"grad_norm": 0.07816061161931821,
"learning_rate": 0.00030344827586206897,
"loss": 0.459,
"step": 22
},
{
"epoch": 0.04801670146137787,
"grad_norm": 0.07796400588496231,
"learning_rate": 0.00031724137931034486,
"loss": 0.5156,
"step": 23
},
{
"epoch": 0.05010438413361169,
"grad_norm": 0.06699113066481945,
"learning_rate": 0.0003310344827586207,
"loss": 0.5508,
"step": 24
},
{
"epoch": 0.05219206680584551,
"grad_norm": 0.07261576114010894,
"learning_rate": 0.0003448275862068965,
"loss": 0.5469,
"step": 25
},
{
"epoch": 0.054279749478079335,
"grad_norm": 0.07464032217874282,
"learning_rate": 0.0003586206896551724,
"loss": 0.4805,
"step": 26
},
{
"epoch": 0.05636743215031315,
"grad_norm": 0.06913377612077472,
"learning_rate": 0.0003724137931034483,
"loss": 0.5117,
"step": 27
},
{
"epoch": 0.05845511482254697,
"grad_norm": 0.061903733751750956,
"learning_rate": 0.0003862068965517242,
"loss": 0.4844,
"step": 28
},
{
"epoch": 0.060542797494780795,
"grad_norm": 0.06908123825325603,
"learning_rate": 0.0004,
"loss": 0.4434,
"step": 29
},
{
"epoch": 0.06263048016701461,
"grad_norm": 0.05444037852658766,
"learning_rate": 0.00039999885641643965,
"loss": 0.4336,
"step": 30
},
{
"epoch": 0.06471816283924843,
"grad_norm": 0.06329218054783699,
"learning_rate": 0.00039999542567883625,
"loss": 0.5078,
"step": 31
},
{
"epoch": 0.06680584551148225,
"grad_norm": 0.05617311135086726,
"learning_rate": 0.0003999897078264233,
"loss": 0.3418,
"step": 32
},
{
"epoch": 0.06889352818371608,
"grad_norm": 0.0631466065976358,
"learning_rate": 0.00039998170292458913,
"loss": 0.4512,
"step": 33
},
{
"epoch": 0.0709812108559499,
"grad_norm": 0.05939776609982831,
"learning_rate": 0.0003999714110648765,
"loss": 0.4727,
"step": 34
},
{
"epoch": 0.07306889352818371,
"grad_norm": 0.07015310870913878,
"learning_rate": 0.0003999588323649815,
"loss": 0.4316,
"step": 35
},
{
"epoch": 0.07515657620041753,
"grad_norm": 0.05541574151493628,
"learning_rate": 0.00039994396696875197,
"loss": 0.3691,
"step": 36
},
{
"epoch": 0.07724425887265135,
"grad_norm": 0.04781737127842468,
"learning_rate": 0.0003999268150461862,
"loss": 0.3711,
"step": 37
},
{
"epoch": 0.07933194154488518,
"grad_norm": 0.052081481582784105,
"learning_rate": 0.00039990737679343073,
"loss": 0.3633,
"step": 38
},
{
"epoch": 0.081419624217119,
"grad_norm": 0.04824424451620144,
"learning_rate": 0.0003998856524327782,
"loss": 0.3867,
"step": 39
},
{
"epoch": 0.08350730688935282,
"grad_norm": 0.048331445388276904,
"learning_rate": 0.0003998616422126649,
"loss": 0.4199,
"step": 40
},
{
"epoch": 0.08559498956158663,
"grad_norm": 0.06796823383551855,
"learning_rate": 0.00039983534640766766,
"loss": 0.4629,
"step": 41
},
{
"epoch": 0.08768267223382047,
"grad_norm": 0.04720979824092969,
"learning_rate": 0.0003998067653185011,
"loss": 0.3594,
"step": 42
},
{
"epoch": 0.08977035490605428,
"grad_norm": 0.05374730136315078,
"learning_rate": 0.00039977589927201373,
"loss": 0.4414,
"step": 43
},
{
"epoch": 0.0918580375782881,
"grad_norm": 0.050894908072710146,
"learning_rate": 0.0003997427486211847,
"loss": 0.4043,
"step": 44
},
{
"epoch": 0.09394572025052192,
"grad_norm": 0.04383506588833311,
"learning_rate": 0.0003997073137451194,
"loss": 0.3281,
"step": 45
},
{
"epoch": 0.09603340292275574,
"grad_norm": 0.055686594769476794,
"learning_rate": 0.00039966959504904515,
"loss": 0.4062,
"step": 46
},
{
"epoch": 0.09812108559498957,
"grad_norm": 0.05244693300757331,
"learning_rate": 0.00039962959296430676,
"loss": 0.3652,
"step": 47
},
{
"epoch": 0.10020876826722339,
"grad_norm": 0.05926741920735249,
"learning_rate": 0.0003995873079483616,
"loss": 0.4844,
"step": 48
},
{
"epoch": 0.1022964509394572,
"grad_norm": 0.04551719284385615,
"learning_rate": 0.00039954274048477405,
"loss": 0.3223,
"step": 49
},
{
"epoch": 0.10438413361169102,
"grad_norm": 0.04385484814869444,
"learning_rate": 0.0003994958910832104,
"loss": 0.4004,
"step": 50
},
{
"epoch": 0.10647181628392484,
"grad_norm": 0.04780597607586747,
"learning_rate": 0.0003994467602794327,
"loss": 0.3125,
"step": 51
},
{
"epoch": 0.10855949895615867,
"grad_norm": 0.08970784709982177,
"learning_rate": 0.00039939534863529265,
"loss": 0.3359,
"step": 52
},
{
"epoch": 0.11064718162839249,
"grad_norm": 0.055961833035488864,
"learning_rate": 0.00039934165673872536,
"loss": 0.4023,
"step": 53
},
{
"epoch": 0.1127348643006263,
"grad_norm": 0.051294633986921585,
"learning_rate": 0.00039928568520374263,
"loss": 0.3848,
"step": 54
},
{
"epoch": 0.11482254697286012,
"grad_norm": 0.05439699876212045,
"learning_rate": 0.0003992274346704257,
"loss": 0.4785,
"step": 55
},
{
"epoch": 0.11691022964509394,
"grad_norm": 0.04990096696869215,
"learning_rate": 0.00039916690580491806,
"loss": 0.377,
"step": 56
},
{
"epoch": 0.11899791231732777,
"grad_norm": 0.04880913842131259,
"learning_rate": 0.00039910409929941787,
"loss": 0.3809,
"step": 57
},
{
"epoch": 0.12108559498956159,
"grad_norm": 0.048110591694635114,
"learning_rate": 0.0003990390158721701,
"loss": 0.375,
"step": 58
},
{
"epoch": 0.12317327766179541,
"grad_norm": 0.05425382131888079,
"learning_rate": 0.0003989716562674579,
"loss": 0.4297,
"step": 59
},
{
"epoch": 0.12526096033402923,
"grad_norm": 0.04140427495861459,
"learning_rate": 0.0003989020212555948,
"loss": 0.3164,
"step": 60
},
{
"epoch": 0.12734864300626306,
"grad_norm": 0.055270560900777276,
"learning_rate": 0.00039883011163291545,
"loss": 0.418,
"step": 61
},
{
"epoch": 0.12943632567849686,
"grad_norm": 0.04546442173482473,
"learning_rate": 0.0003987559282217662,
"loss": 0.3125,
"step": 62
},
{
"epoch": 0.1315240083507307,
"grad_norm": 0.05835477454974449,
"learning_rate": 0.00039867947187049654,
"loss": 0.4199,
"step": 63
},
{
"epoch": 0.1336116910229645,
"grad_norm": 0.06512707863408194,
"learning_rate": 0.00039860074345344875,
"loss": 0.4238,
"step": 64
},
{
"epoch": 0.13569937369519833,
"grad_norm": 0.04896390157005381,
"learning_rate": 0.0003985197438709479,
"loss": 0.2598,
"step": 65
},
{
"epoch": 0.13778705636743216,
"grad_norm": 0.048613271081049014,
"learning_rate": 0.00039843647404929207,
"loss": 0.3633,
"step": 66
},
{
"epoch": 0.13987473903966596,
"grad_norm": 0.050233778334495785,
"learning_rate": 0.0003983509349407412,
"loss": 0.3926,
"step": 67
},
{
"epoch": 0.1419624217118998,
"grad_norm": 0.051442044526522436,
"learning_rate": 0.0003982631275235064,
"loss": 0.4531,
"step": 68
},
{
"epoch": 0.1440501043841336,
"grad_norm": 0.04624056126157752,
"learning_rate": 0.00039817305280173893,
"loss": 0.373,
"step": 69
},
{
"epoch": 0.14613778705636743,
"grad_norm": 0.04350984771291989,
"learning_rate": 0.0003980807118055185,
"loss": 0.377,
"step": 70
},
{
"epoch": 0.14822546972860126,
"grad_norm": 0.04395725697208839,
"learning_rate": 0.0003979861055908415,
"loss": 0.3633,
"step": 71
},
{
"epoch": 0.15031315240083507,
"grad_norm": 0.04706297130636851,
"learning_rate": 0.0003978892352396091,
"loss": 0.4199,
"step": 72
},
{
"epoch": 0.1524008350730689,
"grad_norm": 0.044497773696428565,
"learning_rate": 0.00039779010185961473,
"loss": 0.4707,
"step": 73
},
{
"epoch": 0.1544885177453027,
"grad_norm": 0.0411155822363719,
"learning_rate": 0.0003976887065845314,
"loss": 0.3262,
"step": 74
},
{
"epoch": 0.15657620041753653,
"grad_norm": 0.043812847498547546,
"learning_rate": 0.0003975850505738988,
"loss": 0.3359,
"step": 75
},
{
"epoch": 0.15866388308977036,
"grad_norm": 0.0554940446413714,
"learning_rate": 0.0003974791350131101,
"loss": 0.3867,
"step": 76
},
{
"epoch": 0.16075156576200417,
"grad_norm": 0.04898321080003146,
"learning_rate": 0.00039737096111339804,
"loss": 0.4648,
"step": 77
},
{
"epoch": 0.162839248434238,
"grad_norm": 0.04703455829390756,
"learning_rate": 0.00039726053011182176,
"loss": 0.3945,
"step": 78
},
{
"epoch": 0.1649269311064718,
"grad_norm": 0.048291736323549404,
"learning_rate": 0.00039714784327125196,
"loss": 0.3555,
"step": 79
},
{
"epoch": 0.16701461377870563,
"grad_norm": 0.048158996103635526,
"learning_rate": 0.00039703290188035687,
"loss": 0.3848,
"step": 80
},
{
"epoch": 0.16910229645093947,
"grad_norm": 0.041070665599436396,
"learning_rate": 0.00039691570725358726,
"loss": 0.3672,
"step": 81
},
{
"epoch": 0.17118997912317327,
"grad_norm": 0.04514739709583794,
"learning_rate": 0.0003967962607311617,
"loss": 0.416,
"step": 82
},
{
"epoch": 0.1732776617954071,
"grad_norm": 0.05018154319088866,
"learning_rate": 0.0003966745636790509,
"loss": 0.4219,
"step": 83
},
{
"epoch": 0.17536534446764093,
"grad_norm": 0.04615841699661321,
"learning_rate": 0.00039655061748896243,
"loss": 0.3594,
"step": 84
},
{
"epoch": 0.17745302713987474,
"grad_norm": 0.04770173243618234,
"learning_rate": 0.00039642442357832447,
"loss": 0.3789,
"step": 85
},
{
"epoch": 0.17954070981210857,
"grad_norm": 0.0529492248920703,
"learning_rate": 0.00039629598339026984,
"loss": 0.4102,
"step": 86
},
{
"epoch": 0.18162839248434237,
"grad_norm": 0.04108643805111346,
"learning_rate": 0.00039616529839361946,
"loss": 0.3691,
"step": 87
},
{
"epoch": 0.1837160751565762,
"grad_norm": 0.04230844261748932,
"learning_rate": 0.00039603237008286544,
"loss": 0.3672,
"step": 88
},
{
"epoch": 0.18580375782881003,
"grad_norm": 0.057536412046600406,
"learning_rate": 0.00039589719997815413,
"loss": 0.3574,
"step": 89
},
{
"epoch": 0.18789144050104384,
"grad_norm": 0.03943187889859641,
"learning_rate": 0.0003957597896252685,
"loss": 0.332,
"step": 90
},
{
"epoch": 0.18997912317327767,
"grad_norm": 0.04024451449784526,
"learning_rate": 0.000395620140595611,
"loss": 0.3438,
"step": 91
},
{
"epoch": 0.19206680584551147,
"grad_norm": 0.04428176826646225,
"learning_rate": 0.0003954782544861847,
"loss": 0.3574,
"step": 92
},
{
"epoch": 0.1941544885177453,
"grad_norm": 0.045692937481720604,
"learning_rate": 0.00039533413291957596,
"loss": 0.3574,
"step": 93
},
{
"epoch": 0.19624217118997914,
"grad_norm": 0.04688683747789243,
"learning_rate": 0.00039518777754393535,
"loss": 0.3867,
"step": 94
},
{
"epoch": 0.19832985386221294,
"grad_norm": 0.05536707488449957,
"learning_rate": 0.0003950391900329588,
"loss": 0.3223,
"step": 95
},
{
"epoch": 0.20041753653444677,
"grad_norm": 0.04606231136571311,
"learning_rate": 0.0003948883720858687,
"loss": 0.3535,
"step": 96
},
{
"epoch": 0.20250521920668058,
"grad_norm": 0.04099133317121465,
"learning_rate": 0.00039473532542739435,
"loss": 0.3066,
"step": 97
},
{
"epoch": 0.2045929018789144,
"grad_norm": 0.055252229005040635,
"learning_rate": 0.0003945800518077521,
"loss": 0.3809,
"step": 98
},
{
"epoch": 0.20668058455114824,
"grad_norm": 0.04490759353014913,
"learning_rate": 0.00039442255300262553,
"loss": 0.373,
"step": 99
},
{
"epoch": 0.20876826722338204,
"grad_norm": 0.04889359614788617,
"learning_rate": 0.0003942628308131452,
"loss": 0.377,
"step": 100
},
{
"epoch": 0.21085594989561587,
"grad_norm": 0.04331101172153666,
"learning_rate": 0.0003941008870658677,
"loss": 0.3926,
"step": 101
},
{
"epoch": 0.21294363256784968,
"grad_norm": 0.039340556954521134,
"learning_rate": 0.00039393672361275506,
"loss": 0.3379,
"step": 102
},
{
"epoch": 0.2150313152400835,
"grad_norm": 0.047692728640419506,
"learning_rate": 0.0003937703423311537,
"loss": 0.375,
"step": 103
},
{
"epoch": 0.21711899791231734,
"grad_norm": 0.04193899617132502,
"learning_rate": 0.00039360174512377246,
"loss": 0.3203,
"step": 104
},
{
"epoch": 0.21920668058455114,
"grad_norm": 0.04450971768791051,
"learning_rate": 0.0003934309339186614,
"loss": 0.3789,
"step": 105
},
{
"epoch": 0.22129436325678498,
"grad_norm": 0.04610401641341007,
"learning_rate": 0.00039325791066918925,
"loss": 0.4004,
"step": 106
},
{
"epoch": 0.22338204592901878,
"grad_norm": 0.0394316515696973,
"learning_rate": 0.00039308267735402156,
"loss": 0.2734,
"step": 107
},
{
"epoch": 0.2254697286012526,
"grad_norm": 0.04552211598084851,
"learning_rate": 0.00039290523597709765,
"loss": 0.3555,
"step": 108
},
{
"epoch": 0.22755741127348644,
"grad_norm": 0.044355074180719295,
"learning_rate": 0.00039272558856760797,
"loss": 0.3418,
"step": 109
},
{
"epoch": 0.22964509394572025,
"grad_norm": 0.04664869616431499,
"learning_rate": 0.0003925437371799708,
"loss": 0.3301,
"step": 110
},
{
"epoch": 0.23173277661795408,
"grad_norm": 0.04421582409863553,
"learning_rate": 0.0003923596838938086,
"loss": 0.3242,
"step": 111
},
{
"epoch": 0.23382045929018788,
"grad_norm": 0.04914905461390972,
"learning_rate": 0.00039217343081392463,
"loss": 0.3027,
"step": 112
},
{
"epoch": 0.2359081419624217,
"grad_norm": 0.04801403896872833,
"learning_rate": 0.00039198498007027836,
"loss": 0.377,
"step": 113
},
{
"epoch": 0.23799582463465555,
"grad_norm": 0.039909379565673524,
"learning_rate": 0.0003917943338179616,
"loss": 0.2949,
"step": 114
},
{
"epoch": 0.24008350730688935,
"grad_norm": 0.04562132710784955,
"learning_rate": 0.00039160149423717353,
"loss": 0.3828,
"step": 115
},
{
"epoch": 0.24217118997912318,
"grad_norm": 0.04218187984586478,
"learning_rate": 0.00039140646353319584,
"loss": 0.3379,
"step": 116
},
{
"epoch": 0.24425887265135698,
"grad_norm": 0.13995497079449562,
"learning_rate": 0.0003912092439363677,
"loss": 0.3887,
"step": 117
},
{
"epoch": 0.24634655532359082,
"grad_norm": 0.03956892044992931,
"learning_rate": 0.00039100983770205995,
"loss": 0.3027,
"step": 118
},
{
"epoch": 0.24843423799582465,
"grad_norm": 0.043965744196170775,
"learning_rate": 0.00039080824711064946,
"loss": 0.3223,
"step": 119
},
{
"epoch": 0.25052192066805845,
"grad_norm": 0.044644460638659476,
"learning_rate": 0.00039060447446749314,
"loss": 0.3496,
"step": 120
},
{
"epoch": 0.25260960334029225,
"grad_norm": 0.05275646603548191,
"learning_rate": 0.00039039852210290147,
"loss": 0.3438,
"step": 121
},
{
"epoch": 0.2546972860125261,
"grad_norm": 0.04064451365674439,
"learning_rate": 0.00039019039237211175,
"loss": 0.3359,
"step": 122
},
{
"epoch": 0.2567849686847599,
"grad_norm": 0.048323141931080164,
"learning_rate": 0.00038998008765526136,
"loss": 0.377,
"step": 123
},
{
"epoch": 0.2588726513569937,
"grad_norm": 0.03725290930480903,
"learning_rate": 0.00038976761035736064,
"loss": 0.2988,
"step": 124
},
{
"epoch": 0.2609603340292276,
"grad_norm": 0.04052181522164268,
"learning_rate": 0.00038955296290826483,
"loss": 0.3203,
"step": 125
},
{
"epoch": 0.2630480167014614,
"grad_norm": 0.04516920645500193,
"learning_rate": 0.00038933614776264695,
"loss": 0.3496,
"step": 126
},
{
"epoch": 0.2651356993736952,
"grad_norm": 0.04887108820007405,
"learning_rate": 0.00038911716739996945,
"loss": 0.2734,
"step": 127
},
{
"epoch": 0.267223382045929,
"grad_norm": 0.04662438292740599,
"learning_rate": 0.0003888960243244556,
"loss": 0.2949,
"step": 128
},
{
"epoch": 0.26931106471816285,
"grad_norm": 0.0462101738931956,
"learning_rate": 0.0003886727210650613,
"loss": 0.3711,
"step": 129
},
{
"epoch": 0.27139874739039666,
"grad_norm": 0.045514397288015096,
"learning_rate": 0.00038844726017544606,
"loss": 0.3945,
"step": 130
},
{
"epoch": 0.27348643006263046,
"grad_norm": 0.04564176989369504,
"learning_rate": 0.00038821964423394335,
"loss": 0.3262,
"step": 131
},
{
"epoch": 0.2755741127348643,
"grad_norm": 0.0480634271209225,
"learning_rate": 0.00038798987584353175,
"loss": 0.4297,
"step": 132
},
{
"epoch": 0.2776617954070981,
"grad_norm": 0.04064367156377542,
"learning_rate": 0.0003877579576318048,
"loss": 0.3066,
"step": 133
},
{
"epoch": 0.2797494780793319,
"grad_norm": 0.06187073759126634,
"learning_rate": 0.000387523892250941,
"loss": 0.3828,
"step": 134
},
{
"epoch": 0.2818371607515658,
"grad_norm": 0.047235774434653544,
"learning_rate": 0.0003872876823776737,
"loss": 0.4551,
"step": 135
},
{
"epoch": 0.2839248434237996,
"grad_norm": 0.04613794323687757,
"learning_rate": 0.00038704933071326004,
"loss": 0.3555,
"step": 136
},
{
"epoch": 0.2860125260960334,
"grad_norm": 0.04123661418763717,
"learning_rate": 0.00038680883998345046,
"loss": 0.3848,
"step": 137
},
{
"epoch": 0.2881002087682672,
"grad_norm": 0.03990534224519547,
"learning_rate": 0.0003865662129384575,
"loss": 0.3223,
"step": 138
},
{
"epoch": 0.29018789144050106,
"grad_norm": 0.03964108882883799,
"learning_rate": 0.00038632145235292407,
"loss": 0.3145,
"step": 139
},
{
"epoch": 0.29227557411273486,
"grad_norm": 0.04771985690919652,
"learning_rate": 0.00038607456102589217,
"loss": 0.3555,
"step": 140
},
{
"epoch": 0.29436325678496866,
"grad_norm": 0.041459959268485676,
"learning_rate": 0.0003858255417807701,
"loss": 0.3203,
"step": 141
},
{
"epoch": 0.2964509394572025,
"grad_norm": 0.04488781487797441,
"learning_rate": 0.0003855743974653014,
"loss": 0.3516,
"step": 142
},
{
"epoch": 0.2985386221294363,
"grad_norm": 0.045595447329589846,
"learning_rate": 0.00038532113095153086,
"loss": 0.4121,
"step": 143
},
{
"epoch": 0.30062630480167013,
"grad_norm": 0.051789411309846184,
"learning_rate": 0.0003850657451357729,
"loss": 0.3945,
"step": 144
},
{
"epoch": 0.302713987473904,
"grad_norm": 0.04655813617947128,
"learning_rate": 0.00038480824293857756,
"loss": 0.377,
"step": 145
},
{
"epoch": 0.3048016701461378,
"grad_norm": 0.07648712725066147,
"learning_rate": 0.00038454862730469773,
"loss": 0.3438,
"step": 146
},
{
"epoch": 0.3068893528183716,
"grad_norm": 0.05310195833383878,
"learning_rate": 0.0003842869012030551,
"loss": 0.3594,
"step": 147
},
{
"epoch": 0.3089770354906054,
"grad_norm": 0.04391887584861711,
"learning_rate": 0.00038402306762670636,
"loss": 0.3555,
"step": 148
},
{
"epoch": 0.31106471816283926,
"grad_norm": 0.038985625260836355,
"learning_rate": 0.000383757129592809,
"loss": 0.2969,
"step": 149
},
{
"epoch": 0.31315240083507306,
"grad_norm": 0.050635143690619895,
"learning_rate": 0.00038348909014258647,
"loss": 0.3398,
"step": 150
},
{
"epoch": 0.31524008350730687,
"grad_norm": 0.04475955558328409,
"learning_rate": 0.00038321895234129394,
"loss": 0.3594,
"step": 151
},
{
"epoch": 0.3173277661795407,
"grad_norm": 0.046705792421880006,
"learning_rate": 0.0003829467192781829,
"loss": 0.3906,
"step": 152
},
{
"epoch": 0.31941544885177453,
"grad_norm": 0.04135295822439203,
"learning_rate": 0.00038267239406646595,
"loss": 0.3203,
"step": 153
},
{
"epoch": 0.32150313152400833,
"grad_norm": 0.03936648551856702,
"learning_rate": 0.000382395979843281,
"loss": 0.3086,
"step": 154
},
{
"epoch": 0.3235908141962422,
"grad_norm": 0.04411342571947294,
"learning_rate": 0.0003821174797696558,
"loss": 0.3145,
"step": 155
},
{
"epoch": 0.325678496868476,
"grad_norm": 0.0440989009123292,
"learning_rate": 0.0003818368970304713,
"loss": 0.3398,
"step": 156
},
{
"epoch": 0.3277661795407098,
"grad_norm": 0.04623084733338215,
"learning_rate": 0.00038155423483442565,
"loss": 0.373,
"step": 157
},
{
"epoch": 0.3298538622129436,
"grad_norm": 0.03851250195995776,
"learning_rate": 0.00038126949641399725,
"loss": 0.2773,
"step": 158
},
{
"epoch": 0.33194154488517746,
"grad_norm": 0.04376944212115029,
"learning_rate": 0.0003809826850254078,
"loss": 0.3613,
"step": 159
},
{
"epoch": 0.33402922755741127,
"grad_norm": 0.044952970087986426,
"learning_rate": 0.00038069380394858526,
"loss": 0.2734,
"step": 160
},
{
"epoch": 0.33611691022964507,
"grad_norm": 0.04338817418637471,
"learning_rate": 0.0003804028564871262,
"loss": 0.334,
"step": 161
},
{
"epoch": 0.33820459290187893,
"grad_norm": 0.04540609013390802,
"learning_rate": 0.00038010984596825774,
"loss": 0.3086,
"step": 162
},
{
"epoch": 0.34029227557411273,
"grad_norm": 0.040675218084428574,
"learning_rate": 0.00037981477574280015,
"loss": 0.3105,
"step": 163
},
{
"epoch": 0.34237995824634654,
"grad_norm": 0.04628740772476471,
"learning_rate": 0.000379517649185128,
"loss": 0.375,
"step": 164
},
{
"epoch": 0.3444676409185804,
"grad_norm": 0.045320126800110626,
"learning_rate": 0.0003792184696931318,
"loss": 0.3926,
"step": 165
},
{
"epoch": 0.3465553235908142,
"grad_norm": 0.047079880323708634,
"learning_rate": 0.00037891724068817896,
"loss": 0.3613,
"step": 166
},
{
"epoch": 0.348643006263048,
"grad_norm": 0.042006341910680284,
"learning_rate": 0.0003786139656150749,
"loss": 0.3848,
"step": 167
},
{
"epoch": 0.35073068893528186,
"grad_norm": 0.042457959635259945,
"learning_rate": 0.0003783086479420235,
"loss": 0.3594,
"step": 168
},
{
"epoch": 0.35281837160751567,
"grad_norm": 0.04729445116359038,
"learning_rate": 0.0003780012911605875,
"loss": 0.4023,
"step": 169
},
{
"epoch": 0.35490605427974947,
"grad_norm": 0.04490557924422057,
"learning_rate": 0.00037769189878564843,
"loss": 0.3184,
"step": 170
},
{
"epoch": 0.3569937369519833,
"grad_norm": 0.04659805592024877,
"learning_rate": 0.0003773804743553667,
"loss": 0.3613,
"step": 171
},
{
"epoch": 0.35908141962421714,
"grad_norm": 0.0457558395135587,
"learning_rate": 0.00037706702143114085,
"loss": 0.3848,
"step": 172
},
{
"epoch": 0.36116910229645094,
"grad_norm": 0.042225345661087835,
"learning_rate": 0.0003767515435975671,
"loss": 0.3652,
"step": 173
},
{
"epoch": 0.36325678496868474,
"grad_norm": 0.03862535581909019,
"learning_rate": 0.00037643404446239805,
"loss": 0.2578,
"step": 174
},
{
"epoch": 0.3653444676409186,
"grad_norm": 0.04511245438490728,
"learning_rate": 0.0003761145276565015,
"loss": 0.3789,
"step": 175
},
{
"epoch": 0.3674321503131524,
"grad_norm": 0.04418007608895449,
"learning_rate": 0.0003757929968338193,
"loss": 0.375,
"step": 176
},
{
"epoch": 0.3695198329853862,
"grad_norm": 0.04131570246472954,
"learning_rate": 0.00037546945567132486,
"loss": 0.3262,
"step": 177
},
{
"epoch": 0.37160751565762007,
"grad_norm": 0.04432885957928793,
"learning_rate": 0.0003751439078689819,
"loss": 0.3105,
"step": 178
},
{
"epoch": 0.3736951983298539,
"grad_norm": 0.04130281228669527,
"learning_rate": 0.0003748163571497015,
"loss": 0.3145,
"step": 179
},
{
"epoch": 0.3757828810020877,
"grad_norm": 0.041235332477716484,
"learning_rate": 0.00037448680725929987,
"loss": 0.3086,
"step": 180
},
{
"epoch": 0.3778705636743215,
"grad_norm": 0.046591187622845885,
"learning_rate": 0.00037415526196645535,
"loss": 0.377,
"step": 181
},
{
"epoch": 0.37995824634655534,
"grad_norm": 0.043131270347982746,
"learning_rate": 0.00037382172506266544,
"loss": 0.3223,
"step": 182
},
{
"epoch": 0.38204592901878914,
"grad_norm": 0.04113286864493325,
"learning_rate": 0.00037348620036220325,
"loss": 0.3379,
"step": 183
},
{
"epoch": 0.38413361169102295,
"grad_norm": 0.045126082835651934,
"learning_rate": 0.00037314869170207427,
"loss": 0.3906,
"step": 184
},
{
"epoch": 0.3862212943632568,
"grad_norm": 0.0468308767634132,
"learning_rate": 0.00037280920294197186,
"loss": 0.418,
"step": 185
},
{
"epoch": 0.3883089770354906,
"grad_norm": 0.041909756539268275,
"learning_rate": 0.0003724677379642338,
"loss": 0.3008,
"step": 186
},
{
"epoch": 0.3903966597077244,
"grad_norm": 0.039249830774930286,
"learning_rate": 0.0003721243006737973,
"loss": 0.293,
"step": 187
},
{
"epoch": 0.3924843423799583,
"grad_norm": 0.043933299859950536,
"learning_rate": 0.0003717788949981549,
"loss": 0.3281,
"step": 188
},
{
"epoch": 0.3945720250521921,
"grad_norm": 0.045457197762689416,
"learning_rate": 0.0003714315248873091,
"loss": 0.416,
"step": 189
},
{
"epoch": 0.3966597077244259,
"grad_norm": 0.038856886092809004,
"learning_rate": 0.00037108219431372734,
"loss": 0.3242,
"step": 190
},
{
"epoch": 0.3987473903966597,
"grad_norm": 0.03887481602609131,
"learning_rate": 0.0003707309072722966,
"loss": 0.3047,
"step": 191
},
{
"epoch": 0.40083507306889354,
"grad_norm": 0.04338847469385484,
"learning_rate": 0.0003703776677802778,
"loss": 0.3945,
"step": 192
},
{
"epoch": 0.40292275574112735,
"grad_norm": 0.04874209932196092,
"learning_rate": 0.0003700224798772597,
"loss": 0.4316,
"step": 193
},
{
"epoch": 0.40501043841336115,
"grad_norm": 0.0409033493154823,
"learning_rate": 0.00036966534762511276,
"loss": 0.3301,
"step": 194
},
{
"epoch": 0.407098121085595,
"grad_norm": 0.040056909386716535,
"learning_rate": 0.0003693062751079427,
"loss": 0.3887,
"step": 195
},
{
"epoch": 0.4091858037578288,
"grad_norm": 0.03764639519212107,
"learning_rate": 0.0003689452664320437,
"loss": 0.3125,
"step": 196
},
{
"epoch": 0.4112734864300626,
"grad_norm": 0.04012224378405559,
"learning_rate": 0.00036858232572585173,
"loss": 0.3574,
"step": 197
},
{
"epoch": 0.4133611691022965,
"grad_norm": 0.04149192047957985,
"learning_rate": 0.00036821745713989706,
"loss": 0.3242,
"step": 198
},
{
"epoch": 0.4154488517745303,
"grad_norm": 0.05017075921477889,
"learning_rate": 0.00036785066484675683,
"loss": 0.3652,
"step": 199
},
{
"epoch": 0.4175365344467641,
"grad_norm": 0.04227941718590804,
"learning_rate": 0.00036748195304100735,
"loss": 0.3242,
"step": 200
},
{
"epoch": 0.4196242171189979,
"grad_norm": 0.03768968002704521,
"learning_rate": 0.00036711132593917635,
"loss": 0.3105,
"step": 201
},
{
"epoch": 0.42171189979123175,
"grad_norm": 0.04194019808640771,
"learning_rate": 0.00036673878777969426,
"loss": 0.2891,
"step": 202
},
{
"epoch": 0.42379958246346555,
"grad_norm": 0.04542209099244845,
"learning_rate": 0.00036636434282284633,
"loss": 0.3359,
"step": 203
},
{
"epoch": 0.42588726513569936,
"grad_norm": 0.03795072966054866,
"learning_rate": 0.0003659879953507235,
"loss": 0.2178,
"step": 204
},
{
"epoch": 0.4279749478079332,
"grad_norm": 0.0676312462445889,
"learning_rate": 0.0003656097496671737,
"loss": 0.4004,
"step": 205
},
{
"epoch": 0.430062630480167,
"grad_norm": 0.06595187551394353,
"learning_rate": 0.00036522961009775217,
"loss": 0.3359,
"step": 206
},
{
"epoch": 0.4321503131524008,
"grad_norm": 0.048238875202028735,
"learning_rate": 0.0003648475809896727,
"loss": 0.375,
"step": 207
},
{
"epoch": 0.4342379958246347,
"grad_norm": 0.04774287102162698,
"learning_rate": 0.00036446366671175735,
"loss": 0.3613,
"step": 208
},
{
"epoch": 0.4363256784968685,
"grad_norm": 0.04447429752378728,
"learning_rate": 0.0003640778716543866,
"loss": 0.3594,
"step": 209
},
{
"epoch": 0.4384133611691023,
"grad_norm": 0.04069940254145397,
"learning_rate": 0.00036369020022944946,
"loss": 0.3105,
"step": 210
},
{
"epoch": 0.4405010438413361,
"grad_norm": 0.03646181755503713,
"learning_rate": 0.0003633006568702924,
"loss": 0.3047,
"step": 211
},
{
"epoch": 0.44258872651356995,
"grad_norm": 0.041622368664595316,
"learning_rate": 0.00036290924603166945,
"loss": 0.3418,
"step": 212
},
{
"epoch": 0.44467640918580376,
"grad_norm": 0.04687265549200718,
"learning_rate": 0.00036251597218969045,
"loss": 0.3926,
"step": 213
},
{
"epoch": 0.44676409185803756,
"grad_norm": 0.05629699568400707,
"learning_rate": 0.0003621208398417705,
"loss": 0.3848,
"step": 214
},
{
"epoch": 0.4488517745302714,
"grad_norm": 0.04583925885988264,
"learning_rate": 0.0003617238535065782,
"loss": 0.334,
"step": 215
},
{
"epoch": 0.4509394572025052,
"grad_norm": 0.04693126530215214,
"learning_rate": 0.0003613250177239838,
"loss": 0.3691,
"step": 216
},
{
"epoch": 0.453027139874739,
"grad_norm": 0.05342936402981673,
"learning_rate": 0.00036092433705500797,
"loss": 0.3301,
"step": 217
},
{
"epoch": 0.4551148225469729,
"grad_norm": 0.04089550326814837,
"learning_rate": 0.000360521816081769,
"loss": 0.2793,
"step": 218
},
{
"epoch": 0.4572025052192067,
"grad_norm": 0.04725556069406023,
"learning_rate": 0.00036011745940743036,
"loss": 0.3887,
"step": 219
},
{
"epoch": 0.4592901878914405,
"grad_norm": 0.04629637553798221,
"learning_rate": 0.0003597112716561486,
"loss": 0.3477,
"step": 220
},
{
"epoch": 0.4613778705636743,
"grad_norm": 0.042829084723361624,
"learning_rate": 0.0003593032574730201,
"loss": 0.3379,
"step": 221
},
{
"epoch": 0.46346555323590816,
"grad_norm": 0.04290003185482568,
"learning_rate": 0.000358893421524028,
"loss": 0.3828,
"step": 222
},
{
"epoch": 0.46555323590814196,
"grad_norm": 0.03695634767631775,
"learning_rate": 0.0003584817684959888,
"loss": 0.2539,
"step": 223
},
{
"epoch": 0.46764091858037576,
"grad_norm": 0.043249882204696076,
"learning_rate": 0.00035806830309649887,
"loss": 0.291,
"step": 224
},
{
"epoch": 0.4697286012526096,
"grad_norm": 0.0426499927716365,
"learning_rate": 0.0003576530300538805,
"loss": 0.3262,
"step": 225
},
{
"epoch": 0.4718162839248434,
"grad_norm": 0.038269992068262225,
"learning_rate": 0.00035723595411712793,
"loss": 0.2598,
"step": 226
},
{
"epoch": 0.47390396659707723,
"grad_norm": 0.039227521535724244,
"learning_rate": 0.0003568170800558531,
"loss": 0.2812,
"step": 227
},
{
"epoch": 0.4759916492693111,
"grad_norm": 0.0414687007362602,
"learning_rate": 0.0003563964126602308,
"loss": 0.3184,
"step": 228
},
{
"epoch": 0.4780793319415449,
"grad_norm": 0.06563073201350628,
"learning_rate": 0.0003559739567409443,
"loss": 0.2969,
"step": 229
},
{
"epoch": 0.4801670146137787,
"grad_norm": 0.04223079865704053,
"learning_rate": 0.00035554971712913004,
"loss": 0.334,
"step": 230
},
{
"epoch": 0.4822546972860125,
"grad_norm": 0.042708614625140126,
"learning_rate": 0.00035512369867632236,
"loss": 0.3418,
"step": 231
},
{
"epoch": 0.48434237995824636,
"grad_norm": 0.04063103807491919,
"learning_rate": 0.00035469590625439843,
"loss": 0.3477,
"step": 232
},
{
"epoch": 0.48643006263048016,
"grad_norm": 0.045147549230686396,
"learning_rate": 0.00035426634475552187,
"loss": 0.3516,
"step": 233
},
{
"epoch": 0.48851774530271397,
"grad_norm": 0.04106108808167586,
"learning_rate": 0.00035383501909208755,
"loss": 0.2812,
"step": 234
},
{
"epoch": 0.4906054279749478,
"grad_norm": 0.0428728366141268,
"learning_rate": 0.0003534019341966647,
"loss": 0.3398,
"step": 235
},
{
"epoch": 0.49269311064718163,
"grad_norm": 0.047772722106343216,
"learning_rate": 0.00035296709502194115,
"loss": 0.3848,
"step": 236
},
{
"epoch": 0.49478079331941544,
"grad_norm": 0.04629142912114304,
"learning_rate": 0.000352530506540666,
"loss": 0.2715,
"step": 237
},
{
"epoch": 0.4968684759916493,
"grad_norm": 0.05018937450568595,
"learning_rate": 0.0003520921737455935,
"loss": 0.3027,
"step": 238
},
{
"epoch": 0.4989561586638831,
"grad_norm": 0.04431763914166162,
"learning_rate": 0.00035165210164942535,
"loss": 0.3477,
"step": 239
},
{
"epoch": 0.5010438413361169,
"grad_norm": 0.04080914710858152,
"learning_rate": 0.00035121029528475386,
"loss": 0.3086,
"step": 240
},
{
"epoch": 0.5031315240083507,
"grad_norm": 0.04222028074641329,
"learning_rate": 0.0003507667597040038,
"loss": 0.3672,
"step": 241
},
{
"epoch": 0.5052192066805845,
"grad_norm": 0.04367534710869874,
"learning_rate": 0.0003503214999793752,
"loss": 0.3223,
"step": 242
},
{
"epoch": 0.5073068893528184,
"grad_norm": 0.04813953280858226,
"learning_rate": 0.00034987452120278505,
"loss": 0.3984,
"step": 243
},
{
"epoch": 0.5093945720250522,
"grad_norm": 0.046413128010797784,
"learning_rate": 0.00034942582848580925,
"loss": 0.3848,
"step": 244
},
{
"epoch": 0.511482254697286,
"grad_norm": 0.036986531982911595,
"learning_rate": 0.0003489754269596239,
"loss": 0.3066,
"step": 245
},
{
"epoch": 0.5135699373695198,
"grad_norm": 0.040001114155947204,
"learning_rate": 0.00034852332177494685,
"loss": 0.2852,
"step": 246
},
{
"epoch": 0.5156576200417536,
"grad_norm": 0.04605884877382518,
"learning_rate": 0.0003480695181019786,
"loss": 0.3535,
"step": 247
},
{
"epoch": 0.5177453027139874,
"grad_norm": 0.0406658437248984,
"learning_rate": 0.0003476140211303433,
"loss": 0.334,
"step": 248
},
{
"epoch": 0.5198329853862212,
"grad_norm": 0.03909067687413047,
"learning_rate": 0.0003471568360690296,
"loss": 0.2871,
"step": 249
},
{
"epoch": 0.5219206680584552,
"grad_norm": 0.04831389973643467,
"learning_rate": 0.00034669796814633055,
"loss": 0.3398,
"step": 250
},
{
"epoch": 0.524008350730689,
"grad_norm": 0.04408464727595837,
"learning_rate": 0.0003462374226097844,
"loss": 0.3242,
"step": 251
},
{
"epoch": 0.5260960334029228,
"grad_norm": 0.03798160874377182,
"learning_rate": 0.00034577520472611414,
"loss": 0.252,
"step": 252
},
{
"epoch": 0.5281837160751566,
"grad_norm": 0.04625762493599334,
"learning_rate": 0.0003453113197811675,
"loss": 0.334,
"step": 253
},
{
"epoch": 0.5302713987473904,
"grad_norm": 0.03757333278335248,
"learning_rate": 0.0003448457730798564,
"loss": 0.2734,
"step": 254
},
{
"epoch": 0.5323590814196242,
"grad_norm": 0.038808091023111425,
"learning_rate": 0.0003443785699460965,
"loss": 0.3652,
"step": 255
},
{
"epoch": 0.534446764091858,
"grad_norm": 0.03797009162534055,
"learning_rate": 0.0003439097157227459,
"loss": 0.3223,
"step": 256
},
{
"epoch": 0.5365344467640919,
"grad_norm": 0.038601658815291935,
"learning_rate": 0.0003434392157715445,
"loss": 0.373,
"step": 257
},
{
"epoch": 0.5386221294363257,
"grad_norm": 0.04680506137895324,
"learning_rate": 0.00034296707547305235,
"loss": 0.334,
"step": 258
},
{
"epoch": 0.5407098121085595,
"grad_norm": 0.041499005757660336,
"learning_rate": 0.0003424933002265883,
"loss": 0.3457,
"step": 259
},
{
"epoch": 0.5427974947807933,
"grad_norm": 0.04115196540841461,
"learning_rate": 0.00034201789545016824,
"loss": 0.3242,
"step": 260
},
{
"epoch": 0.5448851774530271,
"grad_norm": 0.04340048389863411,
"learning_rate": 0.00034154086658044297,
"loss": 0.2949,
"step": 261
},
{
"epoch": 0.5469728601252609,
"grad_norm": 0.04378711119505926,
"learning_rate": 0.0003410622190726362,
"loss": 0.3066,
"step": 262
},
{
"epoch": 0.5490605427974948,
"grad_norm": 0.04643412727247773,
"learning_rate": 0.00034058195840048225,
"loss": 0.3652,
"step": 263
},
{
"epoch": 0.5511482254697286,
"grad_norm": 0.042043153196747375,
"learning_rate": 0.0003401000900561631,
"loss": 0.3789,
"step": 264
},
{
"epoch": 0.5532359081419624,
"grad_norm": 0.040040695052475275,
"learning_rate": 0.00033961661955024595,
"loss": 0.3145,
"step": 265
},
{
"epoch": 0.5553235908141962,
"grad_norm": 0.038423365462145156,
"learning_rate": 0.0003391315524116202,
"loss": 0.3223,
"step": 266
},
{
"epoch": 0.55741127348643,
"grad_norm": 0.038539884775496434,
"learning_rate": 0.00033864489418743355,
"loss": 0.3086,
"step": 267
},
{
"epoch": 0.5594989561586639,
"grad_norm": 0.04422443806977669,
"learning_rate": 0.0003381566504430298,
"loss": 0.3945,
"step": 268
},
{
"epoch": 0.5615866388308977,
"grad_norm": 0.0377661934221125,
"learning_rate": 0.00033766682676188386,
"loss": 0.3398,
"step": 269
},
{
"epoch": 0.5636743215031316,
"grad_norm": 0.036534819703584714,
"learning_rate": 0.00033717542874553896,
"loss": 0.3418,
"step": 270
},
{
"epoch": 0.5657620041753654,
"grad_norm": 0.04440346943789282,
"learning_rate": 0.0003366824620135421,
"loss": 0.3984,
"step": 271
},
{
"epoch": 0.5678496868475992,
"grad_norm": 0.037034787962705526,
"learning_rate": 0.00033618793220337965,
"loss": 0.2812,
"step": 272
},
{
"epoch": 0.569937369519833,
"grad_norm": 0.03927620752143223,
"learning_rate": 0.0003356918449704133,
"loss": 0.2695,
"step": 273
},
{
"epoch": 0.5720250521920668,
"grad_norm": 0.04369797108825467,
"learning_rate": 0.00033519420598781505,
"loss": 0.3477,
"step": 274
},
{
"epoch": 0.5741127348643006,
"grad_norm": 0.0379874901088521,
"learning_rate": 0.0003346950209465025,
"loss": 0.248,
"step": 275
},
{
"epoch": 0.5762004175365344,
"grad_norm": 0.042755557159503056,
"learning_rate": 0.0003341942955550738,
"loss": 0.3242,
"step": 276
},
{
"epoch": 0.5782881002087683,
"grad_norm": 0.039892604797522456,
"learning_rate": 0.0003336920355397421,
"loss": 0.3125,
"step": 277
},
{
"epoch": 0.5803757828810021,
"grad_norm": 0.037020438684393274,
"learning_rate": 0.0003331882466442704,
"loss": 0.2832,
"step": 278
},
{
"epoch": 0.5824634655532359,
"grad_norm": 0.0448544789689836,
"learning_rate": 0.0003326829346299057,
"loss": 0.3613,
"step": 279
},
{
"epoch": 0.5845511482254697,
"grad_norm": 0.037923744133296695,
"learning_rate": 0.00033217610527531315,
"loss": 0.3262,
"step": 280
},
{
"epoch": 0.5866388308977035,
"grad_norm": 0.039536585438320955,
"learning_rate": 0.00033166776437650985,
"loss": 0.3301,
"step": 281
},
{
"epoch": 0.5887265135699373,
"grad_norm": 0.042309874793804436,
"learning_rate": 0.0003311579177467988,
"loss": 0.3281,
"step": 282
},
{
"epoch": 0.5908141962421712,
"grad_norm": 0.04247524788442754,
"learning_rate": 0.0003306465712167023,
"loss": 0.3281,
"step": 283
},
{
"epoch": 0.592901878914405,
"grad_norm": 0.038298773282535274,
"learning_rate": 0.0003301337306338951,
"loss": 0.3066,
"step": 284
},
{
"epoch": 0.5949895615866388,
"grad_norm": 0.04845680190635201,
"learning_rate": 0.00032961940186313784,
"loss": 0.3867,
"step": 285
},
{
"epoch": 0.5970772442588727,
"grad_norm": 0.04425999769437242,
"learning_rate": 0.00032910359078620974,
"loss": 0.3848,
"step": 286
},
{
"epoch": 0.5991649269311065,
"grad_norm": 0.04350677333004508,
"learning_rate": 0.0003285863033018416,
"loss": 0.3359,
"step": 287
},
{
"epoch": 0.6012526096033403,
"grad_norm": 0.04313248762187342,
"learning_rate": 0.00032806754532564795,
"loss": 0.3281,
"step": 288
},
{
"epoch": 0.6033402922755741,
"grad_norm": 0.04172934704529679,
"learning_rate": 0.00032754732279005974,
"loss": 0.3887,
"step": 289
},
{
"epoch": 0.605427974947808,
"grad_norm": 0.053530546182614214,
"learning_rate": 0.0003270256416442564,
"loss": 0.4512,
"step": 290
},
{
"epoch": 0.6075156576200418,
"grad_norm": 0.0585588589735363,
"learning_rate": 0.0003265025078540977,
"loss": 0.4023,
"step": 291
},
{
"epoch": 0.6096033402922756,
"grad_norm": 0.046794215765212366,
"learning_rate": 0.00032597792740205573,
"loss": 0.2871,
"step": 292
},
{
"epoch": 0.6116910229645094,
"grad_norm": 0.04076085980117636,
"learning_rate": 0.0003254519062871462,
"loss": 0.3027,
"step": 293
},
{
"epoch": 0.6137787056367432,
"grad_norm": 0.039930806408977594,
"learning_rate": 0.0003249244505248603,
"loss": 0.3086,
"step": 294
},
{
"epoch": 0.615866388308977,
"grad_norm": 0.047339086645206876,
"learning_rate": 0.0003243955661470952,
"loss": 0.373,
"step": 295
},
{
"epoch": 0.6179540709812108,
"grad_norm": 0.041035006122051215,
"learning_rate": 0.00032386525920208574,
"loss": 0.3281,
"step": 296
},
{
"epoch": 0.6200417536534447,
"grad_norm": 0.05042533676976384,
"learning_rate": 0.000323333535754335,
"loss": 0.293,
"step": 297
},
{
"epoch": 0.6221294363256785,
"grad_norm": 0.043959165085212505,
"learning_rate": 0.00032280040188454494,
"loss": 0.3125,
"step": 298
},
{
"epoch": 0.6242171189979123,
"grad_norm": 0.04458635918949398,
"learning_rate": 0.0003222658636895469,
"loss": 0.3105,
"step": 299
},
{
"epoch": 0.6263048016701461,
"grad_norm": 0.04146656432155535,
"learning_rate": 0.0003217299272822316,
"loss": 0.3672,
"step": 300
},
{
"epoch": 0.6283924843423799,
"grad_norm": 0.04285146681080291,
"learning_rate": 0.00032119259879147987,
"loss": 0.2773,
"step": 301
},
{
"epoch": 0.6304801670146137,
"grad_norm": 0.04012756626164825,
"learning_rate": 0.0003206538843620919,
"loss": 0.3145,
"step": 302
},
{
"epoch": 0.6325678496868476,
"grad_norm": 0.03395547556102837,
"learning_rate": 0.0003201137901547175,
"loss": 0.2344,
"step": 303
},
{
"epoch": 0.6346555323590815,
"grad_norm": 0.038474167468473125,
"learning_rate": 0.00031957232234578506,
"loss": 0.3105,
"step": 304
},
{
"epoch": 0.6367432150313153,
"grad_norm": 0.04345850995580467,
"learning_rate": 0.00031902948712743146,
"loss": 0.3711,
"step": 305
},
{
"epoch": 0.6388308977035491,
"grad_norm": 0.04561991595493648,
"learning_rate": 0.00031848529070743113,
"loss": 0.3281,
"step": 306
},
{
"epoch": 0.6409185803757829,
"grad_norm": 0.043639261730465534,
"learning_rate": 0.00031793973930912465,
"loss": 0.3535,
"step": 307
},
{
"epoch": 0.6430062630480167,
"grad_norm": 0.041614849807266444,
"learning_rate": 0.00031739283917134835,
"loss": 0.293,
"step": 308
},
{
"epoch": 0.6450939457202505,
"grad_norm": 0.037829924843835225,
"learning_rate": 0.00031684459654836224,
"loss": 0.2793,
"step": 309
},
{
"epoch": 0.6471816283924844,
"grad_norm": 0.043593747485495736,
"learning_rate": 0.00031629501770977865,
"loss": 0.2871,
"step": 310
},
{
"epoch": 0.6492693110647182,
"grad_norm": 0.043504619259611736,
"learning_rate": 0.0003157441089404911,
"loss": 0.377,
"step": 311
},
{
"epoch": 0.651356993736952,
"grad_norm": 0.042969570910675,
"learning_rate": 0.00031519187654060153,
"loss": 0.3535,
"step": 312
},
{
"epoch": 0.6534446764091858,
"grad_norm": 0.04444146348618118,
"learning_rate": 0.0003146383268253489,
"loss": 0.3633,
"step": 313
},
{
"epoch": 0.6555323590814196,
"grad_norm": 0.0579233541293356,
"learning_rate": 0.00031408346612503684,
"loss": 0.2852,
"step": 314
},
{
"epoch": 0.6576200417536534,
"grad_norm": 0.04116612592402649,
"learning_rate": 0.00031352730078496103,
"loss": 0.3184,
"step": 315
},
{
"epoch": 0.6597077244258872,
"grad_norm": 0.042619485954835536,
"learning_rate": 0.0003129698371653368,
"loss": 0.3555,
"step": 316
},
{
"epoch": 0.6617954070981211,
"grad_norm": 0.03978568537420636,
"learning_rate": 0.00031241108164122667,
"loss": 0.3184,
"step": 317
},
{
"epoch": 0.6638830897703549,
"grad_norm": 0.04398595731711137,
"learning_rate": 0.0003118510406024667,
"loss": 0.3262,
"step": 318
},
{
"epoch": 0.6659707724425887,
"grad_norm": 0.03557319884714253,
"learning_rate": 0.0003112897204535943,
"loss": 0.2539,
"step": 319
},
{
"epoch": 0.6680584551148225,
"grad_norm": 0.04098653755334443,
"learning_rate": 0.0003107271276137745,
"loss": 0.3711,
"step": 320
},
{
"epoch": 0.6701461377870563,
"grad_norm": 0.0377819264468114,
"learning_rate": 0.0003101632685167263,
"loss": 0.2891,
"step": 321
},
{
"epoch": 0.6722338204592901,
"grad_norm": 0.038417025071128603,
"learning_rate": 0.00030959814961064973,
"loss": 0.2891,
"step": 322
},
{
"epoch": 0.6743215031315241,
"grad_norm": 0.038602243878265344,
"learning_rate": 0.0003090317773581518,
"loss": 0.2734,
"step": 323
},
{
"epoch": 0.6764091858037579,
"grad_norm": 0.0487821791964808,
"learning_rate": 0.0003084641582361723,
"loss": 0.3574,
"step": 324
},
{
"epoch": 0.6784968684759917,
"grad_norm": 0.04544872266451946,
"learning_rate": 0.00030789529873591034,
"loss": 0.3008,
"step": 325
},
{
"epoch": 0.6805845511482255,
"grad_norm": 0.0373160417363858,
"learning_rate": 0.0003073252053627496,
"loss": 0.2871,
"step": 326
},
{
"epoch": 0.6826722338204593,
"grad_norm": 0.04358868533626725,
"learning_rate": 0.0003067538846361841,
"loss": 0.3066,
"step": 327
},
{
"epoch": 0.6847599164926931,
"grad_norm": 0.040357136071415334,
"learning_rate": 0.0003061813430897439,
"loss": 0.3516,
"step": 328
},
{
"epoch": 0.6868475991649269,
"grad_norm": 0.04862770193509928,
"learning_rate": 0.0003056075872709199,
"loss": 0.4062,
"step": 329
},
{
"epoch": 0.6889352818371608,
"grad_norm": 0.04688881879011172,
"learning_rate": 0.00030503262374108927,
"loss": 0.4434,
"step": 330
},
{
"epoch": 0.6910229645093946,
"grad_norm": 0.04276287817278344,
"learning_rate": 0.00030445645907544057,
"loss": 0.3105,
"step": 331
},
{
"epoch": 0.6931106471816284,
"grad_norm": 0.03644902974160357,
"learning_rate": 0.0003038790998628981,
"loss": 0.249,
"step": 332
},
{
"epoch": 0.6951983298538622,
"grad_norm": 0.03602308126572102,
"learning_rate": 0.0003033005527060469,
"loss": 0.2891,
"step": 333
},
{
"epoch": 0.697286012526096,
"grad_norm": 0.0397502232733115,
"learning_rate": 0.0003027208242210571,
"loss": 0.334,
"step": 334
},
{
"epoch": 0.6993736951983298,
"grad_norm": 0.042614553937913115,
"learning_rate": 0.0003021399210376084,
"loss": 0.3281,
"step": 335
},
{
"epoch": 0.7014613778705637,
"grad_norm": 0.040360864280806166,
"learning_rate": 0.0003015578497988142,
"loss": 0.3145,
"step": 336
},
{
"epoch": 0.7035490605427975,
"grad_norm": 0.03697519450333998,
"learning_rate": 0.0003009746171611454,
"loss": 0.2852,
"step": 337
},
{
"epoch": 0.7056367432150313,
"grad_norm": 0.03920168369252357,
"learning_rate": 0.0003003902297943545,
"loss": 0.2578,
"step": 338
},
{
"epoch": 0.7077244258872651,
"grad_norm": 0.04596701530986532,
"learning_rate": 0.00029980469438139947,
"loss": 0.3066,
"step": 339
},
{
"epoch": 0.7098121085594989,
"grad_norm": 0.03590053569859233,
"learning_rate": 0.00029921801761836696,
"loss": 0.2598,
"step": 340
},
{
"epoch": 0.7118997912317327,
"grad_norm": 0.039137959610818826,
"learning_rate": 0.0002986302062143961,
"loss": 0.3398,
"step": 341
},
{
"epoch": 0.7139874739039666,
"grad_norm": 0.04445491792602957,
"learning_rate": 0.0002980412668916013,
"loss": 0.3398,
"step": 342
},
{
"epoch": 0.7160751565762005,
"grad_norm": 0.040340445047287156,
"learning_rate": 0.00029745120638499596,
"loss": 0.3516,
"step": 343
},
{
"epoch": 0.7181628392484343,
"grad_norm": 0.04520397395484933,
"learning_rate": 0.00029686003144241494,
"loss": 0.3555,
"step": 344
},
{
"epoch": 0.7202505219206681,
"grad_norm": 0.045695108134425295,
"learning_rate": 0.0002962677488244378,
"loss": 0.2871,
"step": 345
},
{
"epoch": 0.7223382045929019,
"grad_norm": 0.0414553610811136,
"learning_rate": 0.0002956743653043111,
"loss": 0.3828,
"step": 346
},
{
"epoch": 0.7244258872651357,
"grad_norm": 0.043267589906942225,
"learning_rate": 0.00029507988766787135,
"loss": 0.3262,
"step": 347
},
{
"epoch": 0.7265135699373695,
"grad_norm": 0.03665181460868117,
"learning_rate": 0.0002944843227134669,
"loss": 0.3301,
"step": 348
},
{
"epoch": 0.7286012526096033,
"grad_norm": 0.04220684212991339,
"learning_rate": 0.00029388767725188077,
"loss": 0.3223,
"step": 349
},
{
"epoch": 0.7306889352818372,
"grad_norm": 0.04468007941423499,
"learning_rate": 0.0002932899581062524,
"loss": 0.2949,
"step": 350
},
{
"epoch": 0.732776617954071,
"grad_norm": 0.044433764048658723,
"learning_rate": 0.00029269117211199955,
"loss": 0.377,
"step": 351
},
{
"epoch": 0.7348643006263048,
"grad_norm": 0.046209035174827495,
"learning_rate": 0.0002920913261167405,
"loss": 0.3691,
"step": 352
},
{
"epoch": 0.7369519832985386,
"grad_norm": 0.040034133910176115,
"learning_rate": 0.00029149042698021544,
"loss": 0.3105,
"step": 353
},
{
"epoch": 0.7390396659707724,
"grad_norm": 0.04609016944022695,
"learning_rate": 0.00029088848157420813,
"loss": 0.3984,
"step": 354
},
{
"epoch": 0.7411273486430062,
"grad_norm": 0.04480775819697267,
"learning_rate": 0.00029028549678246715,
"loss": 0.332,
"step": 355
},
{
"epoch": 0.7432150313152401,
"grad_norm": 0.04387465996335384,
"learning_rate": 0.0002896814795006277,
"loss": 0.3984,
"step": 356
},
{
"epoch": 0.7453027139874739,
"grad_norm": 0.041864619992452806,
"learning_rate": 0.0002890764366361318,
"loss": 0.3633,
"step": 357
},
{
"epoch": 0.7473903966597077,
"grad_norm": 0.03717012201708029,
"learning_rate": 0.0002884703751081504,
"loss": 0.2324,
"step": 358
},
{
"epoch": 0.7494780793319415,
"grad_norm": 0.043251857860819434,
"learning_rate": 0.0002878633018475035,
"loss": 0.3066,
"step": 359
},
{
"epoch": 0.7515657620041754,
"grad_norm": 0.03591327407860244,
"learning_rate": 0.00028725522379658097,
"loss": 0.2852,
"step": 360
},
{
"epoch": 0.7536534446764092,
"grad_norm": 0.0381908896371197,
"learning_rate": 0.00028664614790926355,
"loss": 0.3477,
"step": 361
},
{
"epoch": 0.755741127348643,
"grad_norm": 0.04392871145759733,
"learning_rate": 0.000286036081150843,
"loss": 0.4414,
"step": 362
},
{
"epoch": 0.7578288100208769,
"grad_norm": 0.04392250979416697,
"learning_rate": 0.0002854250304979423,
"loss": 0.2832,
"step": 363
},
{
"epoch": 0.7599164926931107,
"grad_norm": 0.03445036859255453,
"learning_rate": 0.0002848130029384363,
"loss": 0.3027,
"step": 364
},
{
"epoch": 0.7620041753653445,
"grad_norm": 0.03588315051937214,
"learning_rate": 0.00028420000547137175,
"loss": 0.2793,
"step": 365
},
{
"epoch": 0.7640918580375783,
"grad_norm": 0.043623631697390225,
"learning_rate": 0.00028358604510688666,
"loss": 0.3301,
"step": 366
},
{
"epoch": 0.7661795407098121,
"grad_norm": 0.044899266192686,
"learning_rate": 0.000282971128866131,
"loss": 0.2988,
"step": 367
},
{
"epoch": 0.7682672233820459,
"grad_norm": 0.044441044360934975,
"learning_rate": 0.0002823552637811857,
"loss": 0.3262,
"step": 368
},
{
"epoch": 0.7703549060542797,
"grad_norm": 0.04617336800060674,
"learning_rate": 0.00028173845689498286,
"loss": 0.2812,
"step": 369
},
{
"epoch": 0.7724425887265136,
"grad_norm": 0.03481084134760086,
"learning_rate": 0.0002811207152612243,
"loss": 0.2637,
"step": 370
},
{
"epoch": 0.7745302713987474,
"grad_norm": 0.04359599728927467,
"learning_rate": 0.000280502045944302,
"loss": 0.3516,
"step": 371
},
{
"epoch": 0.7766179540709812,
"grad_norm": 0.03910171209605865,
"learning_rate": 0.0002798824560192164,
"loss": 0.2793,
"step": 372
},
{
"epoch": 0.778705636743215,
"grad_norm": 0.04498875604580853,
"learning_rate": 0.0002792619525714962,
"loss": 0.334,
"step": 373
},
{
"epoch": 0.7807933194154488,
"grad_norm": 0.046945206399684386,
"learning_rate": 0.00027864054269711666,
"loss": 0.3262,
"step": 374
},
{
"epoch": 0.7828810020876826,
"grad_norm": 0.03720210905131818,
"learning_rate": 0.0002780182335024191,
"loss": 0.252,
"step": 375
},
{
"epoch": 0.7849686847599165,
"grad_norm": 0.03971464889707817,
"learning_rate": 0.00027739503210402907,
"loss": 0.334,
"step": 376
},
{
"epoch": 0.7870563674321504,
"grad_norm": 0.047941221203412314,
"learning_rate": 0.00027677094562877534,
"loss": 0.3066,
"step": 377
},
{
"epoch": 0.7891440501043842,
"grad_norm": 0.04116312254250454,
"learning_rate": 0.0002761459812136083,
"loss": 0.3652,
"step": 378
},
{
"epoch": 0.791231732776618,
"grad_norm": 0.0408608818482939,
"learning_rate": 0.00027552014600551813,
"loss": 0.3359,
"step": 379
},
{
"epoch": 0.7933194154488518,
"grad_norm": 0.039519399095894796,
"learning_rate": 0.00027489344716145344,
"loss": 0.3105,
"step": 380
},
{
"epoch": 0.7954070981210856,
"grad_norm": 0.04468756187207934,
"learning_rate": 0.0002742658918482392,
"loss": 0.334,
"step": 381
},
{
"epoch": 0.7974947807933194,
"grad_norm": 0.03799026650779495,
"learning_rate": 0.00027363748724249485,
"loss": 0.3125,
"step": 382
},
{
"epoch": 0.7995824634655533,
"grad_norm": 0.05260052294591301,
"learning_rate": 0.000273008240530552,
"loss": 0.4316,
"step": 383
},
{
"epoch": 0.8016701461377871,
"grad_norm": 0.04069771345832873,
"learning_rate": 0.00027237815890837283,
"loss": 0.2969,
"step": 384
},
{
"epoch": 0.8037578288100209,
"grad_norm": 0.039739122060491815,
"learning_rate": 0.000271747249581467,
"loss": 0.3555,
"step": 385
},
{
"epoch": 0.8058455114822547,
"grad_norm": 0.04610276901781518,
"learning_rate": 0.00027111551976480996,
"loss": 0.3574,
"step": 386
},
{
"epoch": 0.8079331941544885,
"grad_norm": 0.04374672870028719,
"learning_rate": 0.00027048297668276,
"loss": 0.3047,
"step": 387
},
{
"epoch": 0.8100208768267223,
"grad_norm": 0.04631564065384605,
"learning_rate": 0.0002698496275689758,
"loss": 0.2871,
"step": 388
},
{
"epoch": 0.8121085594989561,
"grad_norm": 0.049176525271089545,
"learning_rate": 0.00026921547966633373,
"loss": 0.2871,
"step": 389
},
{
"epoch": 0.81419624217119,
"grad_norm": 0.040234387326007176,
"learning_rate": 0.0002685805402268449,
"loss": 0.3301,
"step": 390
},
{
"epoch": 0.8162839248434238,
"grad_norm": 0.044678234660364975,
"learning_rate": 0.0002679448165115725,
"loss": 0.2852,
"step": 391
},
{
"epoch": 0.8183716075156576,
"grad_norm": 0.043749807508504976,
"learning_rate": 0.00026730831579054823,
"loss": 0.3418,
"step": 392
},
{
"epoch": 0.8204592901878914,
"grad_norm": 0.03702597280717622,
"learning_rate": 0.00026667104534268984,
"loss": 0.334,
"step": 393
},
{
"epoch": 0.8225469728601252,
"grad_norm": 0.04051649281576064,
"learning_rate": 0.00026603301245571733,
"loss": 0.3652,
"step": 394
},
{
"epoch": 0.824634655532359,
"grad_norm": 0.04070349409408308,
"learning_rate": 0.00026539422442607,
"loss": 0.3164,
"step": 395
},
{
"epoch": 0.826722338204593,
"grad_norm": 0.03685594277413145,
"learning_rate": 0.0002647546885588227,
"loss": 0.2695,
"step": 396
},
{
"epoch": 0.8288100208768268,
"grad_norm": 0.033959522628423805,
"learning_rate": 0.00026411441216760243,
"loss": 0.2422,
"step": 397
},
{
"epoch": 0.8308977035490606,
"grad_norm": 0.03649515909769935,
"learning_rate": 0.00026347340257450473,
"loss": 0.252,
"step": 398
},
{
"epoch": 0.8329853862212944,
"grad_norm": 0.05036202800279952,
"learning_rate": 0.00026283166711001,
"loss": 0.3203,
"step": 399
},
{
"epoch": 0.8350730688935282,
"grad_norm": 0.04724871652205951,
"learning_rate": 0.0002621892131128994,
"loss": 0.3555,
"step": 400
},
{
"epoch": 0.837160751565762,
"grad_norm": 0.04523467438521327,
"learning_rate": 0.0002615460479301714,
"loss": 0.3809,
"step": 401
},
{
"epoch": 0.8392484342379958,
"grad_norm": 0.04240307801356352,
"learning_rate": 0.0002609021789169571,
"loss": 0.3574,
"step": 402
},
{
"epoch": 0.8413361169102297,
"grad_norm": 0.03998645664485665,
"learning_rate": 0.0002602576134364369,
"loss": 0.3008,
"step": 403
},
{
"epoch": 0.8434237995824635,
"grad_norm": 0.045774083398434376,
"learning_rate": 0.0002596123588597555,
"loss": 0.4043,
"step": 404
},
{
"epoch": 0.8455114822546973,
"grad_norm": 0.043894774464113946,
"learning_rate": 0.0002589664225659382,
"loss": 0.3301,
"step": 405
},
{
"epoch": 0.8475991649269311,
"grad_norm": 0.04133334314623751,
"learning_rate": 0.0002583198119418063,
"loss": 0.2637,
"step": 406
},
{
"epoch": 0.8496868475991649,
"grad_norm": 0.037286338050560974,
"learning_rate": 0.00025767253438189255,
"loss": 0.3164,
"step": 407
},
{
"epoch": 0.8517745302713987,
"grad_norm": 0.04288387043023777,
"learning_rate": 0.0002570245972883568,
"loss": 0.3164,
"step": 408
},
{
"epoch": 0.8538622129436325,
"grad_norm": 0.04117605927774004,
"learning_rate": 0.000256376008070901,
"loss": 0.3457,
"step": 409
},
{
"epoch": 0.8559498956158664,
"grad_norm": 0.040660730568059064,
"learning_rate": 0.00025572677414668497,
"loss": 0.3691,
"step": 410
},
{
"epoch": 0.8580375782881002,
"grad_norm": 0.03473424285365033,
"learning_rate": 0.00025507690294024105,
"loss": 0.252,
"step": 411
},
{
"epoch": 0.860125260960334,
"grad_norm": 0.04106073793621654,
"learning_rate": 0.00025442640188338945,
"loss": 0.2871,
"step": 412
},
{
"epoch": 0.8622129436325678,
"grad_norm": 0.03780921487044816,
"learning_rate": 0.00025377527841515343,
"loss": 0.2969,
"step": 413
},
{
"epoch": 0.8643006263048016,
"grad_norm": 0.03779696762506096,
"learning_rate": 0.00025312353998167397,
"loss": 0.291,
"step": 414
},
{
"epoch": 0.8663883089770354,
"grad_norm": 0.04781732258675552,
"learning_rate": 0.00025247119403612456,
"loss": 0.3301,
"step": 415
},
{
"epoch": 0.8684759916492694,
"grad_norm": 0.03912752520944136,
"learning_rate": 0.0002518182480386261,
"loss": 0.3477,
"step": 416
},
{
"epoch": 0.8705636743215032,
"grad_norm": 0.043540503401128916,
"learning_rate": 0.0002511647094561619,
"loss": 0.3398,
"step": 417
},
{
"epoch": 0.872651356993737,
"grad_norm": 0.04207586943763798,
"learning_rate": 0.00025051058576249153,
"loss": 0.3398,
"step": 418
},
{
"epoch": 0.8747390396659708,
"grad_norm": 0.03596881591749821,
"learning_rate": 0.00024985588443806624,
"loss": 0.2715,
"step": 419
},
{
"epoch": 0.8768267223382046,
"grad_norm": 0.042287007835322306,
"learning_rate": 0.00024920061296994253,
"loss": 0.3086,
"step": 420
},
{
"epoch": 0.8789144050104384,
"grad_norm": 0.03856865323635185,
"learning_rate": 0.00024854477885169734,
"loss": 0.2871,
"step": 421
},
{
"epoch": 0.8810020876826722,
"grad_norm": 0.03879769693027103,
"learning_rate": 0.0002478883895833417,
"loss": 0.2617,
"step": 422
},
{
"epoch": 0.8830897703549061,
"grad_norm": 0.04500551833152804,
"learning_rate": 0.00024723145267123556,
"loss": 0.2793,
"step": 423
},
{
"epoch": 0.8851774530271399,
"grad_norm": 0.03925914590146329,
"learning_rate": 0.0002465739756280013,
"loss": 0.3086,
"step": 424
},
{
"epoch": 0.8872651356993737,
"grad_norm": 0.04390031842385643,
"learning_rate": 0.0002459159659724383,
"loss": 0.3594,
"step": 425
},
{
"epoch": 0.8893528183716075,
"grad_norm": 0.04015458555102483,
"learning_rate": 0.00024525743122943684,
"loss": 0.2773,
"step": 426
},
{
"epoch": 0.8914405010438413,
"grad_norm": 0.043251502572368115,
"learning_rate": 0.000244598378929892,
"loss": 0.3633,
"step": 427
},
{
"epoch": 0.8935281837160751,
"grad_norm": 0.041718720849523466,
"learning_rate": 0.00024393881661061747,
"loss": 0.3418,
"step": 428
},
{
"epoch": 0.8956158663883089,
"grad_norm": 0.03667112112306563,
"learning_rate": 0.0002432787518142596,
"loss": 0.2754,
"step": 429
},
{
"epoch": 0.8977035490605428,
"grad_norm": 0.04034875824232465,
"learning_rate": 0.0002426181920892108,
"loss": 0.3242,
"step": 430
},
{
"epoch": 0.8997912317327766,
"grad_norm": 0.0416779068349274,
"learning_rate": 0.0002419571449895236,
"loss": 0.2793,
"step": 431
},
{
"epoch": 0.9018789144050104,
"grad_norm": 0.04492298155955886,
"learning_rate": 0.00024129561807482378,
"loss": 0.373,
"step": 432
},
{
"epoch": 0.9039665970772442,
"grad_norm": 0.0358568800030184,
"learning_rate": 0.00024063361891022455,
"loss": 0.293,
"step": 433
},
{
"epoch": 0.906054279749478,
"grad_norm": 0.041240679259523606,
"learning_rate": 0.00023997115506623948,
"loss": 0.3477,
"step": 434
},
{
"epoch": 0.9081419624217119,
"grad_norm": 0.04776747282663658,
"learning_rate": 0.00023930823411869617,
"loss": 0.3633,
"step": 435
},
{
"epoch": 0.9102296450939458,
"grad_norm": 0.03739632540656231,
"learning_rate": 0.00023864486364864963,
"loss": 0.3027,
"step": 436
},
{
"epoch": 0.9123173277661796,
"grad_norm": 0.04013506353404083,
"learning_rate": 0.00023798105124229536,
"loss": 0.2793,
"step": 437
},
{
"epoch": 0.9144050104384134,
"grad_norm": 0.044295244931196354,
"learning_rate": 0.00023731680449088312,
"loss": 0.377,
"step": 438
},
{
"epoch": 0.9164926931106472,
"grad_norm": 0.03805031241226096,
"learning_rate": 0.00023665213099062938,
"loss": 0.2559,
"step": 439
},
{
"epoch": 0.918580375782881,
"grad_norm": 0.03767845062036613,
"learning_rate": 0.00023598703834263113,
"loss": 0.3125,
"step": 440
},
{
"epoch": 0.9206680584551148,
"grad_norm": 0.04968215089246422,
"learning_rate": 0.0002353215341527785,
"loss": 0.3223,
"step": 441
},
{
"epoch": 0.9227557411273486,
"grad_norm": 0.036246340480605606,
"learning_rate": 0.00023465562603166808,
"loss": 0.2734,
"step": 442
},
{
"epoch": 0.9248434237995825,
"grad_norm": 0.034358684978586476,
"learning_rate": 0.00023398932159451557,
"loss": 0.3457,
"step": 443
},
{
"epoch": 0.9269311064718163,
"grad_norm": 0.03631746144581004,
"learning_rate": 0.000233322628461069,
"loss": 0.3672,
"step": 444
},
{
"epoch": 0.9290187891440501,
"grad_norm": 0.037623854588027666,
"learning_rate": 0.0002326555542555215,
"loss": 0.3047,
"step": 445
},
{
"epoch": 0.9311064718162839,
"grad_norm": 0.037886888034629966,
"learning_rate": 0.0002319881066064239,
"loss": 0.2812,
"step": 446
},
{
"epoch": 0.9331941544885177,
"grad_norm": 0.039683782098775454,
"learning_rate": 0.00023132029314659795,
"loss": 0.2969,
"step": 447
},
{
"epoch": 0.9352818371607515,
"grad_norm": 0.032859528564560474,
"learning_rate": 0.00023065212151304843,
"loss": 0.2617,
"step": 448
},
{
"epoch": 0.9373695198329853,
"grad_norm": 0.032092006779359715,
"learning_rate": 0.00022998359934687645,
"loss": 0.2656,
"step": 449
},
{
"epoch": 0.9394572025052192,
"grad_norm": 0.03973715339510534,
"learning_rate": 0.00022931473429319143,
"loss": 0.3242,
"step": 450
},
{
"epoch": 0.941544885177453,
"grad_norm": 0.040296933695116964,
"learning_rate": 0.00022864553400102437,
"loss": 0.3281,
"step": 451
},
{
"epoch": 0.9436325678496869,
"grad_norm": 0.03972360152628702,
"learning_rate": 0.00022797600612323965,
"loss": 0.3301,
"step": 452
},
{
"epoch": 0.9457202505219207,
"grad_norm": 0.03196535128099967,
"learning_rate": 0.00022730615831644807,
"loss": 0.2617,
"step": 453
},
{
"epoch": 0.9478079331941545,
"grad_norm": 0.041518745854850714,
"learning_rate": 0.00022663599824091892,
"loss": 0.3535,
"step": 454
},
{
"epoch": 0.9498956158663883,
"grad_norm": 0.03987544052335102,
"learning_rate": 0.00022596553356049283,
"loss": 0.3184,
"step": 455
},
{
"epoch": 0.9519832985386222,
"grad_norm": 0.03790469940548919,
"learning_rate": 0.0002252947719424935,
"loss": 0.2988,
"step": 456
},
{
"epoch": 0.954070981210856,
"grad_norm": 0.03494334490842234,
"learning_rate": 0.00022462372105764065,
"loss": 0.2793,
"step": 457
},
{
"epoch": 0.9561586638830898,
"grad_norm": 0.03463196563976093,
"learning_rate": 0.0002239523885799618,
"loss": 0.2637,
"step": 458
},
{
"epoch": 0.9582463465553236,
"grad_norm": 0.03593899157528181,
"learning_rate": 0.0002232807821867048,
"loss": 0.3184,
"step": 459
},
{
"epoch": 0.9603340292275574,
"grad_norm": 0.03890353431042402,
"learning_rate": 0.00022260890955825001,
"loss": 0.3008,
"step": 460
},
{
"epoch": 0.9624217118997912,
"grad_norm": 0.033343558090590815,
"learning_rate": 0.00022193677837802235,
"loss": 0.2891,
"step": 461
},
{
"epoch": 0.964509394572025,
"grad_norm": 0.037945187747727475,
"learning_rate": 0.0002212643963324035,
"loss": 0.3574,
"step": 462
},
{
"epoch": 0.9665970772442589,
"grad_norm": 0.03770022520852902,
"learning_rate": 0.00022059177111064391,
"loss": 0.3262,
"step": 463
},
{
"epoch": 0.9686847599164927,
"grad_norm": 0.03875032196468757,
"learning_rate": 0.00021991891040477516,
"loss": 0.3145,
"step": 464
},
{
"epoch": 0.9707724425887265,
"grad_norm": 0.03788954429707523,
"learning_rate": 0.00021924582190952153,
"loss": 0.3008,
"step": 465
},
{
"epoch": 0.9728601252609603,
"grad_norm": 0.035023715952115834,
"learning_rate": 0.00021857251332221256,
"loss": 0.2412,
"step": 466
},
{
"epoch": 0.9749478079331941,
"grad_norm": 0.037210987286251065,
"learning_rate": 0.0002178989923426944,
"loss": 0.3398,
"step": 467
},
{
"epoch": 0.9770354906054279,
"grad_norm": 0.03207771649660305,
"learning_rate": 0.0002172252666732424,
"loss": 0.2578,
"step": 468
},
{
"epoch": 0.9791231732776617,
"grad_norm": 0.031951869684034837,
"learning_rate": 0.00021655134401847245,
"loss": 0.2695,
"step": 469
},
{
"epoch": 0.9812108559498957,
"grad_norm": 0.03674391020210522,
"learning_rate": 0.00021587723208525334,
"loss": 0.3066,
"step": 470
},
{
"epoch": 0.9832985386221295,
"grad_norm": 0.035932078354231324,
"learning_rate": 0.00021520293858261823,
"loss": 0.3086,
"step": 471
},
{
"epoch": 0.9853862212943633,
"grad_norm": 0.03876657300516048,
"learning_rate": 0.00021452847122167677,
"loss": 0.3008,
"step": 472
},
{
"epoch": 0.9874739039665971,
"grad_norm": 0.032766584147543416,
"learning_rate": 0.00021385383771552684,
"loss": 0.3047,
"step": 473
},
{
"epoch": 0.9895615866388309,
"grad_norm": 0.04086665364700512,
"learning_rate": 0.00021317904577916635,
"loss": 0.3223,
"step": 474
},
{
"epoch": 0.9916492693110647,
"grad_norm": 0.043130532139654114,
"learning_rate": 0.0002125041031294049,
"loss": 0.3242,
"step": 475
},
{
"epoch": 0.9937369519832986,
"grad_norm": 0.04273564060035245,
"learning_rate": 0.00021182901748477564,
"loss": 0.4023,
"step": 476
},
{
"epoch": 0.9958246346555324,
"grad_norm": 0.03628393896960175,
"learning_rate": 0.00021115379656544715,
"loss": 0.2969,
"step": 477
},
{
"epoch": 0.9979123173277662,
"grad_norm": 0.03800046175114251,
"learning_rate": 0.00021047844809313467,
"loss": 0.332,
"step": 478
},
{
"epoch": 1.0,
"grad_norm": 0.038197115177252855,
"learning_rate": 0.00020980297979101252,
"loss": 0.2695,
"step": 479
},
{
"epoch": 1.0020876826722338,
"grad_norm": 0.032423776045183476,
"learning_rate": 0.000209127399383625,
"loss": 0.2021,
"step": 480
},
{
"epoch": 1.0041753653444676,
"grad_norm": 0.03224992988614844,
"learning_rate": 0.00020845171459679866,
"loss": 0.2139,
"step": 481
},
{
"epoch": 1.0062630480167014,
"grad_norm": 0.03676757603527534,
"learning_rate": 0.00020777593315755358,
"loss": 0.249,
"step": 482
},
{
"epoch": 1.0083507306889352,
"grad_norm": 0.03889498255916948,
"learning_rate": 0.00020710006279401531,
"loss": 0.3047,
"step": 483
},
{
"epoch": 1.010438413361169,
"grad_norm": 0.03394341377626214,
"learning_rate": 0.00020642411123532605,
"loss": 0.2598,
"step": 484
},
{
"epoch": 1.0125260960334028,
"grad_norm": 0.03195140878316827,
"learning_rate": 0.0002057480862115569,
"loss": 0.2266,
"step": 485
},
{
"epoch": 1.0146137787056368,
"grad_norm": 0.03236779419185144,
"learning_rate": 0.0002050719954536187,
"loss": 0.2373,
"step": 486
},
{
"epoch": 1.0167014613778707,
"grad_norm": 0.033762294027972366,
"learning_rate": 0.0002043958466931744,
"loss": 0.2324,
"step": 487
},
{
"epoch": 1.0187891440501045,
"grad_norm": 0.03225364955535692,
"learning_rate": 0.00020371964766254988,
"loss": 0.2246,
"step": 488
},
{
"epoch": 1.0208768267223383,
"grad_norm": 0.03549374570066917,
"learning_rate": 0.00020304340609464627,
"loss": 0.2559,
"step": 489
},
{
"epoch": 1.022964509394572,
"grad_norm": 0.039376031742883684,
"learning_rate": 0.00020236712972285088,
"loss": 0.291,
"step": 490
},
{
"epoch": 1.0250521920668059,
"grad_norm": 0.033112169954733074,
"learning_rate": 0.0002016908262809491,
"loss": 0.2188,
"step": 491
},
{
"epoch": 1.0271398747390397,
"grad_norm": 0.03556499196371877,
"learning_rate": 0.00020101450350303596,
"loss": 0.2363,
"step": 492
},
{
"epoch": 1.0292275574112735,
"grad_norm": 0.03747952002586393,
"learning_rate": 0.0002003381691234275,
"loss": 0.2734,
"step": 493
},
{
"epoch": 1.0313152400835073,
"grad_norm": 0.039212245318626486,
"learning_rate": 0.00019966183087657255,
"loss": 0.2812,
"step": 494
},
{
"epoch": 1.033402922755741,
"grad_norm": 0.038556823852449205,
"learning_rate": 0.00019898549649696409,
"loss": 0.252,
"step": 495
},
{
"epoch": 1.0354906054279749,
"grad_norm": 0.03678903694409147,
"learning_rate": 0.00019830917371905095,
"loss": 0.2227,
"step": 496
},
{
"epoch": 1.0375782881002087,
"grad_norm": 0.035536317646152164,
"learning_rate": 0.00019763287027714917,
"loss": 0.2451,
"step": 497
},
{
"epoch": 1.0396659707724425,
"grad_norm": 0.03669268286478292,
"learning_rate": 0.00019695659390535377,
"loss": 0.2158,
"step": 498
},
{
"epoch": 1.0417536534446765,
"grad_norm": 0.0365044069944666,
"learning_rate": 0.0001962803523374501,
"loss": 0.2334,
"step": 499
},
{
"epoch": 1.0438413361169103,
"grad_norm": 0.04311308542212124,
"learning_rate": 0.00019560415330682567,
"loss": 0.2754,
"step": 500
},
{
"epoch": 1.0459290187891441,
"grad_norm": 0.03807877662781654,
"learning_rate": 0.00019492800454638128,
"loss": 0.2207,
"step": 501
},
{
"epoch": 1.048016701461378,
"grad_norm": 0.037347719889777004,
"learning_rate": 0.00019425191378844315,
"loss": 0.2598,
"step": 502
},
{
"epoch": 1.0501043841336117,
"grad_norm": 0.041720879231187474,
"learning_rate": 0.00019357588876467403,
"loss": 0.2832,
"step": 503
},
{
"epoch": 1.0521920668058455,
"grad_norm": 0.03611878979996833,
"learning_rate": 0.0001928999372059848,
"loss": 0.2305,
"step": 504
},
{
"epoch": 1.0542797494780793,
"grad_norm": 0.03914121207636423,
"learning_rate": 0.00019222406684244652,
"loss": 0.2715,
"step": 505
},
{
"epoch": 1.0563674321503131,
"grad_norm": 0.045008260139354196,
"learning_rate": 0.00019154828540320141,
"loss": 0.3496,
"step": 506
},
{
"epoch": 1.058455114822547,
"grad_norm": 0.06925075294932237,
"learning_rate": 0.00019087260061637507,
"loss": 0.21,
"step": 507
},
{
"epoch": 1.0605427974947808,
"grad_norm": 0.042949891828889544,
"learning_rate": 0.00019019702020898753,
"loss": 0.3164,
"step": 508
},
{
"epoch": 1.0626304801670146,
"grad_norm": 0.03336924307296428,
"learning_rate": 0.00018952155190686535,
"loss": 0.2334,
"step": 509
},
{
"epoch": 1.0647181628392484,
"grad_norm": 0.037171672304770545,
"learning_rate": 0.00018884620343455292,
"loss": 0.2432,
"step": 510
},
{
"epoch": 1.0668058455114822,
"grad_norm": 0.03845099190077751,
"learning_rate": 0.0001881709825152244,
"loss": 0.2793,
"step": 511
},
{
"epoch": 1.068893528183716,
"grad_norm": 0.03411076783878538,
"learning_rate": 0.00018749589687059513,
"loss": 0.2041,
"step": 512
},
{
"epoch": 1.07098121085595,
"grad_norm": 0.03701566955647055,
"learning_rate": 0.00018682095422083373,
"loss": 0.2305,
"step": 513
},
{
"epoch": 1.0730688935281838,
"grad_norm": 0.03954320852756579,
"learning_rate": 0.00018614616228447318,
"loss": 0.2354,
"step": 514
},
{
"epoch": 1.0751565762004176,
"grad_norm": 0.04313297344221612,
"learning_rate": 0.00018547152877832328,
"loss": 0.2656,
"step": 515
},
{
"epoch": 1.0772442588726514,
"grad_norm": 0.03799814412465588,
"learning_rate": 0.0001847970614173818,
"loss": 0.2285,
"step": 516
},
{
"epoch": 1.0793319415448852,
"grad_norm": 0.04096742193043013,
"learning_rate": 0.0001841227679147467,
"loss": 0.25,
"step": 517
},
{
"epoch": 1.081419624217119,
"grad_norm": 0.03795069745138144,
"learning_rate": 0.00018344865598152754,
"loss": 0.2471,
"step": 518
},
{
"epoch": 1.0835073068893528,
"grad_norm": 0.04325821353197005,
"learning_rate": 0.00018277473332675758,
"loss": 0.2754,
"step": 519
},
{
"epoch": 1.0855949895615866,
"grad_norm": 0.04070425486872679,
"learning_rate": 0.0001821010076573056,
"loss": 0.2617,
"step": 520
},
{
"epoch": 1.0876826722338204,
"grad_norm": 0.031587246842812465,
"learning_rate": 0.00018142748667778746,
"loss": 0.1777,
"step": 521
},
{
"epoch": 1.0897703549060542,
"grad_norm": 0.036539737613616244,
"learning_rate": 0.00018075417809047857,
"loss": 0.2773,
"step": 522
},
{
"epoch": 1.091858037578288,
"grad_norm": 0.041178092025696736,
"learning_rate": 0.00018008108959522497,
"loss": 0.2676,
"step": 523
},
{
"epoch": 1.0939457202505218,
"grad_norm": 0.04046942553591543,
"learning_rate": 0.00017940822888935616,
"loss": 0.2217,
"step": 524
},
{
"epoch": 1.0960334029227556,
"grad_norm": 0.03449629028853095,
"learning_rate": 0.0001787356036675966,
"loss": 0.2061,
"step": 525
},
{
"epoch": 1.0981210855949897,
"grad_norm": 0.03986588393895996,
"learning_rate": 0.0001780632216219777,
"loss": 0.3125,
"step": 526
},
{
"epoch": 1.1002087682672235,
"grad_norm": 0.0333875432706545,
"learning_rate": 0.00017739109044175003,
"loss": 0.1787,
"step": 527
},
{
"epoch": 1.1022964509394573,
"grad_norm": 0.04046361146111627,
"learning_rate": 0.00017671921781329522,
"loss": 0.2715,
"step": 528
},
{
"epoch": 1.104384133611691,
"grad_norm": 0.04262435868048576,
"learning_rate": 0.00017604761142003827,
"loss": 0.2559,
"step": 529
},
{
"epoch": 1.1064718162839249,
"grad_norm": 0.03428115599372914,
"learning_rate": 0.0001753762789423594,
"loss": 0.2061,
"step": 530
},
{
"epoch": 1.1085594989561587,
"grad_norm": 0.03868138089293832,
"learning_rate": 0.00017470522805750654,
"loss": 0.2236,
"step": 531
},
{
"epoch": 1.1106471816283925,
"grad_norm": 0.03475622072699226,
"learning_rate": 0.00017403446643950724,
"loss": 0.248,
"step": 532
},
{
"epoch": 1.1127348643006263,
"grad_norm": 0.03979226115172334,
"learning_rate": 0.00017336400175908112,
"loss": 0.2988,
"step": 533
},
{
"epoch": 1.11482254697286,
"grad_norm": 0.036336277616070035,
"learning_rate": 0.000172693841683552,
"loss": 0.2236,
"step": 534
},
{
"epoch": 1.116910229645094,
"grad_norm": 0.039263581746387936,
"learning_rate": 0.0001720239938767604,
"loss": 0.332,
"step": 535
},
{
"epoch": 1.1189979123173277,
"grad_norm": 0.04395872414762282,
"learning_rate": 0.00017135446599897565,
"loss": 0.2734,
"step": 536
},
{
"epoch": 1.1210855949895615,
"grad_norm": 0.03867598581621191,
"learning_rate": 0.00017068526570680853,
"loss": 0.249,
"step": 537
},
{
"epoch": 1.1231732776617953,
"grad_norm": 0.035662617079617494,
"learning_rate": 0.00017001640065312357,
"loss": 0.2041,
"step": 538
},
{
"epoch": 1.1252609603340291,
"grad_norm": 0.04078186211455469,
"learning_rate": 0.00016934787848695156,
"loss": 0.2119,
"step": 539
},
{
"epoch": 1.1273486430062631,
"grad_norm": 0.03947326929596921,
"learning_rate": 0.00016867970685340215,
"loss": 0.2393,
"step": 540
},
{
"epoch": 1.129436325678497,
"grad_norm": 0.04449645249301931,
"learning_rate": 0.00016801189339357616,
"loss": 0.3203,
"step": 541
},
{
"epoch": 1.1315240083507307,
"grad_norm": 0.046340346999829446,
"learning_rate": 0.00016734444574447862,
"loss": 0.2598,
"step": 542
},
{
"epoch": 1.1336116910229646,
"grad_norm": 0.03957931201804222,
"learning_rate": 0.00016667737153893105,
"loss": 0.2402,
"step": 543
},
{
"epoch": 1.1356993736951984,
"grad_norm": 0.03703757920078097,
"learning_rate": 0.00016601067840548453,
"loss": 0.1924,
"step": 544
},
{
"epoch": 1.1377870563674322,
"grad_norm": 0.03682994100461504,
"learning_rate": 0.000165344373968332,
"loss": 0.2383,
"step": 545
},
{
"epoch": 1.139874739039666,
"grad_norm": 0.03920731523127231,
"learning_rate": 0.00016467846584722153,
"loss": 0.2734,
"step": 546
},
{
"epoch": 1.1419624217118998,
"grad_norm": 0.04276682539134295,
"learning_rate": 0.0001640129616573689,
"loss": 0.2852,
"step": 547
},
{
"epoch": 1.1440501043841336,
"grad_norm": 0.03614032190490994,
"learning_rate": 0.00016334786900937067,
"loss": 0.2354,
"step": 548
},
{
"epoch": 1.1461377870563674,
"grad_norm": 0.041006439179447914,
"learning_rate": 0.00016268319550911695,
"loss": 0.2432,
"step": 549
},
{
"epoch": 1.1482254697286012,
"grad_norm": 0.044517563506872825,
"learning_rate": 0.00016201894875770466,
"loss": 0.2217,
"step": 550
},
{
"epoch": 1.150313152400835,
"grad_norm": 0.03655321665381233,
"learning_rate": 0.00016135513635135045,
"loss": 0.2031,
"step": 551
},
{
"epoch": 1.152400835073069,
"grad_norm": 0.04079189189547784,
"learning_rate": 0.00016069176588130388,
"loss": 0.3027,
"step": 552
},
{
"epoch": 1.1544885177453028,
"grad_norm": 0.03485321031694839,
"learning_rate": 0.00016002884493376054,
"loss": 0.2275,
"step": 553
},
{
"epoch": 1.1565762004175366,
"grad_norm": 0.03052141642274577,
"learning_rate": 0.00015936638108977546,
"loss": 0.1992,
"step": 554
},
{
"epoch": 1.1586638830897704,
"grad_norm": 0.03724204209659217,
"learning_rate": 0.0001587043819251762,
"loss": 0.2207,
"step": 555
},
{
"epoch": 1.1607515657620042,
"grad_norm": 0.03701626710699115,
"learning_rate": 0.00015804285501047646,
"loss": 0.2041,
"step": 556
},
{
"epoch": 1.162839248434238,
"grad_norm": 0.04171363907307022,
"learning_rate": 0.00015738180791078922,
"loss": 0.2715,
"step": 557
},
{
"epoch": 1.1649269311064718,
"grad_norm": 0.0392243815886844,
"learning_rate": 0.00015672124818574042,
"loss": 0.2734,
"step": 558
},
{
"epoch": 1.1670146137787056,
"grad_norm": 0.034576244303318175,
"learning_rate": 0.0001560611833893826,
"loss": 0.207,
"step": 559
},
{
"epoch": 1.1691022964509394,
"grad_norm": 0.03659195217316462,
"learning_rate": 0.0001554016210701081,
"loss": 0.2178,
"step": 560
},
{
"epoch": 1.1711899791231732,
"grad_norm": 0.04227483515507797,
"learning_rate": 0.00015474256877056326,
"loss": 0.3086,
"step": 561
},
{
"epoch": 1.173277661795407,
"grad_norm": 0.0359857035728927,
"learning_rate": 0.00015408403402756177,
"loss": 0.2617,
"step": 562
},
{
"epoch": 1.1753653444676408,
"grad_norm": 0.03680442854270533,
"learning_rate": 0.0001534260243719988,
"loss": 0.209,
"step": 563
},
{
"epoch": 1.1774530271398747,
"grad_norm": 0.03670404417452906,
"learning_rate": 0.0001527685473287645,
"loss": 0.2197,
"step": 564
},
{
"epoch": 1.1795407098121085,
"grad_norm": 0.0393433887764882,
"learning_rate": 0.00015211161041665832,
"loss": 0.2324,
"step": 565
},
{
"epoch": 1.1816283924843423,
"grad_norm": 0.03642939621217633,
"learning_rate": 0.00015145522114830268,
"loss": 0.2393,
"step": 566
},
{
"epoch": 1.1837160751565763,
"grad_norm": 0.03878579783362816,
"learning_rate": 0.00015079938703005752,
"loss": 0.2559,
"step": 567
},
{
"epoch": 1.18580375782881,
"grad_norm": 0.040970554736426454,
"learning_rate": 0.0001501441155619338,
"loss": 0.2139,
"step": 568
},
{
"epoch": 1.187891440501044,
"grad_norm": 0.036820951922352504,
"learning_rate": 0.0001494894142375085,
"loss": 0.1719,
"step": 569
},
{
"epoch": 1.1899791231732777,
"grad_norm": 0.04473872009108077,
"learning_rate": 0.00014883529054383817,
"loss": 0.2734,
"step": 570
},
{
"epoch": 1.1920668058455115,
"grad_norm": 0.037460855453091094,
"learning_rate": 0.00014818175196137392,
"loss": 0.2334,
"step": 571
},
{
"epoch": 1.1941544885177453,
"grad_norm": 0.03980728407908227,
"learning_rate": 0.00014752880596387551,
"loss": 0.2295,
"step": 572
},
{
"epoch": 1.196242171189979,
"grad_norm": 0.03562870229391469,
"learning_rate": 0.00014687646001832608,
"loss": 0.1953,
"step": 573
},
{
"epoch": 1.198329853862213,
"grad_norm": 0.0412827193725494,
"learning_rate": 0.00014622472158484654,
"loss": 0.25,
"step": 574
},
{
"epoch": 1.2004175365344467,
"grad_norm": 0.03634424947703343,
"learning_rate": 0.00014557359811661054,
"loss": 0.207,
"step": 575
},
{
"epoch": 1.2025052192066805,
"grad_norm": 0.036441371511458535,
"learning_rate": 0.000144923097059759,
"loss": 0.249,
"step": 576
},
{
"epoch": 1.2045929018789143,
"grad_norm": 0.04200794927737242,
"learning_rate": 0.00014427322585331505,
"loss": 0.3027,
"step": 577
},
{
"epoch": 1.2066805845511483,
"grad_norm": 0.03620959513479551,
"learning_rate": 0.00014362399192909908,
"loss": 0.2324,
"step": 578
},
{
"epoch": 1.2087682672233822,
"grad_norm": 0.03930165573165628,
"learning_rate": 0.0001429754027116433,
"loss": 0.2812,
"step": 579
},
{
"epoch": 1.210855949895616,
"grad_norm": 0.040322537436277076,
"learning_rate": 0.00014232746561810747,
"loss": 0.2773,
"step": 580
},
{
"epoch": 1.2129436325678498,
"grad_norm": 0.03781024017014511,
"learning_rate": 0.00014168018805819376,
"loss": 0.2354,
"step": 581
},
{
"epoch": 1.2150313152400836,
"grad_norm": 0.038595664424750396,
"learning_rate": 0.00014103357743406182,
"loss": 0.2715,
"step": 582
},
{
"epoch": 1.2171189979123174,
"grad_norm": 0.034975311451833,
"learning_rate": 0.00014038764114024456,
"loss": 0.2041,
"step": 583
},
{
"epoch": 1.2192066805845512,
"grad_norm": 0.03825635258528123,
"learning_rate": 0.00013974238656356313,
"loss": 0.2344,
"step": 584
},
{
"epoch": 1.221294363256785,
"grad_norm": 0.03735073552974533,
"learning_rate": 0.0001390978210830429,
"loss": 0.2559,
"step": 585
},
{
"epoch": 1.2233820459290188,
"grad_norm": 0.040138739862162154,
"learning_rate": 0.00013845395206982863,
"loss": 0.2656,
"step": 586
},
{
"epoch": 1.2254697286012526,
"grad_norm": 0.04418658015359039,
"learning_rate": 0.00013781078688710063,
"loss": 0.291,
"step": 587
},
{
"epoch": 1.2275574112734864,
"grad_norm": 0.04239030702531795,
"learning_rate": 0.00013716833288999005,
"loss": 0.2637,
"step": 588
},
{
"epoch": 1.2296450939457202,
"grad_norm": 0.03818928317637622,
"learning_rate": 0.00013652659742549532,
"loss": 0.252,
"step": 589
},
{
"epoch": 1.231732776617954,
"grad_norm": 0.033789428188020935,
"learning_rate": 0.00013588558783239762,
"loss": 0.1826,
"step": 590
},
{
"epoch": 1.2338204592901878,
"grad_norm": 0.03479799715006895,
"learning_rate": 0.00013524531144117736,
"loss": 0.208,
"step": 591
},
{
"epoch": 1.2359081419624216,
"grad_norm": 0.0407139048904091,
"learning_rate": 0.00013460577557393,
"loss": 0.2832,
"step": 592
},
{
"epoch": 1.2379958246346556,
"grad_norm": 0.04211809721321409,
"learning_rate": 0.00013396698754428266,
"loss": 0.2402,
"step": 593
},
{
"epoch": 1.2400835073068894,
"grad_norm": 0.038313063106902565,
"learning_rate": 0.00013332895465731018,
"loss": 0.2539,
"step": 594
},
{
"epoch": 1.2421711899791232,
"grad_norm": 0.034451480337492464,
"learning_rate": 0.00013269168420945178,
"loss": 0.209,
"step": 595
},
{
"epoch": 1.244258872651357,
"grad_norm": 0.042570134603052,
"learning_rate": 0.0001320551834884276,
"loss": 0.2734,
"step": 596
},
{
"epoch": 1.2463465553235908,
"grad_norm": 0.041331289867428476,
"learning_rate": 0.00013141945977315515,
"loss": 0.3066,
"step": 597
},
{
"epoch": 1.2484342379958246,
"grad_norm": 0.03567452834168155,
"learning_rate": 0.00013078452033366635,
"loss": 0.2129,
"step": 598
},
{
"epoch": 1.2505219206680585,
"grad_norm": 0.03702357216857016,
"learning_rate": 0.00013015037243102425,
"loss": 0.2471,
"step": 599
},
{
"epoch": 1.2526096033402923,
"grad_norm": 0.03423169810999704,
"learning_rate": 0.00012951702331724007,
"loss": 0.2236,
"step": 600
},
{
"epoch": 1.254697286012526,
"grad_norm": 0.04865070295457581,
"learning_rate": 0.00012888448023519006,
"loss": 0.3223,
"step": 601
},
{
"epoch": 1.2567849686847599,
"grad_norm": 0.031921521655174964,
"learning_rate": 0.000128252750418533,
"loss": 0.1934,
"step": 602
},
{
"epoch": 1.2588726513569937,
"grad_norm": 0.03207714489804328,
"learning_rate": 0.0001276218410916272,
"loss": 0.1768,
"step": 603
},
{
"epoch": 1.2609603340292277,
"grad_norm": 0.03764702756962339,
"learning_rate": 0.000126991759469448,
"loss": 0.207,
"step": 604
},
{
"epoch": 1.2630480167014615,
"grad_norm": 0.03494076286574369,
"learning_rate": 0.0001263625127575052,
"loss": 0.2129,
"step": 605
},
{
"epoch": 1.2651356993736953,
"grad_norm": 0.03530986655804941,
"learning_rate": 0.00012573410815176083,
"loss": 0.2617,
"step": 606
},
{
"epoch": 1.267223382045929,
"grad_norm": 0.03741233532866471,
"learning_rate": 0.00012510655283854658,
"loss": 0.1914,
"step": 607
},
{
"epoch": 1.269311064718163,
"grad_norm": 0.03133581344814849,
"learning_rate": 0.00012447985399448194,
"loss": 0.1631,
"step": 608
},
{
"epoch": 1.2713987473903967,
"grad_norm": 0.03819672658693745,
"learning_rate": 0.00012385401878639175,
"loss": 0.2891,
"step": 609
},
{
"epoch": 1.2734864300626305,
"grad_norm": 0.04476513417542424,
"learning_rate": 0.00012322905437122468,
"loss": 0.3379,
"step": 610
},
{
"epoch": 1.2755741127348643,
"grad_norm": 0.037096901415977844,
"learning_rate": 0.00012260496789597093,
"loss": 0.2637,
"step": 611
},
{
"epoch": 1.2776617954070981,
"grad_norm": 0.033560968570005754,
"learning_rate": 0.00012198176649758091,
"loss": 0.1768,
"step": 612
},
{
"epoch": 1.279749478079332,
"grad_norm": 0.03915286043749694,
"learning_rate": 0.00012135945730288333,
"loss": 0.2402,
"step": 613
},
{
"epoch": 1.2818371607515657,
"grad_norm": 0.04243780465562693,
"learning_rate": 0.00012073804742850385,
"loss": 0.2266,
"step": 614
},
{
"epoch": 1.2839248434237995,
"grad_norm": 0.035340613835254,
"learning_rate": 0.00012011754398078369,
"loss": 0.2021,
"step": 615
},
{
"epoch": 1.2860125260960333,
"grad_norm": 0.038064762794890517,
"learning_rate": 0.00011949795405569813,
"loss": 0.2373,
"step": 616
},
{
"epoch": 1.2881002087682671,
"grad_norm": 0.03189685893047627,
"learning_rate": 0.00011887928473877581,
"loss": 0.2178,
"step": 617
},
{
"epoch": 1.290187891440501,
"grad_norm": 0.04119691632833074,
"learning_rate": 0.00011826154310501723,
"loss": 0.2832,
"step": 618
},
{
"epoch": 1.2922755741127347,
"grad_norm": 0.035970451962649,
"learning_rate": 0.00011764473621881427,
"loss": 0.2295,
"step": 619
},
{
"epoch": 1.2943632567849686,
"grad_norm": 0.03644862411137844,
"learning_rate": 0.000117028871133869,
"loss": 0.293,
"step": 620
},
{
"epoch": 1.2964509394572026,
"grad_norm": 0.04232916111512329,
"learning_rate": 0.00011641395489311334,
"loss": 0.2432,
"step": 621
},
{
"epoch": 1.2985386221294364,
"grad_norm": 0.03815713518197317,
"learning_rate": 0.00011579999452862834,
"loss": 0.2363,
"step": 622
},
{
"epoch": 1.3006263048016702,
"grad_norm": 0.03522231596610688,
"learning_rate": 0.00011518699706156373,
"loss": 0.2275,
"step": 623
},
{
"epoch": 1.302713987473904,
"grad_norm": 0.03763846715271956,
"learning_rate": 0.00011457496950205784,
"loss": 0.2139,
"step": 624
},
{
"epoch": 1.3048016701461378,
"grad_norm": 0.03937303225722648,
"learning_rate": 0.00011396391884915707,
"loss": 0.2246,
"step": 625
},
{
"epoch": 1.3068893528183716,
"grad_norm": 0.03259155185424707,
"learning_rate": 0.00011335385209073645,
"loss": 0.2148,
"step": 626
},
{
"epoch": 1.3089770354906054,
"grad_norm": 0.036437191147767146,
"learning_rate": 0.00011274477620341906,
"loss": 0.2168,
"step": 627
},
{
"epoch": 1.3110647181628392,
"grad_norm": 0.03676944565811173,
"learning_rate": 0.00011213669815249659,
"loss": 0.21,
"step": 628
},
{
"epoch": 1.313152400835073,
"grad_norm": 0.03659844810941453,
"learning_rate": 0.00011152962489184955,
"loss": 0.2139,
"step": 629
},
{
"epoch": 1.3152400835073068,
"grad_norm": 0.03855990494361526,
"learning_rate": 0.0001109235633638682,
"loss": 0.2373,
"step": 630
},
{
"epoch": 1.3173277661795408,
"grad_norm": 0.034719164079528235,
"learning_rate": 0.00011031852049937237,
"loss": 0.2236,
"step": 631
},
{
"epoch": 1.3194154488517746,
"grad_norm": 0.03650236752350518,
"learning_rate": 0.00010971450321753276,
"loss": 0.2197,
"step": 632
},
{
"epoch": 1.3215031315240084,
"grad_norm": 0.035674494077417296,
"learning_rate": 0.00010911151842579195,
"loss": 0.25,
"step": 633
},
{
"epoch": 1.3235908141962422,
"grad_norm": 0.040631215715819284,
"learning_rate": 0.00010850957301978462,
"loss": 0.2891,
"step": 634
},
{
"epoch": 1.325678496868476,
"grad_norm": 0.03124048788618308,
"learning_rate": 0.00010790867388325951,
"loss": 0.208,
"step": 635
},
{
"epoch": 1.3277661795407099,
"grad_norm": 0.04271594640915922,
"learning_rate": 0.0001073088278880005,
"loss": 0.3184,
"step": 636
},
{
"epoch": 1.3298538622129437,
"grad_norm": 0.03507030432270385,
"learning_rate": 0.0001067100418937477,
"loss": 0.208,
"step": 637
},
{
"epoch": 1.3319415448851775,
"grad_norm": 0.04020676296142397,
"learning_rate": 0.00010611232274811929,
"loss": 0.2988,
"step": 638
},
{
"epoch": 1.3340292275574113,
"grad_norm": 0.034794700042634165,
"learning_rate": 0.00010551567728653311,
"loss": 0.2061,
"step": 639
},
{
"epoch": 1.336116910229645,
"grad_norm": 0.04251059198573156,
"learning_rate": 0.00010492011233212871,
"loss": 0.2578,
"step": 640
},
{
"epoch": 1.3382045929018789,
"grad_norm": 0.031789932037755535,
"learning_rate": 0.0001043256346956889,
"loss": 0.2109,
"step": 641
},
{
"epoch": 1.3402922755741127,
"grad_norm": 0.039155457576247846,
"learning_rate": 0.00010373225117556224,
"loss": 0.2422,
"step": 642
},
{
"epoch": 1.3423799582463465,
"grad_norm": 0.035388665066062634,
"learning_rate": 0.00010313996855758504,
"loss": 0.2002,
"step": 643
},
{
"epoch": 1.3444676409185803,
"grad_norm": 0.0385841616358797,
"learning_rate": 0.00010254879361500407,
"loss": 0.2041,
"step": 644
},
{
"epoch": 1.346555323590814,
"grad_norm": 0.03971686342813934,
"learning_rate": 0.00010195873310839874,
"loss": 0.2256,
"step": 645
},
{
"epoch": 1.348643006263048,
"grad_norm": 0.04152102958645101,
"learning_rate": 0.00010136979378560398,
"loss": 0.2734,
"step": 646
},
{
"epoch": 1.350730688935282,
"grad_norm": 0.036651673265288884,
"learning_rate": 0.00010078198238163299,
"loss": 0.2148,
"step": 647
},
{
"epoch": 1.3528183716075157,
"grad_norm": 0.032824107867213306,
"learning_rate": 0.00010019530561860051,
"loss": 0.208,
"step": 648
},
{
"epoch": 1.3549060542797495,
"grad_norm": 0.03503308809553616,
"learning_rate": 9.96097702056455e-05,
"loss": 0.2041,
"step": 649
},
{
"epoch": 1.3569937369519833,
"grad_norm": 0.0404355664516907,
"learning_rate": 9.902538283885465e-05,
"loss": 0.2021,
"step": 650
},
{
"epoch": 1.3590814196242171,
"grad_norm": 0.047056225295053385,
"learning_rate": 9.844215020118576e-05,
"loss": 0.3125,
"step": 651
},
{
"epoch": 1.361169102296451,
"grad_norm": 0.0328129234031199,
"learning_rate": 9.78600789623916e-05,
"loss": 0.1973,
"step": 652
},
{
"epoch": 1.3632567849686847,
"grad_norm": 0.03175285628250321,
"learning_rate": 9.727917577894297e-05,
"loss": 0.1914,
"step": 653
},
{
"epoch": 1.3653444676409185,
"grad_norm": 0.04520466671433204,
"learning_rate": 9.669944729395316e-05,
"loss": 0.3125,
"step": 654
},
{
"epoch": 1.3674321503131524,
"grad_norm": 0.03901825127204986,
"learning_rate": 9.612090013710195e-05,
"loss": 0.2412,
"step": 655
},
{
"epoch": 1.3695198329853862,
"grad_norm": 0.037647523878022705,
"learning_rate": 9.554354092455949e-05,
"loss": 0.2227,
"step": 656
},
{
"epoch": 1.3716075156576202,
"grad_norm": 0.03796904304768066,
"learning_rate": 9.496737625891076e-05,
"loss": 0.2402,
"step": 657
},
{
"epoch": 1.373695198329854,
"grad_norm": 0.036620990435136655,
"learning_rate": 9.439241272908012e-05,
"loss": 0.21,
"step": 658
},
{
"epoch": 1.3757828810020878,
"grad_norm": 0.03753406391892656,
"learning_rate": 9.381865691025613e-05,
"loss": 0.2471,
"step": 659
},
{
"epoch": 1.3778705636743216,
"grad_norm": 0.03675743788954959,
"learning_rate": 9.324611536381591e-05,
"loss": 0.2256,
"step": 660
},
{
"epoch": 1.3799582463465554,
"grad_norm": 0.04379732725831618,
"learning_rate": 9.267479463725048e-05,
"loss": 0.3184,
"step": 661
},
{
"epoch": 1.3820459290187892,
"grad_norm": 0.04108282244231379,
"learning_rate": 9.210470126408966e-05,
"loss": 0.3145,
"step": 662
},
{
"epoch": 1.384133611691023,
"grad_norm": 0.033523837380523475,
"learning_rate": 9.15358417638277e-05,
"loss": 0.2021,
"step": 663
},
{
"epoch": 1.3862212943632568,
"grad_norm": 0.03998207042720944,
"learning_rate": 9.096822264184825e-05,
"loss": 0.2441,
"step": 664
},
{
"epoch": 1.3883089770354906,
"grad_norm": 0.037635714292856164,
"learning_rate": 9.040185038935029e-05,
"loss": 0.1934,
"step": 665
},
{
"epoch": 1.3903966597077244,
"grad_norm": 0.04063187756459359,
"learning_rate": 8.983673148327369e-05,
"loss": 0.2812,
"step": 666
},
{
"epoch": 1.3924843423799582,
"grad_norm": 0.03568609472936473,
"learning_rate": 8.927287238622555e-05,
"loss": 0.2578,
"step": 667
},
{
"epoch": 1.394572025052192,
"grad_norm": 0.044075873922673904,
"learning_rate": 8.871027954640567e-05,
"loss": 0.2949,
"step": 668
},
{
"epoch": 1.3966597077244258,
"grad_norm": 0.04018365813830774,
"learning_rate": 8.814895939753331e-05,
"loss": 0.3203,
"step": 669
},
{
"epoch": 1.3987473903966596,
"grad_norm": 0.038371138538880856,
"learning_rate": 8.758891835877335e-05,
"loss": 0.2451,
"step": 670
},
{
"epoch": 1.4008350730688934,
"grad_norm": 0.03644115907045684,
"learning_rate": 8.703016283466323e-05,
"loss": 0.2109,
"step": 671
},
{
"epoch": 1.4029227557411272,
"grad_norm": 0.0355184313530291,
"learning_rate": 8.64726992150391e-05,
"loss": 0.2256,
"step": 672
},
{
"epoch": 1.405010438413361,
"grad_norm": 0.032473676946887235,
"learning_rate": 8.59165338749632e-05,
"loss": 0.2266,
"step": 673
},
{
"epoch": 1.407098121085595,
"grad_norm": 0.03821259918901156,
"learning_rate": 8.536167317465114e-05,
"loss": 0.2305,
"step": 674
},
{
"epoch": 1.4091858037578289,
"grad_norm": 0.035842354722894494,
"learning_rate": 8.480812345939855e-05,
"loss": 0.2363,
"step": 675
},
{
"epoch": 1.4112734864300627,
"grad_norm": 0.034936854456220405,
"learning_rate": 8.4255891059509e-05,
"loss": 0.2344,
"step": 676
},
{
"epoch": 1.4133611691022965,
"grad_norm": 0.0327536953633252,
"learning_rate": 8.370498229022133e-05,
"loss": 0.2354,
"step": 677
},
{
"epoch": 1.4154488517745303,
"grad_norm": 0.03332967518468726,
"learning_rate": 8.315540345163783e-05,
"loss": 0.21,
"step": 678
},
{
"epoch": 1.417536534446764,
"grad_norm": 0.039757185578627195,
"learning_rate": 8.26071608286517e-05,
"loss": 0.2451,
"step": 679
},
{
"epoch": 1.4196242171189979,
"grad_norm": 0.033351639676043855,
"learning_rate": 8.206026069087538e-05,
"loss": 0.1963,
"step": 680
},
{
"epoch": 1.4217118997912317,
"grad_norm": 0.036626791510792185,
"learning_rate": 8.151470929256893e-05,
"loss": 0.2471,
"step": 681
},
{
"epoch": 1.4237995824634655,
"grad_norm": 0.03663750757496257,
"learning_rate": 8.097051287256854e-05,
"loss": 0.2559,
"step": 682
},
{
"epoch": 1.4258872651356993,
"grad_norm": 0.03533804105379715,
"learning_rate": 8.042767765421499e-05,
"loss": 0.2334,
"step": 683
},
{
"epoch": 1.4279749478079333,
"grad_norm": 0.03548267765082115,
"learning_rate": 7.988620984528248e-05,
"loss": 0.1992,
"step": 684
},
{
"epoch": 1.4300626304801671,
"grad_norm": 0.03365970412354974,
"learning_rate": 7.934611563790803e-05,
"loss": 0.2129,
"step": 685
},
{
"epoch": 1.432150313152401,
"grad_norm": 0.036222173404344804,
"learning_rate": 7.880740120852012e-05,
"loss": 0.2441,
"step": 686
},
{
"epoch": 1.4342379958246347,
"grad_norm": 0.03545069244831471,
"learning_rate": 7.827007271776843e-05,
"loss": 0.2246,
"step": 687
},
{
"epoch": 1.4363256784968685,
"grad_norm": 0.036141104648819485,
"learning_rate": 7.773413631045314e-05,
"loss": 0.2334,
"step": 688
},
{
"epoch": 1.4384133611691023,
"grad_norm": 0.03552223808507066,
"learning_rate": 7.719959811545512e-05,
"loss": 0.207,
"step": 689
},
{
"epoch": 1.4405010438413361,
"grad_norm": 0.040531388046511914,
"learning_rate": 7.666646424566508e-05,
"loss": 0.25,
"step": 690
},
{
"epoch": 1.44258872651357,
"grad_norm": 0.03642114412895087,
"learning_rate": 7.613474079791432e-05,
"loss": 0.2275,
"step": 691
},
{
"epoch": 1.4446764091858038,
"grad_norm": 0.03334766371093395,
"learning_rate": 7.56044338529049e-05,
"loss": 0.2246,
"step": 692
},
{
"epoch": 1.4467640918580376,
"grad_norm": 0.038435220776801955,
"learning_rate": 7.50755494751398e-05,
"loss": 0.2871,
"step": 693
},
{
"epoch": 1.4488517745302714,
"grad_norm": 0.03639976375126217,
"learning_rate": 7.454809371285381e-05,
"loss": 0.2236,
"step": 694
},
{
"epoch": 1.4509394572025052,
"grad_norm": 0.04139491524391803,
"learning_rate": 7.402207259794428e-05,
"loss": 0.2617,
"step": 695
},
{
"epoch": 1.453027139874739,
"grad_norm": 0.031147759796875986,
"learning_rate": 7.34974921459023e-05,
"loss": 0.1904,
"step": 696
},
{
"epoch": 1.4551148225469728,
"grad_norm": 0.04026433329347746,
"learning_rate": 7.297435835574362e-05,
"loss": 0.2393,
"step": 697
},
{
"epoch": 1.4572025052192066,
"grad_norm": 0.03965005806725716,
"learning_rate": 7.24526772099403e-05,
"loss": 0.2715,
"step": 698
},
{
"epoch": 1.4592901878914404,
"grad_norm": 0.040893406654203844,
"learning_rate": 7.193245467435206e-05,
"loss": 0.2471,
"step": 699
},
{
"epoch": 1.4613778705636742,
"grad_norm": 0.03636456124885245,
"learning_rate": 7.141369669815841e-05,
"loss": 0.2168,
"step": 700
},
{
"epoch": 1.4634655532359082,
"grad_norm": 0.03647208734905981,
"learning_rate": 7.089640921379026e-05,
"loss": 0.2178,
"step": 701
},
{
"epoch": 1.465553235908142,
"grad_norm": 0.0398294556445215,
"learning_rate": 7.038059813686224e-05,
"loss": 0.2695,
"step": 702
},
{
"epoch": 1.4676409185803758,
"grad_norm": 0.03444283718095077,
"learning_rate": 6.986626936610491e-05,
"loss": 0.2109,
"step": 703
},
{
"epoch": 1.4697286012526096,
"grad_norm": 0.036117772890174496,
"learning_rate": 6.935342878329774e-05,
"loss": 0.1855,
"step": 704
},
{
"epoch": 1.4718162839248434,
"grad_norm": 0.03485285291792577,
"learning_rate": 6.884208225320121e-05,
"loss": 0.1953,
"step": 705
},
{
"epoch": 1.4739039665970772,
"grad_norm": 0.03913119058204658,
"learning_rate": 6.833223562349018e-05,
"loss": 0.291,
"step": 706
},
{
"epoch": 1.475991649269311,
"grad_norm": 0.03619007457603297,
"learning_rate": 6.782389472468687e-05,
"loss": 0.2256,
"step": 707
},
{
"epoch": 1.4780793319415448,
"grad_norm": 0.03954136112257214,
"learning_rate": 6.731706537009437e-05,
"loss": 0.2148,
"step": 708
},
{
"epoch": 1.4801670146137786,
"grad_norm": 0.039048253382048426,
"learning_rate": 6.68117533557297e-05,
"loss": 0.2119,
"step": 709
},
{
"epoch": 1.4822546972860124,
"grad_norm": 0.03633979845664376,
"learning_rate": 6.630796446025793e-05,
"loss": 0.207,
"step": 710
},
{
"epoch": 1.4843423799582465,
"grad_norm": 0.04042749139792089,
"learning_rate": 6.580570444492626e-05,
"loss": 0.2617,
"step": 711
},
{
"epoch": 1.4864300626304803,
"grad_norm": 0.04112986991098296,
"learning_rate": 6.530497905349753e-05,
"loss": 0.2578,
"step": 712
},
{
"epoch": 1.488517745302714,
"grad_norm": 0.04014939278609065,
"learning_rate": 6.480579401218502e-05,
"loss": 0.2324,
"step": 713
},
{
"epoch": 1.4906054279749479,
"grad_norm": 0.03662270065596823,
"learning_rate": 6.430815502958674e-05,
"loss": 0.1689,
"step": 714
},
{
"epoch": 1.4926931106471817,
"grad_norm": 0.033001164537146806,
"learning_rate": 6.381206779662039e-05,
"loss": 0.1826,
"step": 715
},
{
"epoch": 1.4947807933194155,
"grad_norm": 0.04093484928652971,
"learning_rate": 6.331753798645796e-05,
"loss": 0.249,
"step": 716
},
{
"epoch": 1.4968684759916493,
"grad_norm": 0.03400770412940897,
"learning_rate": 6.282457125446109e-05,
"loss": 0.1709,
"step": 717
},
{
"epoch": 1.498956158663883,
"grad_norm": 0.03449610130369155,
"learning_rate": 6.233317323811615e-05,
"loss": 0.2168,
"step": 718
},
{
"epoch": 1.501043841336117,
"grad_norm": 0.03943007646918304,
"learning_rate": 6.184334955697028e-05,
"loss": 0.2617,
"step": 719
},
{
"epoch": 1.5031315240083507,
"grad_norm": 0.04046270086854637,
"learning_rate": 6.135510581256647e-05,
"loss": 0.2715,
"step": 720
},
{
"epoch": 1.5052192066805845,
"grad_norm": 0.036042076960770504,
"learning_rate": 6.086844758837991e-05,
"loss": 0.2275,
"step": 721
},
{
"epoch": 1.5073068893528183,
"grad_norm": 0.038478916176100005,
"learning_rate": 6.0383380449754004e-05,
"loss": 0.1924,
"step": 722
},
{
"epoch": 1.5093945720250521,
"grad_norm": 0.0578411545976244,
"learning_rate": 5.98999099438369e-05,
"loss": 0.2852,
"step": 723
},
{
"epoch": 1.511482254697286,
"grad_norm": 0.03739145944754586,
"learning_rate": 5.941804159951778e-05,
"loss": 0.2197,
"step": 724
},
{
"epoch": 1.5135699373695197,
"grad_norm": 0.036832588153257496,
"learning_rate": 5.893778092736382e-05,
"loss": 0.2422,
"step": 725
},
{
"epoch": 1.5156576200417535,
"grad_norm": 0.03671021108564163,
"learning_rate": 5.845913341955711e-05,
"loss": 0.1572,
"step": 726
},
{
"epoch": 1.5177453027139873,
"grad_norm": 0.03573073453794959,
"learning_rate": 5.798210454983184e-05,
"loss": 0.2168,
"step": 727
},
{
"epoch": 1.5198329853862211,
"grad_norm": 0.035959632617295005,
"learning_rate": 5.7506699773411764e-05,
"loss": 0.2451,
"step": 728
},
{
"epoch": 1.5219206680584552,
"grad_norm": 0.038728578675763334,
"learning_rate": 5.703292452694771e-05,
"loss": 0.2139,
"step": 729
},
{
"epoch": 1.524008350730689,
"grad_norm": 0.035308915371426546,
"learning_rate": 5.6560784228455586e-05,
"loss": 0.2021,
"step": 730
},
{
"epoch": 1.5260960334029228,
"grad_norm": 0.039840996287124834,
"learning_rate": 5.609028427725418e-05,
"loss": 0.1953,
"step": 731
},
{
"epoch": 1.5281837160751566,
"grad_norm": 0.04087634507351788,
"learning_rate": 5.562143005390361e-05,
"loss": 0.2617,
"step": 732
},
{
"epoch": 1.5302713987473904,
"grad_norm": 0.03746689779610756,
"learning_rate": 5.5154226920143626e-05,
"loss": 0.2197,
"step": 733
},
{
"epoch": 1.5323590814196242,
"grad_norm": 0.04226091666932121,
"learning_rate": 5.468868021883256e-05,
"loss": 0.2773,
"step": 734
},
{
"epoch": 1.534446764091858,
"grad_norm": 0.03255698564490662,
"learning_rate": 5.422479527388591e-05,
"loss": 0.1816,
"step": 735
},
{
"epoch": 1.536534446764092,
"grad_norm": 0.03534998758506692,
"learning_rate": 5.376257739021564e-05,
"loss": 0.2158,
"step": 736
},
{
"epoch": 1.5386221294363258,
"grad_norm": 0.04286917750609741,
"learning_rate": 5.330203185366942e-05,
"loss": 0.2559,
"step": 737
},
{
"epoch": 1.5407098121085596,
"grad_norm": 0.03987124501813775,
"learning_rate": 5.284316393097042e-05,
"loss": 0.2197,
"step": 738
},
{
"epoch": 1.5427974947807934,
"grad_norm": 0.0357279862009252,
"learning_rate": 5.2385978869656705e-05,
"loss": 0.209,
"step": 739
},
{
"epoch": 1.5448851774530272,
"grad_norm": 0.03917856556644758,
"learning_rate": 5.193048189802143e-05,
"loss": 0.2402,
"step": 740
},
{
"epoch": 1.546972860125261,
"grad_norm": 0.05340411374410496,
"learning_rate": 5.147667822505318e-05,
"loss": 0.2695,
"step": 741
},
{
"epoch": 1.5490605427974948,
"grad_norm": 0.041333285407966595,
"learning_rate": 5.1024573040376115e-05,
"loss": 0.2969,
"step": 742
},
{
"epoch": 1.5511482254697286,
"grad_norm": 0.039402130485741114,
"learning_rate": 5.057417151419077e-05,
"loss": 0.2402,
"step": 743
},
{
"epoch": 1.5532359081419624,
"grad_norm": 0.035671385682246794,
"learning_rate": 5.012547879721494e-05,
"loss": 0.2158,
"step": 744
},
{
"epoch": 1.5553235908141962,
"grad_norm": 0.038417929606667435,
"learning_rate": 4.967850002062491e-05,
"loss": 0.2236,
"step": 745
},
{
"epoch": 1.55741127348643,
"grad_norm": 0.03949905474465816,
"learning_rate": 4.923324029599632e-05,
"loss": 0.2715,
"step": 746
},
{
"epoch": 1.5594989561586639,
"grad_norm": 0.0352045032702957,
"learning_rate": 4.878970471524622e-05,
"loss": 0.2275,
"step": 747
},
{
"epoch": 1.5615866388308977,
"grad_norm": 0.036303248719977955,
"learning_rate": 4.834789835057465e-05,
"loss": 0.2021,
"step": 748
},
{
"epoch": 1.5636743215031315,
"grad_norm": 0.03443800350271572,
"learning_rate": 4.790782625440655e-05,
"loss": 0.1992,
"step": 749
},
{
"epoch": 1.5657620041753653,
"grad_norm": 0.04231633067945213,
"learning_rate": 4.7469493459334066e-05,
"loss": 0.2617,
"step": 750
},
{
"epoch": 1.567849686847599,
"grad_norm": 0.037830454141616246,
"learning_rate": 4.70329049780589e-05,
"loss": 0.2217,
"step": 751
},
{
"epoch": 1.5699373695198329,
"grad_norm": 0.037662635732679015,
"learning_rate": 4.65980658033353e-05,
"loss": 0.2188,
"step": 752
},
{
"epoch": 1.5720250521920667,
"grad_norm": 0.04067293838178101,
"learning_rate": 4.616498090791248e-05,
"loss": 0.293,
"step": 753
},
{
"epoch": 1.5741127348643005,
"grad_norm": 0.03901581955251283,
"learning_rate": 4.5733655244478146e-05,
"loss": 0.2266,
"step": 754
},
{
"epoch": 1.5762004175365343,
"grad_norm": 0.03554948290229258,
"learning_rate": 4.5304093745601605e-05,
"loss": 0.1943,
"step": 755
},
{
"epoch": 1.5782881002087683,
"grad_norm": 0.0405280460733433,
"learning_rate": 4.487630132367764e-05,
"loss": 0.2617,
"step": 756
},
{
"epoch": 1.5803757828810021,
"grad_norm": 0.03823926366289705,
"learning_rate": 4.445028287087003e-05,
"loss": 0.252,
"step": 757
},
{
"epoch": 1.582463465553236,
"grad_norm": 0.03721532941881517,
"learning_rate": 4.402604325905575e-05,
"loss": 0.2129,
"step": 758
},
{
"epoch": 1.5845511482254697,
"grad_norm": 0.029782665453401677,
"learning_rate": 4.360358733976919e-05,
"loss": 0.1826,
"step": 759
},
{
"epoch": 1.5866388308977035,
"grad_norm": 0.03989589641175334,
"learning_rate": 4.3182919944146936e-05,
"loss": 0.2734,
"step": 760
},
{
"epoch": 1.5887265135699373,
"grad_norm": 0.04383942244265336,
"learning_rate": 4.2764045882872086e-05,
"loss": 0.3008,
"step": 761
},
{
"epoch": 1.5908141962421714,
"grad_norm": 0.034818289814213685,
"learning_rate": 4.2346969946119574e-05,
"loss": 0.2188,
"step": 762
},
{
"epoch": 1.5929018789144052,
"grad_norm": 0.03770186022731002,
"learning_rate": 4.1931696903501204e-05,
"loss": 0.2314,
"step": 763
},
{
"epoch": 1.594989561586639,
"grad_norm": 0.03973578816213652,
"learning_rate": 4.151823150401124e-05,
"loss": 0.2402,
"step": 764
},
{
"epoch": 1.5970772442588728,
"grad_norm": 0.03236007667945323,
"learning_rate": 4.1106578475972035e-05,
"loss": 0.1924,
"step": 765
},
{
"epoch": 1.5991649269311066,
"grad_norm": 0.03764508247239107,
"learning_rate": 4.069674252697988e-05,
"loss": 0.1748,
"step": 766
},
{
"epoch": 1.6012526096033404,
"grad_norm": 0.03731281581422325,
"learning_rate": 4.02887283438514e-05,
"loss": 0.2559,
"step": 767
},
{
"epoch": 1.6033402922755742,
"grad_norm": 0.0372624393342072,
"learning_rate": 3.988254059256971e-05,
"loss": 0.2188,
"step": 768
},
{
"epoch": 1.605427974947808,
"grad_norm": 0.039361210964991795,
"learning_rate": 3.9478183918231106e-05,
"loss": 0.2578,
"step": 769
},
{
"epoch": 1.6075156576200418,
"grad_norm": 0.03750381742110697,
"learning_rate": 3.907566294499201e-05,
"loss": 0.1982,
"step": 770
},
{
"epoch": 1.6096033402922756,
"grad_norm": 0.03771991482783373,
"learning_rate": 3.8674982276016205e-05,
"loss": 0.2412,
"step": 771
},
{
"epoch": 1.6116910229645094,
"grad_norm": 0.04091075279365844,
"learning_rate": 3.82761464934219e-05,
"loss": 0.2793,
"step": 772
},
{
"epoch": 1.6137787056367432,
"grad_norm": 0.03859340968976196,
"learning_rate": 3.787916015822954e-05,
"loss": 0.2471,
"step": 773
},
{
"epoch": 1.615866388308977,
"grad_norm": 0.03200544736921389,
"learning_rate": 3.748402781030955e-05,
"loss": 0.1826,
"step": 774
},
{
"epoch": 1.6179540709812108,
"grad_norm": 0.0340269199931104,
"learning_rate": 3.709075396833057e-05,
"loss": 0.1855,
"step": 775
},
{
"epoch": 1.6200417536534446,
"grad_norm": 0.038072978162411425,
"learning_rate": 3.66993431297076e-05,
"loss": 0.2432,
"step": 776
},
{
"epoch": 1.6221294363256784,
"grad_norm": 0.04102889623967079,
"learning_rate": 3.63097997705506e-05,
"loss": 0.3145,
"step": 777
},
{
"epoch": 1.6242171189979122,
"grad_norm": 0.04007477882751308,
"learning_rate": 3.592212834561337e-05,
"loss": 0.2432,
"step": 778
},
{
"epoch": 1.626304801670146,
"grad_norm": 0.03819036990721501,
"learning_rate": 3.553633328824266e-05,
"loss": 0.1973,
"step": 779
},
{
"epoch": 1.6283924843423798,
"grad_norm": 0.038350798236554316,
"learning_rate": 3.515241901032731e-05,
"loss": 0.2676,
"step": 780
},
{
"epoch": 1.6304801670146136,
"grad_norm": 0.03548951603993466,
"learning_rate": 3.477038990224786e-05,
"loss": 0.2002,
"step": 781
},
{
"epoch": 1.6325678496868476,
"grad_norm": 0.034615335672358714,
"learning_rate": 3.439025033282639e-05,
"loss": 0.1963,
"step": 782
},
{
"epoch": 1.6346555323590815,
"grad_norm": 0.04345337713073096,
"learning_rate": 3.401200464927654e-05,
"loss": 0.2578,
"step": 783
},
{
"epoch": 1.6367432150313153,
"grad_norm": 0.03791537054424221,
"learning_rate": 3.363565717715373e-05,
"loss": 0.2227,
"step": 784
},
{
"epoch": 1.638830897703549,
"grad_norm": 0.04113195871715081,
"learning_rate": 3.326121222030578e-05,
"loss": 0.2012,
"step": 785
},
{
"epoch": 1.6409185803757829,
"grad_norm": 0.041884239195247946,
"learning_rate": 3.288867406082372e-05,
"loss": 0.2236,
"step": 786
},
{
"epoch": 1.6430062630480167,
"grad_norm": 0.03652594444269533,
"learning_rate": 3.251804695899267e-05,
"loss": 0.2139,
"step": 787
},
{
"epoch": 1.6450939457202505,
"grad_norm": 0.05643896674418451,
"learning_rate": 3.214933515324323e-05,
"loss": 0.2422,
"step": 788
},
{
"epoch": 1.6471816283924845,
"grad_norm": 0.04098991740754866,
"learning_rate": 3.178254286010296e-05,
"loss": 0.2422,
"step": 789
},
{
"epoch": 1.6492693110647183,
"grad_norm": 0.04098022159134268,
"learning_rate": 3.1417674274148276e-05,
"loss": 0.2402,
"step": 790
},
{
"epoch": 1.651356993736952,
"grad_norm": 0.037087942509587524,
"learning_rate": 3.105473356795634e-05,
"loss": 0.2295,
"step": 791
},
{
"epoch": 1.653444676409186,
"grad_norm": 0.03775161487801411,
"learning_rate": 3.0693724892057396e-05,
"loss": 0.207,
"step": 792
},
{
"epoch": 1.6555323590814197,
"grad_norm": 0.04001811834655189,
"learning_rate": 3.0334652374887263e-05,
"loss": 0.2773,
"step": 793
},
{
"epoch": 1.6576200417536535,
"grad_norm": 0.03971800362165568,
"learning_rate": 2.997752012274031e-05,
"loss": 0.2139,
"step": 794
},
{
"epoch": 1.6597077244258873,
"grad_norm": 0.03971183328914648,
"learning_rate": 2.9622332219722193e-05,
"loss": 0.2422,
"step": 795
},
{
"epoch": 1.6617954070981211,
"grad_norm": 0.03937647373747184,
"learning_rate": 2.9269092727703375e-05,
"loss": 0.252,
"step": 796
},
{
"epoch": 1.663883089770355,
"grad_norm": 0.03693074285981622,
"learning_rate": 2.891780568627267e-05,
"loss": 0.1992,
"step": 797
},
{
"epoch": 1.6659707724425887,
"grad_norm": 0.03819446049329271,
"learning_rate": 2.8568475112690918e-05,
"loss": 0.2383,
"step": 798
},
{
"epoch": 1.6680584551148225,
"grad_norm": 0.03385681858741079,
"learning_rate": 2.8221105001845095e-05,
"loss": 0.1914,
"step": 799
},
{
"epoch": 1.6701461377870563,
"grad_norm": 0.03481284448136041,
"learning_rate": 2.7875699326202663e-05,
"loss": 0.1943,
"step": 800
},
{
"epoch": 1.6722338204592901,
"grad_norm": 0.038887635380198286,
"learning_rate": 2.7532262035766288e-05,
"loss": 0.2236,
"step": 801
},
{
"epoch": 1.674321503131524,
"grad_norm": 0.039268882753983984,
"learning_rate": 2.7190797058028207e-05,
"loss": 0.2402,
"step": 802
},
{
"epoch": 1.6764091858037578,
"grad_norm": 0.03561285649555011,
"learning_rate": 2.685130829792577e-05,
"loss": 0.2041,
"step": 803
},
{
"epoch": 1.6784968684759916,
"grad_norm": 0.03548955540319579,
"learning_rate": 2.6513799637796745e-05,
"loss": 0.2188,
"step": 804
},
{
"epoch": 1.6805845511482254,
"grad_norm": 0.03987654406934741,
"learning_rate": 2.61782749373346e-05,
"loss": 0.2334,
"step": 805
},
{
"epoch": 1.6826722338204592,
"grad_norm": 0.03591235670312746,
"learning_rate": 2.584473803354468e-05,
"loss": 0.2217,
"step": 806
},
{
"epoch": 1.684759916492693,
"grad_norm": 0.03563754130765487,
"learning_rate": 2.5513192740700144e-05,
"loss": 0.2061,
"step": 807
},
{
"epoch": 1.6868475991649268,
"grad_norm": 0.03453991302292082,
"learning_rate": 2.5183642850298505e-05,
"loss": 0.207,
"step": 808
},
{
"epoch": 1.6889352818371608,
"grad_norm": 0.040646656045912735,
"learning_rate": 2.4856092131018137e-05,
"loss": 0.293,
"step": 809
},
{
"epoch": 1.6910229645093946,
"grad_norm": 0.03639422717002704,
"learning_rate": 2.4530544328675186e-05,
"loss": 0.2236,
"step": 810
},
{
"epoch": 1.6931106471816284,
"grad_norm": 0.04124184507078617,
"learning_rate": 2.4207003166180765e-05,
"loss": 0.2344,
"step": 811
},
{
"epoch": 1.6951983298538622,
"grad_norm": 0.04821861121500109,
"learning_rate": 2.3885472343498515e-05,
"loss": 0.2227,
"step": 812
},
{
"epoch": 1.697286012526096,
"grad_norm": 0.036261622126831795,
"learning_rate": 2.3565955537602014e-05,
"loss": 0.2266,
"step": 813
},
{
"epoch": 1.6993736951983298,
"grad_norm": 0.03833360109114772,
"learning_rate": 2.3248456402432918e-05,
"loss": 0.2598,
"step": 814
},
{
"epoch": 1.7014613778705638,
"grad_norm": 0.03811210564749943,
"learning_rate": 2.293297856885912e-05,
"loss": 0.2197,
"step": 815
},
{
"epoch": 1.7035490605427976,
"grad_norm": 0.03440881800931742,
"learning_rate": 2.261952564463332e-05,
"loss": 0.1865,
"step": 816
},
{
"epoch": 1.7056367432150314,
"grad_norm": 0.03784666030148996,
"learning_rate": 2.2308101214351628e-05,
"loss": 0.2324,
"step": 817
},
{
"epoch": 1.7077244258872653,
"grad_norm": 0.03706179096561306,
"learning_rate": 2.1998708839412597e-05,
"loss": 0.21,
"step": 818
},
{
"epoch": 1.709812108559499,
"grad_norm": 0.036230331032483534,
"learning_rate": 2.1691352057976565e-05,
"loss": 0.2471,
"step": 819
},
{
"epoch": 1.7118997912317329,
"grad_norm": 0.04090973690460557,
"learning_rate": 2.138603438492517e-05,
"loss": 0.2637,
"step": 820
},
{
"epoch": 1.7139874739039667,
"grad_norm": 0.051662701812649245,
"learning_rate": 2.108275931182111e-05,
"loss": 0.1953,
"step": 821
},
{
"epoch": 1.7160751565762005,
"grad_norm": 0.038240986750146325,
"learning_rate": 2.0781530306868246e-05,
"loss": 0.2393,
"step": 822
},
{
"epoch": 1.7181628392484343,
"grad_norm": 0.04106815130974131,
"learning_rate": 2.048235081487202e-05,
"loss": 0.2637,
"step": 823
},
{
"epoch": 1.720250521920668,
"grad_norm": 0.03673610429332063,
"learning_rate": 2.0185224257199908e-05,
"loss": 0.2207,
"step": 824
},
{
"epoch": 1.7223382045929019,
"grad_norm": 0.04334868227291063,
"learning_rate": 1.9890154031742325e-05,
"loss": 0.3223,
"step": 825
},
{
"epoch": 1.7244258872651357,
"grad_norm": 0.04017653068637208,
"learning_rate": 1.9597143512873872e-05,
"loss": 0.2656,
"step": 826
},
{
"epoch": 1.7265135699373695,
"grad_norm": 0.04143141277717955,
"learning_rate": 1.9306196051414728e-05,
"loss": 0.252,
"step": 827
},
{
"epoch": 1.7286012526096033,
"grad_norm": 0.030689616701832047,
"learning_rate": 1.9017314974592203e-05,
"loss": 0.1982,
"step": 828
},
{
"epoch": 1.730688935281837,
"grad_norm": 0.04149364456774439,
"learning_rate": 1.873050358600279e-05,
"loss": 0.2451,
"step": 829
},
{
"epoch": 1.732776617954071,
"grad_norm": 0.037822515436121175,
"learning_rate": 1.8445765165574324e-05,
"loss": 0.2246,
"step": 830
},
{
"epoch": 1.7348643006263047,
"grad_norm": 0.036300250288018185,
"learning_rate": 1.8163102969528702e-05,
"loss": 0.1709,
"step": 831
},
{
"epoch": 1.7369519832985385,
"grad_norm": 0.03813579853432033,
"learning_rate": 1.7882520230344225e-05,
"loss": 0.1982,
"step": 832
},
{
"epoch": 1.7390396659707723,
"grad_norm": 0.03490543301641405,
"learning_rate": 1.7604020156719026e-05,
"loss": 0.2256,
"step": 833
},
{
"epoch": 1.7411273486430061,
"grad_norm": 0.03767801651422133,
"learning_rate": 1.7327605933534087e-05,
"loss": 0.248,
"step": 834
},
{
"epoch": 1.7432150313152401,
"grad_norm": 0.03559965744869205,
"learning_rate": 1.705328072181711e-05,
"loss": 0.1963,
"step": 835
},
{
"epoch": 1.745302713987474,
"grad_norm": 0.03620103246900227,
"learning_rate": 1.6781047658706094e-05,
"loss": 0.2363,
"step": 836
},
{
"epoch": 1.7473903966597077,
"grad_norm": 0.03363208108698153,
"learning_rate": 1.6510909857413593e-05,
"loss": 0.208,
"step": 837
},
{
"epoch": 1.7494780793319415,
"grad_norm": 0.038864402335742426,
"learning_rate": 1.624287040719108e-05,
"loss": 0.2217,
"step": 838
},
{
"epoch": 1.7515657620041754,
"grad_norm": 0.03604996210348334,
"learning_rate": 1.597693237329363e-05,
"loss": 0.2363,
"step": 839
},
{
"epoch": 1.7536534446764092,
"grad_norm": 0.03870805026122142,
"learning_rate": 1.5713098796944913e-05,
"loss": 0.2363,
"step": 840
},
{
"epoch": 1.755741127348643,
"grad_norm": 0.037967280022889786,
"learning_rate": 1.545137269530228e-05,
"loss": 0.2031,
"step": 841
},
{
"epoch": 1.757828810020877,
"grad_norm": 0.032100601105533956,
"learning_rate": 1.5191757061422484e-05,
"loss": 0.1963,
"step": 842
},
{
"epoch": 1.7599164926931108,
"grad_norm": 0.03677180115256404,
"learning_rate": 1.493425486422717e-05,
"loss": 0.2324,
"step": 843
},
{
"epoch": 1.7620041753653446,
"grad_norm": 0.04179333438984321,
"learning_rate": 1.467886904846918e-05,
"loss": 0.2197,
"step": 844
},
{
"epoch": 1.7640918580375784,
"grad_norm": 0.03330069993588093,
"learning_rate": 1.4425602534698645e-05,
"loss": 0.1992,
"step": 845
},
{
"epoch": 1.7661795407098122,
"grad_norm": 0.03921593097418753,
"learning_rate": 1.4174458219229868e-05,
"loss": 0.2734,
"step": 846
},
{
"epoch": 1.768267223382046,
"grad_norm": 0.0384869660070795,
"learning_rate": 1.3925438974107918e-05,
"loss": 0.2285,
"step": 847
},
{
"epoch": 1.7703549060542798,
"grad_norm": 0.03235612278723638,
"learning_rate": 1.3678547647075946e-05,
"loss": 0.2021,
"step": 848
},
{
"epoch": 1.7724425887265136,
"grad_norm": 0.034557637083813396,
"learning_rate": 1.3433787061542525e-05,
"loss": 0.1982,
"step": 849
},
{
"epoch": 1.7745302713987474,
"grad_norm": 0.036691512115311166,
"learning_rate": 1.3191160016549564e-05,
"loss": 0.2695,
"step": 850
},
{
"epoch": 1.7766179540709812,
"grad_norm": 0.036890816812044876,
"learning_rate": 1.2950669286740024e-05,
"loss": 0.2119,
"step": 851
},
{
"epoch": 1.778705636743215,
"grad_norm": 0.03801656481599902,
"learning_rate": 1.271231762232632e-05,
"loss": 0.2354,
"step": 852
},
{
"epoch": 1.7807933194154488,
"grad_norm": 0.0431132757880389,
"learning_rate": 1.2476107749058986e-05,
"loss": 0.2305,
"step": 853
},
{
"epoch": 1.7828810020876826,
"grad_norm": 0.034619924337342727,
"learning_rate": 1.2242042368195218e-05,
"loss": 0.1914,
"step": 854
},
{
"epoch": 1.7849686847599164,
"grad_norm": 0.034941322602160714,
"learning_rate": 1.2010124156468294e-05,
"loss": 0.2051,
"step": 855
},
{
"epoch": 1.7870563674321502,
"grad_norm": 0.03577282594283845,
"learning_rate": 1.1780355766056694e-05,
"loss": 0.2334,
"step": 856
},
{
"epoch": 1.789144050104384,
"grad_norm": 0.03923352027261708,
"learning_rate": 1.1552739824554026e-05,
"loss": 0.2217,
"step": 857
},
{
"epoch": 1.7912317327766178,
"grad_norm": 0.03464671275803716,
"learning_rate": 1.1327278934938723e-05,
"loss": 0.1992,
"step": 858
},
{
"epoch": 1.7933194154488517,
"grad_norm": 0.04157357259303964,
"learning_rate": 1.1103975675544443e-05,
"loss": 0.2461,
"step": 859
},
{
"epoch": 1.7954070981210855,
"grad_norm": 0.040227594429237123,
"learning_rate": 1.0882832600030624e-05,
"loss": 0.2441,
"step": 860
},
{
"epoch": 1.7974947807933193,
"grad_norm": 0.03488250123821703,
"learning_rate": 1.0663852237353067e-05,
"loss": 0.1963,
"step": 861
},
{
"epoch": 1.7995824634655533,
"grad_norm": 0.03979475512791435,
"learning_rate": 1.0447037091735223e-05,
"loss": 0.2471,
"step": 862
},
{
"epoch": 1.801670146137787,
"grad_norm": 0.03997223499064563,
"learning_rate": 1.0232389642639395e-05,
"loss": 0.2471,
"step": 863
},
{
"epoch": 1.803757828810021,
"grad_norm": 0.04038899286733726,
"learning_rate": 1.0019912344738625e-05,
"loss": 0.2598,
"step": 864
},
{
"epoch": 1.8058455114822547,
"grad_norm": 0.03748451896837008,
"learning_rate": 9.809607627888296e-06,
"loss": 0.2002,
"step": 865
},
{
"epoch": 1.8079331941544885,
"grad_norm": 0.03427034514456546,
"learning_rate": 9.601477897098576e-06,
"loss": 0.1729,
"step": 866
},
{
"epoch": 1.8100208768267223,
"grad_norm": 0.04471055633543089,
"learning_rate": 9.395525532506866e-06,
"loss": 0.2256,
"step": 867
},
{
"epoch": 1.812108559498956,
"grad_norm": 0.03475782341698621,
"learning_rate": 9.191752889350547e-06,
"loss": 0.2207,
"step": 868
},
{
"epoch": 1.8141962421711901,
"grad_norm": 0.03828798248246293,
"learning_rate": 8.990162297940097e-06,
"loss": 0.2109,
"step": 869
},
{
"epoch": 1.816283924843424,
"grad_norm": 0.04180601763110725,
"learning_rate": 8.79075606363231e-06,
"loss": 0.2676,
"step": 870
},
{
"epoch": 1.8183716075156577,
"grad_norm": 0.03646735664376503,
"learning_rate": 8.593536466804142e-06,
"loss": 0.207,
"step": 871
},
{
"epoch": 1.8204592901878915,
"grad_norm": 0.03218572550634406,
"learning_rate": 8.398505762826503e-06,
"loss": 0.1797,
"step": 872
},
{
"epoch": 1.8225469728601253,
"grad_norm": 0.03982709976184709,
"learning_rate": 8.205666182038418e-06,
"loss": 0.2432,
"step": 873
},
{
"epoch": 1.8246346555323592,
"grad_norm": 0.037982327489911684,
"learning_rate": 8.015019929721668e-06,
"loss": 0.2344,
"step": 874
},
{
"epoch": 1.826722338204593,
"grad_norm": 0.03580893689768604,
"learning_rate": 7.826569186075428e-06,
"loss": 0.252,
"step": 875
},
{
"epoch": 1.8288100208768268,
"grad_norm": 0.035854237119299116,
"learning_rate": 7.640316106191403e-06,
"loss": 0.1963,
"step": 876
},
{
"epoch": 1.8308977035490606,
"grad_norm": 0.03548915356785303,
"learning_rate": 7.4562628200292475e-06,
"loss": 0.2148,
"step": 877
},
{
"epoch": 1.8329853862212944,
"grad_norm": 0.03888900644009877,
"learning_rate": 7.274411432392026e-06,
"loss": 0.2812,
"step": 878
},
{
"epoch": 1.8350730688935282,
"grad_norm": 0.036848786139394796,
"learning_rate": 7.0947640229023675e-06,
"loss": 0.1904,
"step": 879
},
{
"epoch": 1.837160751565762,
"grad_norm": 0.032594039620939604,
"learning_rate": 6.917322645978463e-06,
"loss": 0.1904,
"step": 880
},
{
"epoch": 1.8392484342379958,
"grad_norm": 0.038807798544498055,
"learning_rate": 6.742089330810774e-06,
"loss": 0.2383,
"step": 881
},
{
"epoch": 1.8413361169102296,
"grad_norm": 0.042279119809553053,
"learning_rate": 6.5690660813386484e-06,
"loss": 0.3184,
"step": 882
},
{
"epoch": 1.8434237995824634,
"grad_norm": 0.03575523378084652,
"learning_rate": 6.398254876227561e-06,
"loss": 0.1875,
"step": 883
},
{
"epoch": 1.8455114822546972,
"grad_norm": 0.03438492023581685,
"learning_rate": 6.229657668846333e-06,
"loss": 0.21,
"step": 884
},
{
"epoch": 1.847599164926931,
"grad_norm": 0.04355612584591555,
"learning_rate": 6.0632763872449535e-06,
"loss": 0.2754,
"step": 885
},
{
"epoch": 1.8496868475991648,
"grad_norm": 0.03533181091838774,
"learning_rate": 5.899112934132345e-06,
"loss": 0.1689,
"step": 886
},
{
"epoch": 1.8517745302713986,
"grad_norm": 0.040660080641486476,
"learning_rate": 5.73716918685483e-06,
"loss": 0.2471,
"step": 887
},
{
"epoch": 1.8538622129436324,
"grad_norm": 0.03235795768355878,
"learning_rate": 5.577446997374458e-06,
"loss": 0.1826,
"step": 888
},
{
"epoch": 1.8559498956158664,
"grad_norm": 0.03609407610302125,
"learning_rate": 5.4199481922479146e-06,
"loss": 0.1816,
"step": 889
},
{
"epoch": 1.8580375782881002,
"grad_norm": 0.03647085365814291,
"learning_rate": 5.264674572605665e-06,
"loss": 0.2393,
"step": 890
},
{
"epoch": 1.860125260960334,
"grad_norm": 0.03892481200719314,
"learning_rate": 5.11162791413129e-06,
"loss": 0.2021,
"step": 891
},
{
"epoch": 1.8622129436325678,
"grad_norm": 0.03859819801659933,
"learning_rate": 4.9608099670412025e-06,
"loss": 0.2402,
"step": 892
},
{
"epoch": 1.8643006263048016,
"grad_norm": 0.03638768037937149,
"learning_rate": 4.812222456064697e-06,
"loss": 0.1729,
"step": 893
},
{
"epoch": 1.8663883089770354,
"grad_norm": 0.03578385304837865,
"learning_rate": 4.665867080424046e-06,
"loss": 0.2217,
"step": 894
},
{
"epoch": 1.8684759916492695,
"grad_norm": 0.03899470496061886,
"learning_rate": 4.521745513815345e-06,
"loss": 0.2236,
"step": 895
},
{
"epoch": 1.8705636743215033,
"grad_norm": 0.03565199656237529,
"learning_rate": 4.379859404389075e-06,
"loss": 0.2158,
"step": 896
},
{
"epoch": 1.872651356993737,
"grad_norm": 0.03700782502510176,
"learning_rate": 4.240210374731479e-06,
"loss": 0.2441,
"step": 897
},
{
"epoch": 1.8747390396659709,
"grad_norm": 0.034012775228610695,
"learning_rate": 4.10280002184591e-06,
"loss": 0.1934,
"step": 898
},
{
"epoch": 1.8768267223382047,
"grad_norm": 0.032231872294556166,
"learning_rate": 3.967629917134574e-06,
"loss": 0.1689,
"step": 899
},
{
"epoch": 1.8789144050104385,
"grad_norm": 0.039125390388194826,
"learning_rate": 3.834701606380575e-06,
"loss": 0.2285,
"step": 900
},
{
"epoch": 1.8810020876826723,
"grad_norm": 0.04416962700457997,
"learning_rate": 3.7040166097301877e-06,
"loss": 0.2656,
"step": 901
},
{
"epoch": 1.883089770354906,
"grad_norm": 0.04111412261545973,
"learning_rate": 3.575576421675586e-06,
"loss": 0.248,
"step": 902
},
{
"epoch": 1.88517745302714,
"grad_norm": 0.03397805823890809,
"learning_rate": 3.449382511037613e-06,
"loss": 0.1904,
"step": 903
},
{
"epoch": 1.8872651356993737,
"grad_norm": 0.041722415283545866,
"learning_rate": 3.3254363209491046e-06,
"loss": 0.3301,
"step": 904
},
{
"epoch": 1.8893528183716075,
"grad_norm": 0.047301660349510964,
"learning_rate": 3.203739268838324e-06,
"loss": 0.2988,
"step": 905
},
{
"epoch": 1.8914405010438413,
"grad_norm": 0.03091322302919706,
"learning_rate": 3.0842927464127537e-06,
"loss": 0.1592,
"step": 906
},
{
"epoch": 1.8935281837160751,
"grad_norm": 0.038250787995771175,
"learning_rate": 2.9670981196431525e-06,
"loss": 0.248,
"step": 907
},
{
"epoch": 1.895615866388309,
"grad_norm": 0.04184624767282992,
"learning_rate": 2.8521567287480344e-06,
"loss": 0.2461,
"step": 908
},
{
"epoch": 1.8977035490605427,
"grad_norm": 0.03669678971086905,
"learning_rate": 2.7394698881782367e-06,
"loss": 0.1895,
"step": 909
},
{
"epoch": 1.8997912317327765,
"grad_norm": 0.03332356389562326,
"learning_rate": 2.6290388866019533e-06,
"loss": 0.1934,
"step": 910
},
{
"epoch": 1.9018789144050103,
"grad_norm": 0.03567454628961636,
"learning_rate": 2.5208649868899704e-06,
"loss": 0.2207,
"step": 911
},
{
"epoch": 1.9039665970772441,
"grad_norm": 0.03723478614951914,
"learning_rate": 2.4149494261012097e-06,
"loss": 0.2002,
"step": 912
},
{
"epoch": 1.906054279749478,
"grad_norm": 0.03840886510699952,
"learning_rate": 2.3112934154686296e-06,
"loss": 0.1982,
"step": 913
},
{
"epoch": 1.9081419624217117,
"grad_norm": 0.03658052263392586,
"learning_rate": 2.2098981403852804e-06,
"loss": 0.2295,
"step": 914
},
{
"epoch": 1.9102296450939458,
"grad_norm": 0.042943855139820596,
"learning_rate": 2.1107647603908933e-06,
"loss": 0.2422,
"step": 915
},
{
"epoch": 1.9123173277661796,
"grad_norm": 0.038181568479852754,
"learning_rate": 2.0138944091585123e-06,
"loss": 0.1768,
"step": 916
},
{
"epoch": 1.9144050104384134,
"grad_norm": 0.04044402128707661,
"learning_rate": 1.91928819448155e-06,
"loss": 0.249,
"step": 917
},
{
"epoch": 1.9164926931106472,
"grad_norm": 0.04021345081714543,
"learning_rate": 1.8269471982611086e-06,
"loss": 0.2578,
"step": 918
},
{
"epoch": 1.918580375782881,
"grad_norm": 0.03765027742041835,
"learning_rate": 1.7368724764936118e-06,
"loss": 0.2275,
"step": 919
},
{
"epoch": 1.9206680584551148,
"grad_norm": 0.03276750255084186,
"learning_rate": 1.6490650592588363e-06,
"loss": 0.1797,
"step": 920
},
{
"epoch": 1.9227557411273486,
"grad_norm": 0.03836166999423338,
"learning_rate": 1.5635259507079224e-06,
"loss": 0.2656,
"step": 921
},
{
"epoch": 1.9248434237995826,
"grad_norm": 0.040716455351115934,
"learning_rate": 1.4802561290520934e-06,
"loss": 0.2373,
"step": 922
},
{
"epoch": 1.9269311064718164,
"grad_norm": 0.044722992742083724,
"learning_rate": 1.3992565465512863e-06,
"loss": 0.2832,
"step": 923
},
{
"epoch": 1.9290187891440502,
"grad_norm": 0.035718501429723326,
"learning_rate": 1.3205281295034512e-06,
"loss": 0.2109,
"step": 924
},
{
"epoch": 1.931106471816284,
"grad_norm": 0.040897036162254374,
"learning_rate": 1.2440717782338018e-06,
"loss": 0.248,
"step": 925
},
{
"epoch": 1.9331941544885178,
"grad_norm": 0.037701179364082175,
"learning_rate": 1.169888367084626e-06,
"loss": 0.2021,
"step": 926
},
{
"epoch": 1.9352818371607516,
"grad_norm": 0.03800126744938023,
"learning_rate": 1.0979787444051815e-06,
"loss": 0.2051,
"step": 927
},
{
"epoch": 1.9373695198329854,
"grad_norm": 0.0376053096363303,
"learning_rate": 1.0283437325421252e-06,
"loss": 0.25,
"step": 928
},
{
"epoch": 1.9394572025052192,
"grad_norm": 0.03852373604309701,
"learning_rate": 9.60984127829989e-07,
"loss": 0.1895,
"step": 929
},
{
"epoch": 1.941544885177453,
"grad_norm": 0.05141719488909021,
"learning_rate": 8.95900700582164e-07,
"loss": 0.2969,
"step": 930
},
{
"epoch": 1.9436325678496869,
"grad_norm": 0.0422557444402279,
"learning_rate": 8.330941950819738e-07,
"loss": 0.2793,
"step": 931
},
{
"epoch": 1.9457202505219207,
"grad_norm": 0.03878673278759856,
"learning_rate": 7.725653295743484e-07,
"loss": 0.2412,
"step": 932
},
{
"epoch": 1.9478079331941545,
"grad_norm": 0.04515324231545769,
"learning_rate": 7.14314796257387e-07,
"loss": 0.2715,
"step": 933
},
{
"epoch": 1.9498956158663883,
"grad_norm": 0.04060496658163304,
"learning_rate": 6.58343261274652e-07,
"loss": 0.2559,
"step": 934
},
{
"epoch": 1.951983298538622,
"grad_norm": 0.0410883973174675,
"learning_rate": 6.046513647074203e-07,
"loss": 0.2676,
"step": 935
},
{
"epoch": 1.9540709812108559,
"grad_norm": 0.037425826708713576,
"learning_rate": 5.532397205673556e-07,
"loss": 0.208,
"step": 936
},
{
"epoch": 1.9561586638830897,
"grad_norm": 0.03711360512013879,
"learning_rate": 5.04108916789603e-07,
"loss": 0.2246,
"step": 937
},
{
"epoch": 1.9582463465553235,
"grad_norm": 0.04129502522262085,
"learning_rate": 4.5725951522592734e-07,
"loss": 0.2812,
"step": 938
},
{
"epoch": 1.9603340292275573,
"grad_norm": 0.042768062565203456,
"learning_rate": 4.126920516384303e-07,
"loss": 0.2734,
"step": 939
},
{
"epoch": 1.962421711899791,
"grad_norm": 0.03636718305173902,
"learning_rate": 3.704070356932432e-07,
"loss": 0.2275,
"step": 940
},
{
"epoch": 1.964509394572025,
"grad_norm": 0.03914519289291409,
"learning_rate": 3.3040495095491006e-07,
"loss": 0.2051,
"step": 941
},
{
"epoch": 1.966597077244259,
"grad_norm": 0.036435492987632524,
"learning_rate": 2.926862548806364e-07,
"loss": 0.21,
"step": 942
},
{
"epoch": 1.9686847599164927,
"grad_norm": 0.04520438858168889,
"learning_rate": 2.5725137881529306e-07,
"loss": 0.2852,
"step": 943
},
{
"epoch": 1.9707724425887265,
"grad_norm": 0.036719387383711895,
"learning_rate": 2.2410072798624283e-07,
"loss": 0.2422,
"step": 944
},
{
"epoch": 1.9728601252609603,
"grad_norm": 0.03890839218462144,
"learning_rate": 1.9323468149892165e-07,
"loss": 0.2617,
"step": 945
},
{
"epoch": 1.9749478079331941,
"grad_norm": 0.04401800316091134,
"learning_rate": 1.6465359233233114e-07,
"loss": 0.2314,
"step": 946
},
{
"epoch": 1.977035490605428,
"grad_norm": 0.042526472328545395,
"learning_rate": 1.383577873351305e-07,
"loss": 0.2734,
"step": 947
},
{
"epoch": 1.9791231732776617,
"grad_norm": 0.04123640104260365,
"learning_rate": 1.143475672218175e-07,
"loss": 0.2715,
"step": 948
},
{
"epoch": 1.9812108559498958,
"grad_norm": 0.041143122691388076,
"learning_rate": 9.26232065693089e-08,
"loss": 0.2578,
"step": 949
},
{
"epoch": 1.9832985386221296,
"grad_norm": 0.03646932624947557,
"learning_rate": 7.318495381383184e-08,
"loss": 0.1953,
"step": 950
},
{
"epoch": 1.9853862212943634,
"grad_norm": 0.042625916797236235,
"learning_rate": 5.603303124803727e-08,
"loss": 0.2344,
"step": 951
},
{
"epoch": 1.9874739039665972,
"grad_norm": 0.039620902455050594,
"learning_rate": 4.1167635018513064e-08,
"loss": 0.2598,
"step": 952
},
{
"epoch": 1.989561586638831,
"grad_norm": 0.0329100227494744,
"learning_rate": 2.8588935123474714e-08,
"loss": 0.1758,
"step": 953
},
{
"epoch": 1.9916492693110648,
"grad_norm": 0.03829557846462901,
"learning_rate": 1.8297075410877997e-08,
"loss": 0.2676,
"step": 954
},
{
"epoch": 1.9937369519832986,
"grad_norm": 0.037593692719248624,
"learning_rate": 1.02921735767314e-08,
"loss": 0.1943,
"step": 955
},
{
"epoch": 1.9958246346555324,
"grad_norm": 0.03806855037144764,
"learning_rate": 4.574321163763884e-09,
"loss": 0.2383,
"step": 956
},
{
"epoch": 1.9979123173277662,
"grad_norm": 0.03559582629136701,
"learning_rate": 1.1435835604034495e-09,
"loss": 0.1787,
"step": 957
},
{
"epoch": 2.0,
"grad_norm": 0.03239360059033573,
"learning_rate": 0.0,
"loss": 0.167,
"step": 958
},
{
"epoch": 2.0,
"step": 958,
"total_flos": 5.650597015940235e+18,
"train_loss": 0.3088689816022965,
"train_runtime": 6060.4121,
"train_samples_per_second": 90.951,
"train_steps_per_second": 0.158
}
],
"logging_steps": 1.0,
"max_steps": 958,
"num_input_tokens_seen": 0,
"num_train_epochs": 2,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 5.650597015940235e+18,
"train_batch_size": 72,
"trial_name": null,
"trial_params": null
}