TLDR_code_release / trainer_state.json
lizhongzhi2022's picture
Upload folder using huggingface_hub
d90b602 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.416,
"eval_steps": 32,
"global_step": 832,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.0005,
"grad_norm": 1.1535558177055725,
"learning_rate": 1e-05,
"loss": 0.4543,
"step": 1
},
{
"epoch": 0.001,
"grad_norm": 0.5039215671909039,
"learning_rate": 1e-05,
"loss": 0.3002,
"step": 2
},
{
"epoch": 0.0015,
"grad_norm": 0.4111085873239227,
"learning_rate": 1e-05,
"loss": 0.2138,
"step": 3
},
{
"epoch": 0.002,
"grad_norm": 0.3258240952011462,
"learning_rate": 1e-05,
"loss": 0.1761,
"step": 4
},
{
"epoch": 0.0025,
"grad_norm": 0.25417983371771585,
"learning_rate": 1e-05,
"loss": 0.1923,
"step": 5
},
{
"epoch": 0.003,
"grad_norm": 0.2023054568437071,
"learning_rate": 1e-05,
"loss": 0.2035,
"step": 6
},
{
"epoch": 0.0035,
"grad_norm": 0.29479546251817934,
"learning_rate": 1e-05,
"loss": 0.166,
"step": 7
},
{
"epoch": 0.004,
"grad_norm": 0.2603707531918473,
"learning_rate": 1e-05,
"loss": 0.1563,
"step": 8
},
{
"epoch": 0.0045,
"grad_norm": 0.27772543950112105,
"learning_rate": 1e-05,
"loss": 0.1742,
"step": 9
},
{
"epoch": 0.005,
"grad_norm": 0.33899668368493546,
"learning_rate": 1e-05,
"loss": 0.1623,
"step": 10
},
{
"epoch": 0.0055,
"grad_norm": 0.17293404759978015,
"learning_rate": 1e-05,
"loss": 0.1902,
"step": 11
},
{
"epoch": 0.006,
"grad_norm": 0.16812733049304054,
"learning_rate": 1e-05,
"loss": 0.1505,
"step": 12
},
{
"epoch": 0.0065,
"grad_norm": 0.24635507048250077,
"learning_rate": 1e-05,
"loss": 0.1496,
"step": 13
},
{
"epoch": 0.007,
"grad_norm": 0.23026986852648482,
"learning_rate": 1e-05,
"loss": 0.1428,
"step": 14
},
{
"epoch": 0.0075,
"grad_norm": 0.20312493098149043,
"learning_rate": 1e-05,
"loss": 0.1411,
"step": 15
},
{
"epoch": 0.008,
"grad_norm": 0.19056569111470645,
"learning_rate": 1e-05,
"loss": 0.1556,
"step": 16
},
{
"epoch": 0.0085,
"grad_norm": 0.19180586046592904,
"learning_rate": 1e-05,
"loss": 0.1448,
"step": 17
},
{
"epoch": 0.009,
"grad_norm": 0.23670452313674922,
"learning_rate": 1e-05,
"loss": 0.1272,
"step": 18
},
{
"epoch": 0.0095,
"grad_norm": 0.2241037709056984,
"learning_rate": 1e-05,
"loss": 0.1584,
"step": 19
},
{
"epoch": 0.01,
"grad_norm": 0.1828368631692288,
"learning_rate": 1e-05,
"loss": 0.1347,
"step": 20
},
{
"epoch": 0.0105,
"grad_norm": 0.23115912688140353,
"learning_rate": 1e-05,
"loss": 0.1293,
"step": 21
},
{
"epoch": 0.011,
"grad_norm": 0.18886903067681482,
"learning_rate": 1e-05,
"loss": 0.1175,
"step": 22
},
{
"epoch": 0.0115,
"grad_norm": 0.19723236256202736,
"learning_rate": 1e-05,
"loss": 0.1665,
"step": 23
},
{
"epoch": 0.012,
"grad_norm": 0.2141017706399641,
"learning_rate": 1e-05,
"loss": 0.1491,
"step": 24
},
{
"epoch": 0.0125,
"grad_norm": 0.19090893978368054,
"learning_rate": 1e-05,
"loss": 0.1113,
"step": 25
},
{
"epoch": 0.013,
"grad_norm": 0.2460352826446744,
"learning_rate": 1e-05,
"loss": 0.1142,
"step": 26
},
{
"epoch": 0.0135,
"grad_norm": 0.21847319513439203,
"learning_rate": 1e-05,
"loss": 0.1273,
"step": 27
},
{
"epoch": 0.014,
"grad_norm": 0.18979686333150375,
"learning_rate": 1e-05,
"loss": 0.1441,
"step": 28
},
{
"epoch": 0.0145,
"grad_norm": 0.18886921913659824,
"learning_rate": 1e-05,
"loss": 0.1481,
"step": 29
},
{
"epoch": 0.015,
"grad_norm": 0.2115532479997522,
"learning_rate": 1e-05,
"loss": 0.0792,
"step": 30
},
{
"epoch": 0.0155,
"grad_norm": 0.20266301051804922,
"learning_rate": 1e-05,
"loss": 0.1379,
"step": 31
},
{
"epoch": 0.016,
"grad_norm": 0.1865900767381873,
"learning_rate": 1e-05,
"loss": 0.1453,
"step": 32
},
{
"epoch": 0.016,
"eval_dev_acc": 0.61328125,
"eval_dev_token": 5204.017578125,
"eval_runtime": 351.6008,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.003,
"step": 32
},
{
"epoch": 0.0165,
"grad_norm": 0.23402804608211233,
"learning_rate": 1e-05,
"loss": 0.1461,
"step": 33
},
{
"epoch": 0.017,
"grad_norm": 0.2191224488373381,
"learning_rate": 1e-05,
"loss": 0.1076,
"step": 34
},
{
"epoch": 0.0175,
"grad_norm": 0.18221820677824999,
"learning_rate": 1e-05,
"loss": 0.138,
"step": 35
},
{
"epoch": 0.018,
"grad_norm": 0.20326615377722793,
"learning_rate": 1e-05,
"loss": 0.1021,
"step": 36
},
{
"epoch": 0.0185,
"grad_norm": 0.2449658821374275,
"learning_rate": 1e-05,
"loss": 0.0881,
"step": 37
},
{
"epoch": 0.019,
"grad_norm": 0.2068478649138205,
"learning_rate": 1e-05,
"loss": 0.1249,
"step": 38
},
{
"epoch": 0.0195,
"grad_norm": 0.22630770783782977,
"learning_rate": 1e-05,
"loss": 0.1259,
"step": 39
},
{
"epoch": 0.02,
"grad_norm": 0.21173641583720768,
"learning_rate": 1e-05,
"loss": 0.1331,
"step": 40
},
{
"epoch": 0.0205,
"grad_norm": 0.23616623193061376,
"learning_rate": 1e-05,
"loss": 0.1192,
"step": 41
},
{
"epoch": 0.021,
"grad_norm": 0.20789242969830385,
"learning_rate": 1e-05,
"loss": 0.159,
"step": 42
},
{
"epoch": 0.0215,
"grad_norm": 0.21662842275351119,
"learning_rate": 1e-05,
"loss": 0.1455,
"step": 43
},
{
"epoch": 0.022,
"grad_norm": 0.224914886884455,
"learning_rate": 1e-05,
"loss": 0.1565,
"step": 44
},
{
"epoch": 0.0225,
"grad_norm": 0.17642201019062015,
"learning_rate": 1e-05,
"loss": 0.1022,
"step": 45
},
{
"epoch": 0.023,
"grad_norm": 0.19476006095917964,
"learning_rate": 1e-05,
"loss": 0.1738,
"step": 46
},
{
"epoch": 0.0235,
"grad_norm": 0.2041987726527936,
"learning_rate": 1e-05,
"loss": 0.1545,
"step": 47
},
{
"epoch": 0.024,
"grad_norm": 0.15364946520211809,
"learning_rate": 1e-05,
"loss": 0.141,
"step": 48
},
{
"epoch": 0.0245,
"grad_norm": 0.1632596100583654,
"learning_rate": 1e-05,
"loss": 0.1493,
"step": 49
},
{
"epoch": 0.025,
"grad_norm": 0.23305215604624085,
"learning_rate": 1e-05,
"loss": 0.1164,
"step": 50
},
{
"epoch": 0.0255,
"grad_norm": 0.2697225282405861,
"learning_rate": 1e-05,
"loss": 0.0921,
"step": 51
},
{
"epoch": 0.026,
"grad_norm": 0.19242004892152365,
"learning_rate": 1e-05,
"loss": 0.1224,
"step": 52
},
{
"epoch": 0.0265,
"grad_norm": 0.21422189358330607,
"learning_rate": 1e-05,
"loss": 0.1084,
"step": 53
},
{
"epoch": 0.027,
"grad_norm": 0.22368520523252378,
"learning_rate": 1e-05,
"loss": 0.1018,
"step": 54
},
{
"epoch": 0.0275,
"grad_norm": 0.15315126193234804,
"learning_rate": 1e-05,
"loss": 0.117,
"step": 55
},
{
"epoch": 0.028,
"grad_norm": 0.2042745134173473,
"learning_rate": 1e-05,
"loss": 0.1355,
"step": 56
},
{
"epoch": 0.0285,
"grad_norm": 0.23889007522498773,
"learning_rate": 1e-05,
"loss": 0.1387,
"step": 57
},
{
"epoch": 0.029,
"grad_norm": 0.17240068749452392,
"learning_rate": 1e-05,
"loss": 0.1634,
"step": 58
},
{
"epoch": 0.0295,
"grad_norm": 0.1899912754500666,
"learning_rate": 1e-05,
"loss": 0.1027,
"step": 59
},
{
"epoch": 0.03,
"grad_norm": 0.17118724561465037,
"learning_rate": 1e-05,
"loss": 0.1398,
"step": 60
},
{
"epoch": 0.0305,
"grad_norm": 0.19950059905690032,
"learning_rate": 1e-05,
"loss": 0.1118,
"step": 61
},
{
"epoch": 0.031,
"grad_norm": 0.19928365636144094,
"learning_rate": 1e-05,
"loss": 0.1146,
"step": 62
},
{
"epoch": 0.0315,
"grad_norm": 0.23146246196039105,
"learning_rate": 1e-05,
"loss": 0.0871,
"step": 63
},
{
"epoch": 0.032,
"grad_norm": 0.22746980065915193,
"learning_rate": 1e-05,
"loss": 0.1673,
"step": 64
},
{
"epoch": 0.032,
"eval_dev_acc": 0.541015625,
"eval_dev_token": 5677.76171875,
"eval_runtime": 358.7405,
"eval_samples_per_second": 0.178,
"eval_steps_per_second": 0.003,
"step": 64
},
{
"epoch": 0.0325,
"grad_norm": 0.1992092305273338,
"learning_rate": 1e-05,
"loss": 0.1292,
"step": 65
},
{
"epoch": 0.033,
"grad_norm": 0.19429880128063629,
"learning_rate": 1e-05,
"loss": 0.0911,
"step": 66
},
{
"epoch": 0.0335,
"grad_norm": 0.1287473705431077,
"learning_rate": 1e-05,
"loss": 0.1377,
"step": 67
},
{
"epoch": 0.034,
"grad_norm": 0.2228992327697556,
"learning_rate": 1e-05,
"loss": 0.0933,
"step": 68
},
{
"epoch": 0.0345,
"grad_norm": 0.18793882506839266,
"learning_rate": 1e-05,
"loss": 0.1097,
"step": 69
},
{
"epoch": 0.035,
"grad_norm": 0.1932965061071618,
"learning_rate": 1e-05,
"loss": 0.1062,
"step": 70
},
{
"epoch": 0.0355,
"grad_norm": 0.20585403698562318,
"learning_rate": 1e-05,
"loss": 0.1279,
"step": 71
},
{
"epoch": 0.036,
"grad_norm": 0.17833203451544913,
"learning_rate": 1e-05,
"loss": 0.1353,
"step": 72
},
{
"epoch": 0.0365,
"grad_norm": 0.21658266347490054,
"learning_rate": 1e-05,
"loss": 0.0961,
"step": 73
},
{
"epoch": 0.037,
"grad_norm": 0.2364173046757495,
"learning_rate": 1e-05,
"loss": 0.1038,
"step": 74
},
{
"epoch": 0.0375,
"grad_norm": 0.20844999333456934,
"learning_rate": 1e-05,
"loss": 0.1266,
"step": 75
},
{
"epoch": 0.038,
"grad_norm": 0.2288998611422715,
"learning_rate": 1e-05,
"loss": 0.0872,
"step": 76
},
{
"epoch": 0.0385,
"grad_norm": 0.1878179335817694,
"learning_rate": 1e-05,
"loss": 0.1038,
"step": 77
},
{
"epoch": 0.039,
"grad_norm": 0.19984002821227043,
"learning_rate": 1e-05,
"loss": 0.1376,
"step": 78
},
{
"epoch": 0.0395,
"grad_norm": 0.20050327741314344,
"learning_rate": 1e-05,
"loss": 0.135,
"step": 79
},
{
"epoch": 0.04,
"grad_norm": 0.19353835667751798,
"learning_rate": 1e-05,
"loss": 0.12,
"step": 80
},
{
"epoch": 0.0405,
"grad_norm": 0.12986340217496,
"learning_rate": 1e-05,
"loss": 0.1754,
"step": 81
},
{
"epoch": 0.041,
"grad_norm": 0.1889393654868388,
"learning_rate": 1e-05,
"loss": 0.1206,
"step": 82
},
{
"epoch": 0.0415,
"grad_norm": 0.17201063783314552,
"learning_rate": 1e-05,
"loss": 0.1401,
"step": 83
},
{
"epoch": 0.042,
"grad_norm": 0.19004092111820917,
"learning_rate": 1e-05,
"loss": 0.122,
"step": 84
},
{
"epoch": 0.0425,
"grad_norm": 0.21797692428743218,
"learning_rate": 1e-05,
"loss": 0.127,
"step": 85
},
{
"epoch": 0.043,
"grad_norm": 0.16779726003397347,
"learning_rate": 1e-05,
"loss": 0.1437,
"step": 86
},
{
"epoch": 0.0435,
"grad_norm": 0.23214262430834917,
"learning_rate": 1e-05,
"loss": 0.1187,
"step": 87
},
{
"epoch": 0.044,
"grad_norm": 0.19415443790822307,
"learning_rate": 1e-05,
"loss": 0.1063,
"step": 88
},
{
"epoch": 0.0445,
"grad_norm": 0.19254517148646239,
"learning_rate": 1e-05,
"loss": 0.1039,
"step": 89
},
{
"epoch": 0.045,
"grad_norm": 0.17307659461239167,
"learning_rate": 1e-05,
"loss": 0.1554,
"step": 90
},
{
"epoch": 0.0455,
"grad_norm": 0.24265029122082277,
"learning_rate": 1e-05,
"loss": 0.1387,
"step": 91
},
{
"epoch": 0.046,
"grad_norm": 0.17107218032177454,
"learning_rate": 1e-05,
"loss": 0.1301,
"step": 92
},
{
"epoch": 0.0465,
"grad_norm": 0.18075891622609033,
"learning_rate": 1e-05,
"loss": 0.1174,
"step": 93
},
{
"epoch": 0.047,
"grad_norm": 0.21595935885391185,
"learning_rate": 1e-05,
"loss": 0.1129,
"step": 94
},
{
"epoch": 0.0475,
"grad_norm": 0.23262812126963384,
"learning_rate": 1e-05,
"loss": 0.0983,
"step": 95
},
{
"epoch": 0.048,
"grad_norm": 0.20420999614021612,
"learning_rate": 1e-05,
"loss": 0.1161,
"step": 96
},
{
"epoch": 0.048,
"eval_dev_acc": 0.63671875,
"eval_dev_token": 5515.37109375,
"eval_runtime": 348.8656,
"eval_samples_per_second": 0.183,
"eval_steps_per_second": 0.003,
"step": 96
},
{
"epoch": 0.0485,
"grad_norm": 0.22980199994832454,
"learning_rate": 1e-05,
"loss": 0.1,
"step": 97
},
{
"epoch": 0.049,
"grad_norm": 0.19262135795217625,
"learning_rate": 1e-05,
"loss": 0.1124,
"step": 98
},
{
"epoch": 0.0495,
"grad_norm": 0.22869326649776367,
"learning_rate": 1e-05,
"loss": 0.1094,
"step": 99
},
{
"epoch": 0.05,
"grad_norm": 0.24173954720541516,
"learning_rate": 1e-05,
"loss": 0.0847,
"step": 100
},
{
"epoch": 0.0505,
"grad_norm": 0.2332215752101726,
"learning_rate": 1e-05,
"loss": 0.0946,
"step": 101
},
{
"epoch": 0.051,
"grad_norm": 0.26475494931892063,
"learning_rate": 1e-05,
"loss": 0.1094,
"step": 102
},
{
"epoch": 0.0515,
"grad_norm": 0.20833383536152603,
"learning_rate": 1e-05,
"loss": 0.0931,
"step": 103
},
{
"epoch": 0.052,
"grad_norm": 0.22676431614598225,
"learning_rate": 1e-05,
"loss": 0.0866,
"step": 104
},
{
"epoch": 0.0525,
"grad_norm": 0.22204744101927545,
"learning_rate": 1e-05,
"loss": 0.0984,
"step": 105
},
{
"epoch": 0.053,
"grad_norm": 0.2012349417230909,
"learning_rate": 1e-05,
"loss": 0.0899,
"step": 106
},
{
"epoch": 0.0535,
"grad_norm": 0.20131082411517556,
"learning_rate": 1e-05,
"loss": 0.095,
"step": 107
},
{
"epoch": 0.054,
"grad_norm": 0.2501268960199406,
"learning_rate": 1e-05,
"loss": 0.0872,
"step": 108
},
{
"epoch": 0.0545,
"grad_norm": 0.1877564425582315,
"learning_rate": 1e-05,
"loss": 0.1447,
"step": 109
},
{
"epoch": 0.055,
"grad_norm": 0.26373944955124323,
"learning_rate": 1e-05,
"loss": 0.1104,
"step": 110
},
{
"epoch": 0.0555,
"grad_norm": 0.2201267469286863,
"learning_rate": 1e-05,
"loss": 0.0864,
"step": 111
},
{
"epoch": 0.056,
"grad_norm": 0.2584324977531668,
"learning_rate": 1e-05,
"loss": 0.1243,
"step": 112
},
{
"epoch": 0.0565,
"grad_norm": 0.21251509346212935,
"learning_rate": 1e-05,
"loss": 0.1287,
"step": 113
},
{
"epoch": 0.057,
"grad_norm": 0.2391921924682281,
"learning_rate": 1e-05,
"loss": 0.1174,
"step": 114
},
{
"epoch": 0.0575,
"grad_norm": 0.2250627442441596,
"learning_rate": 1e-05,
"loss": 0.0961,
"step": 115
},
{
"epoch": 0.058,
"grad_norm": 0.21589217619835932,
"learning_rate": 1e-05,
"loss": 0.1041,
"step": 116
},
{
"epoch": 0.0585,
"grad_norm": 0.23947622053978754,
"learning_rate": 1e-05,
"loss": 0.1027,
"step": 117
},
{
"epoch": 0.059,
"grad_norm": 0.20944871960722214,
"learning_rate": 1e-05,
"loss": 0.1117,
"step": 118
},
{
"epoch": 0.0595,
"grad_norm": 0.2025724984657677,
"learning_rate": 1e-05,
"loss": 0.1102,
"step": 119
},
{
"epoch": 0.06,
"grad_norm": 0.1801908473330023,
"learning_rate": 1e-05,
"loss": 0.1072,
"step": 120
},
{
"epoch": 0.0605,
"grad_norm": 0.1940191661946107,
"learning_rate": 1e-05,
"loss": 0.0904,
"step": 121
},
{
"epoch": 0.061,
"grad_norm": 0.21867530545592728,
"learning_rate": 1e-05,
"loss": 0.1193,
"step": 122
},
{
"epoch": 0.0615,
"grad_norm": 0.22225462024793874,
"learning_rate": 1e-05,
"loss": 0.0782,
"step": 123
},
{
"epoch": 0.062,
"grad_norm": 0.24241920616546134,
"learning_rate": 1e-05,
"loss": 0.1099,
"step": 124
},
{
"epoch": 0.0625,
"grad_norm": 0.21360306106975577,
"learning_rate": 1e-05,
"loss": 0.091,
"step": 125
},
{
"epoch": 0.063,
"grad_norm": 0.2007423950283759,
"learning_rate": 1e-05,
"loss": 0.1152,
"step": 126
},
{
"epoch": 0.0635,
"grad_norm": 0.19343038731295426,
"learning_rate": 1e-05,
"loss": 0.1181,
"step": 127
},
{
"epoch": 0.064,
"grad_norm": 0.234786663580031,
"learning_rate": 1e-05,
"loss": 0.1149,
"step": 128
},
{
"epoch": 0.064,
"eval_dev_acc": 0.548828125,
"eval_dev_token": 5586.20703125,
"eval_runtime": 364.0015,
"eval_samples_per_second": 0.176,
"eval_steps_per_second": 0.003,
"step": 128
},
{
"epoch": 0.0645,
"grad_norm": 0.20468481909378916,
"learning_rate": 1e-05,
"loss": 0.0976,
"step": 129
},
{
"epoch": 0.065,
"grad_norm": 0.25320635754138643,
"learning_rate": 1e-05,
"loss": 0.1183,
"step": 130
},
{
"epoch": 0.0655,
"grad_norm": 0.21530698126365438,
"learning_rate": 1e-05,
"loss": 0.1156,
"step": 131
},
{
"epoch": 0.066,
"grad_norm": 0.20489101859011527,
"learning_rate": 1e-05,
"loss": 0.0991,
"step": 132
},
{
"epoch": 0.0665,
"grad_norm": 0.21113632835377186,
"learning_rate": 1e-05,
"loss": 0.0838,
"step": 133
},
{
"epoch": 0.067,
"grad_norm": 0.18341595697478763,
"learning_rate": 1e-05,
"loss": 0.1036,
"step": 134
},
{
"epoch": 0.0675,
"grad_norm": 0.1762785527366556,
"learning_rate": 1e-05,
"loss": 0.1674,
"step": 135
},
{
"epoch": 0.068,
"grad_norm": 0.2632947258005063,
"learning_rate": 1e-05,
"loss": 0.1202,
"step": 136
},
{
"epoch": 0.0685,
"grad_norm": 0.21085919863317307,
"learning_rate": 1e-05,
"loss": 0.1131,
"step": 137
},
{
"epoch": 0.069,
"grad_norm": 0.19457697084640746,
"learning_rate": 1e-05,
"loss": 0.1005,
"step": 138
},
{
"epoch": 0.0695,
"grad_norm": 0.17119664823512107,
"learning_rate": 1e-05,
"loss": 0.1389,
"step": 139
},
{
"epoch": 0.07,
"grad_norm": 0.19084161070023772,
"learning_rate": 1e-05,
"loss": 0.1527,
"step": 140
},
{
"epoch": 0.0705,
"grad_norm": 0.19580784724822164,
"learning_rate": 1e-05,
"loss": 0.1175,
"step": 141
},
{
"epoch": 0.071,
"grad_norm": 0.2110588281462844,
"learning_rate": 1e-05,
"loss": 0.1062,
"step": 142
},
{
"epoch": 0.0715,
"grad_norm": 0.20012341575489243,
"learning_rate": 1e-05,
"loss": 0.0848,
"step": 143
},
{
"epoch": 0.072,
"grad_norm": 0.1940479696118561,
"learning_rate": 1e-05,
"loss": 0.0694,
"step": 144
},
{
"epoch": 0.0725,
"grad_norm": 0.21593290579494073,
"learning_rate": 1e-05,
"loss": 0.0766,
"step": 145
},
{
"epoch": 0.073,
"grad_norm": 0.22638726501654005,
"learning_rate": 1e-05,
"loss": 0.1084,
"step": 146
},
{
"epoch": 0.0735,
"grad_norm": 0.20997037800742063,
"learning_rate": 1e-05,
"loss": 0.0761,
"step": 147
},
{
"epoch": 0.074,
"grad_norm": 0.2380179005894331,
"learning_rate": 1e-05,
"loss": 0.0927,
"step": 148
},
{
"epoch": 0.0745,
"grad_norm": 0.23889516090857615,
"learning_rate": 1e-05,
"loss": 0.0827,
"step": 149
},
{
"epoch": 0.075,
"grad_norm": 0.18767850358859223,
"learning_rate": 1e-05,
"loss": 0.0881,
"step": 150
},
{
"epoch": 0.0755,
"grad_norm": 0.19463069265385494,
"learning_rate": 1e-05,
"loss": 0.0917,
"step": 151
},
{
"epoch": 0.076,
"grad_norm": 0.212834963744102,
"learning_rate": 1e-05,
"loss": 0.0756,
"step": 152
},
{
"epoch": 0.0765,
"grad_norm": 0.20584172308777918,
"learning_rate": 1e-05,
"loss": 0.0762,
"step": 153
},
{
"epoch": 0.077,
"grad_norm": 0.2371449651260928,
"learning_rate": 1e-05,
"loss": 0.0978,
"step": 154
},
{
"epoch": 0.0775,
"grad_norm": 0.2049083024101962,
"learning_rate": 1e-05,
"loss": 0.0925,
"step": 155
},
{
"epoch": 0.078,
"grad_norm": 0.21217273061342656,
"learning_rate": 1e-05,
"loss": 0.0825,
"step": 156
},
{
"epoch": 0.0785,
"grad_norm": 0.20105825530151383,
"learning_rate": 1e-05,
"loss": 0.0858,
"step": 157
},
{
"epoch": 0.079,
"grad_norm": 0.2257052714675071,
"learning_rate": 1e-05,
"loss": 0.1105,
"step": 158
},
{
"epoch": 0.0795,
"grad_norm": 0.20210858652943217,
"learning_rate": 1e-05,
"loss": 0.1022,
"step": 159
},
{
"epoch": 0.08,
"grad_norm": 0.19482796495242663,
"learning_rate": 1e-05,
"loss": 0.1262,
"step": 160
},
{
"epoch": 0.08,
"eval_dev_acc": 0.6015625,
"eval_dev_token": 5543.318359375,
"eval_runtime": 353.7068,
"eval_samples_per_second": 0.181,
"eval_steps_per_second": 0.003,
"step": 160
},
{
"epoch": 0.0805,
"grad_norm": 0.22955090034416561,
"learning_rate": 1e-05,
"loss": 0.0988,
"step": 161
},
{
"epoch": 0.081,
"grad_norm": 0.23547588568917174,
"learning_rate": 1e-05,
"loss": 0.0803,
"step": 162
},
{
"epoch": 0.0815,
"grad_norm": 0.230658904399123,
"learning_rate": 1e-05,
"loss": 0.0881,
"step": 163
},
{
"epoch": 0.082,
"grad_norm": 0.2595571094242936,
"learning_rate": 1e-05,
"loss": 0.0881,
"step": 164
},
{
"epoch": 0.0825,
"grad_norm": 0.26763452927239884,
"learning_rate": 1e-05,
"loss": 0.114,
"step": 165
},
{
"epoch": 0.083,
"grad_norm": 0.257549186353109,
"learning_rate": 1e-05,
"loss": 0.1045,
"step": 166
},
{
"epoch": 0.0835,
"grad_norm": 0.19712751160118708,
"learning_rate": 1e-05,
"loss": 0.0899,
"step": 167
},
{
"epoch": 0.084,
"grad_norm": 0.17276675854807147,
"learning_rate": 1e-05,
"loss": 0.0684,
"step": 168
},
{
"epoch": 0.0845,
"grad_norm": 0.21103265575626073,
"learning_rate": 1e-05,
"loss": 0.0821,
"step": 169
},
{
"epoch": 0.085,
"grad_norm": 0.22292947141761962,
"learning_rate": 1e-05,
"loss": 0.0983,
"step": 170
},
{
"epoch": 0.0855,
"grad_norm": 0.21612821069411284,
"learning_rate": 1e-05,
"loss": 0.0917,
"step": 171
},
{
"epoch": 0.086,
"grad_norm": 0.20757781370778242,
"learning_rate": 1e-05,
"loss": 0.0919,
"step": 172
},
{
"epoch": 0.0865,
"grad_norm": 0.2054200483785948,
"learning_rate": 1e-05,
"loss": 0.077,
"step": 173
},
{
"epoch": 0.087,
"grad_norm": 0.24143931624393172,
"learning_rate": 1e-05,
"loss": 0.0919,
"step": 174
},
{
"epoch": 0.0875,
"grad_norm": 0.22708368965968964,
"learning_rate": 1e-05,
"loss": 0.0931,
"step": 175
},
{
"epoch": 0.088,
"grad_norm": 0.20044838709826737,
"learning_rate": 1e-05,
"loss": 0.0808,
"step": 176
},
{
"epoch": 0.0885,
"grad_norm": 0.20148090317828546,
"learning_rate": 1e-05,
"loss": 0.0905,
"step": 177
},
{
"epoch": 0.089,
"grad_norm": 0.2090109676571514,
"learning_rate": 1e-05,
"loss": 0.0801,
"step": 178
},
{
"epoch": 0.0895,
"grad_norm": 0.19147542578517765,
"learning_rate": 1e-05,
"loss": 0.0774,
"step": 179
},
{
"epoch": 0.09,
"grad_norm": 0.22051053694604383,
"learning_rate": 1e-05,
"loss": 0.0949,
"step": 180
},
{
"epoch": 0.0905,
"grad_norm": 0.20152485004966214,
"learning_rate": 1e-05,
"loss": 0.077,
"step": 181
},
{
"epoch": 0.091,
"grad_norm": 0.18668034234550573,
"learning_rate": 1e-05,
"loss": 0.0887,
"step": 182
},
{
"epoch": 0.0915,
"grad_norm": 0.21497572062563422,
"learning_rate": 1e-05,
"loss": 0.095,
"step": 183
},
{
"epoch": 0.092,
"grad_norm": 0.21716742542098177,
"learning_rate": 1e-05,
"loss": 0.0822,
"step": 184
},
{
"epoch": 0.0925,
"grad_norm": 0.20262525326126424,
"learning_rate": 1e-05,
"loss": 0.0804,
"step": 185
},
{
"epoch": 0.093,
"grad_norm": 0.18652482669251277,
"learning_rate": 1e-05,
"loss": 0.0959,
"step": 186
},
{
"epoch": 0.0935,
"grad_norm": 0.232270946947485,
"learning_rate": 1e-05,
"loss": 0.0785,
"step": 187
},
{
"epoch": 0.094,
"grad_norm": 0.22559277541611453,
"learning_rate": 1e-05,
"loss": 0.0868,
"step": 188
},
{
"epoch": 0.0945,
"grad_norm": 0.21772738934026295,
"learning_rate": 1e-05,
"loss": 0.1384,
"step": 189
},
{
"epoch": 0.095,
"grad_norm": 0.19366625753900965,
"learning_rate": 1e-05,
"loss": 0.0962,
"step": 190
},
{
"epoch": 0.0955,
"grad_norm": 0.2162137483161777,
"learning_rate": 1e-05,
"loss": 0.0753,
"step": 191
},
{
"epoch": 0.096,
"grad_norm": 0.2111612755929646,
"learning_rate": 1e-05,
"loss": 0.0776,
"step": 192
},
{
"epoch": 0.096,
"eval_dev_acc": 0.546875,
"eval_dev_token": 5439.14453125,
"eval_runtime": 358.6784,
"eval_samples_per_second": 0.178,
"eval_steps_per_second": 0.003,
"step": 192
},
{
"epoch": 0.0965,
"grad_norm": 0.20427587800007568,
"learning_rate": 1e-05,
"loss": 0.1026,
"step": 193
},
{
"epoch": 0.097,
"grad_norm": 0.1913558266102393,
"learning_rate": 1e-05,
"loss": 0.0947,
"step": 194
},
{
"epoch": 0.0975,
"grad_norm": 0.21823580107793827,
"learning_rate": 1e-05,
"loss": 0.0852,
"step": 195
},
{
"epoch": 0.098,
"grad_norm": 0.21947391378003933,
"learning_rate": 1e-05,
"loss": 0.0834,
"step": 196
},
{
"epoch": 0.0985,
"grad_norm": 0.2234108965736748,
"learning_rate": 1e-05,
"loss": 0.0922,
"step": 197
},
{
"epoch": 0.099,
"grad_norm": 0.20738196532743278,
"learning_rate": 1e-05,
"loss": 0.0814,
"step": 198
},
{
"epoch": 0.0995,
"grad_norm": 0.21586171895392783,
"learning_rate": 1e-05,
"loss": 0.0988,
"step": 199
},
{
"epoch": 0.1,
"grad_norm": 0.20785742252782521,
"learning_rate": 1e-05,
"loss": 0.0872,
"step": 200
},
{
"epoch": 0.1005,
"grad_norm": 0.22570317690675268,
"learning_rate": 1e-05,
"loss": 0.0937,
"step": 201
},
{
"epoch": 0.101,
"grad_norm": 0.19453877023547578,
"learning_rate": 1e-05,
"loss": 0.1083,
"step": 202
},
{
"epoch": 0.1015,
"grad_norm": 0.20591293588894416,
"learning_rate": 1e-05,
"loss": 0.0767,
"step": 203
},
{
"epoch": 0.102,
"grad_norm": 0.1798554117116506,
"learning_rate": 1e-05,
"loss": 0.1017,
"step": 204
},
{
"epoch": 0.1025,
"grad_norm": 0.1927231622743242,
"learning_rate": 1e-05,
"loss": 0.0911,
"step": 205
},
{
"epoch": 0.103,
"grad_norm": 0.21444122564196544,
"learning_rate": 1e-05,
"loss": 0.0856,
"step": 206
},
{
"epoch": 0.1035,
"grad_norm": 0.17259842125017608,
"learning_rate": 1e-05,
"loss": 0.0904,
"step": 207
},
{
"epoch": 0.104,
"grad_norm": 0.23979518316140722,
"learning_rate": 1e-05,
"loss": 0.084,
"step": 208
},
{
"epoch": 0.1045,
"grad_norm": 0.22442151254111703,
"learning_rate": 1e-05,
"loss": 0.0949,
"step": 209
},
{
"epoch": 0.105,
"grad_norm": 0.19615294291751353,
"learning_rate": 1e-05,
"loss": 0.0864,
"step": 210
},
{
"epoch": 0.1055,
"grad_norm": 0.18344154651920094,
"learning_rate": 1e-05,
"loss": 0.0843,
"step": 211
},
{
"epoch": 0.106,
"grad_norm": 0.21335203803361255,
"learning_rate": 1e-05,
"loss": 0.077,
"step": 212
},
{
"epoch": 0.1065,
"grad_norm": 0.252518444172673,
"learning_rate": 1e-05,
"loss": 0.0888,
"step": 213
},
{
"epoch": 0.107,
"grad_norm": 0.2116629068211744,
"learning_rate": 1e-05,
"loss": 0.098,
"step": 214
},
{
"epoch": 0.1075,
"grad_norm": 0.21920052784491295,
"learning_rate": 1e-05,
"loss": 0.0823,
"step": 215
},
{
"epoch": 0.108,
"grad_norm": 0.19779399311082105,
"learning_rate": 1e-05,
"loss": 0.0752,
"step": 216
},
{
"epoch": 0.1085,
"grad_norm": 0.19973286372655655,
"learning_rate": 1e-05,
"loss": 0.074,
"step": 217
},
{
"epoch": 0.109,
"grad_norm": 0.22343594632856933,
"learning_rate": 1e-05,
"loss": 0.0907,
"step": 218
},
{
"epoch": 0.1095,
"grad_norm": 0.25396426928555105,
"learning_rate": 1e-05,
"loss": 0.1075,
"step": 219
},
{
"epoch": 0.11,
"grad_norm": 0.1945715284952783,
"learning_rate": 1e-05,
"loss": 0.1053,
"step": 220
},
{
"epoch": 0.1105,
"grad_norm": 0.21559958220587308,
"learning_rate": 1e-05,
"loss": 0.0865,
"step": 221
},
{
"epoch": 0.111,
"grad_norm": 0.22369755043282374,
"learning_rate": 1e-05,
"loss": 0.0841,
"step": 222
},
{
"epoch": 0.1115,
"grad_norm": 0.2097379198995065,
"learning_rate": 1e-05,
"loss": 0.0753,
"step": 223
},
{
"epoch": 0.112,
"grad_norm": 0.17224880947033328,
"learning_rate": 1e-05,
"loss": 0.0699,
"step": 224
},
{
"epoch": 0.112,
"eval_dev_acc": 0.560546875,
"eval_dev_token": 5619.009765625,
"eval_runtime": 357.3428,
"eval_samples_per_second": 0.179,
"eval_steps_per_second": 0.003,
"step": 224
},
{
"epoch": 0.1125,
"grad_norm": 0.19442868536212735,
"learning_rate": 1e-05,
"loss": 0.0846,
"step": 225
},
{
"epoch": 0.113,
"grad_norm": 0.1573515275602218,
"learning_rate": 1e-05,
"loss": 0.0634,
"step": 226
},
{
"epoch": 0.1135,
"grad_norm": 0.17680777096637068,
"learning_rate": 1e-05,
"loss": 0.0761,
"step": 227
},
{
"epoch": 0.114,
"grad_norm": 0.20935704393341403,
"learning_rate": 1e-05,
"loss": 0.0549,
"step": 228
},
{
"epoch": 0.1145,
"grad_norm": 0.19829321187204563,
"learning_rate": 1e-05,
"loss": 0.05,
"step": 229
},
{
"epoch": 0.115,
"grad_norm": 0.18833561824346334,
"learning_rate": 1e-05,
"loss": 0.0656,
"step": 230
},
{
"epoch": 0.1155,
"grad_norm": 0.17277292328026173,
"learning_rate": 1e-05,
"loss": 0.08,
"step": 231
},
{
"epoch": 0.116,
"grad_norm": 0.2083709354078263,
"learning_rate": 1e-05,
"loss": 0.0628,
"step": 232
},
{
"epoch": 0.1165,
"grad_norm": 0.18113171413223286,
"learning_rate": 1e-05,
"loss": 0.0633,
"step": 233
},
{
"epoch": 0.117,
"grad_norm": 0.19985236934976783,
"learning_rate": 1e-05,
"loss": 0.0676,
"step": 234
},
{
"epoch": 0.1175,
"grad_norm": 0.2023196182410912,
"learning_rate": 1e-05,
"loss": 0.062,
"step": 235
},
{
"epoch": 0.118,
"grad_norm": 0.19446408540605106,
"learning_rate": 1e-05,
"loss": 0.0609,
"step": 236
},
{
"epoch": 0.1185,
"grad_norm": 0.1879635955015942,
"learning_rate": 1e-05,
"loss": 0.0631,
"step": 237
},
{
"epoch": 0.119,
"grad_norm": 0.18000098291861588,
"learning_rate": 1e-05,
"loss": 0.0628,
"step": 238
},
{
"epoch": 0.1195,
"grad_norm": 0.2091766063587954,
"learning_rate": 1e-05,
"loss": 0.0642,
"step": 239
},
{
"epoch": 0.12,
"grad_norm": 0.19781458462773657,
"learning_rate": 1e-05,
"loss": 0.0551,
"step": 240
},
{
"epoch": 0.1205,
"grad_norm": 0.20570535681429145,
"learning_rate": 1e-05,
"loss": 0.062,
"step": 241
},
{
"epoch": 0.121,
"grad_norm": 0.17479813291913535,
"learning_rate": 1e-05,
"loss": 0.0657,
"step": 242
},
{
"epoch": 0.1215,
"grad_norm": 0.20023576215167263,
"learning_rate": 1e-05,
"loss": 0.0545,
"step": 243
},
{
"epoch": 0.122,
"grad_norm": 0.21569894284947272,
"learning_rate": 1e-05,
"loss": 0.0641,
"step": 244
},
{
"epoch": 0.1225,
"grad_norm": 0.16426082027771785,
"learning_rate": 1e-05,
"loss": 0.0488,
"step": 245
},
{
"epoch": 0.123,
"grad_norm": 0.23142807048539513,
"learning_rate": 1e-05,
"loss": 0.0617,
"step": 246
},
{
"epoch": 0.1235,
"grad_norm": 0.21054969399806525,
"learning_rate": 1e-05,
"loss": 0.0566,
"step": 247
},
{
"epoch": 0.124,
"grad_norm": 0.1533567582820314,
"learning_rate": 1e-05,
"loss": 0.0559,
"step": 248
},
{
"epoch": 0.1245,
"grad_norm": 0.19025683614022437,
"learning_rate": 1e-05,
"loss": 0.051,
"step": 249
},
{
"epoch": 0.125,
"grad_norm": 0.16411772241541067,
"learning_rate": 1e-05,
"loss": 0.068,
"step": 250
},
{
"epoch": 0.1255,
"grad_norm": 0.23699209914417435,
"learning_rate": 1e-05,
"loss": 0.0845,
"step": 251
},
{
"epoch": 0.126,
"grad_norm": 0.2047456890167515,
"learning_rate": 1e-05,
"loss": 0.0588,
"step": 252
},
{
"epoch": 0.1265,
"grad_norm": 0.20625503133501016,
"learning_rate": 1e-05,
"loss": 0.0543,
"step": 253
},
{
"epoch": 0.127,
"grad_norm": 0.16081505489801892,
"learning_rate": 1e-05,
"loss": 0.0413,
"step": 254
},
{
"epoch": 0.1275,
"grad_norm": 0.20221086748641462,
"learning_rate": 1e-05,
"loss": 0.0704,
"step": 255
},
{
"epoch": 0.128,
"grad_norm": 0.20711399972324054,
"learning_rate": 1e-05,
"loss": 0.0563,
"step": 256
},
{
"epoch": 0.128,
"eval_dev_acc": 0.587890625,
"eval_dev_token": 5704.3125,
"eval_runtime": 356.7726,
"eval_samples_per_second": 0.179,
"eval_steps_per_second": 0.003,
"step": 256
},
{
"epoch": 0.1285,
"grad_norm": 0.18942704583355485,
"learning_rate": 1e-05,
"loss": 0.0692,
"step": 257
},
{
"epoch": 0.129,
"grad_norm": 0.19171984906136896,
"learning_rate": 1e-05,
"loss": 0.05,
"step": 258
},
{
"epoch": 0.1295,
"grad_norm": 0.2125382567332232,
"learning_rate": 1e-05,
"loss": 0.0615,
"step": 259
},
{
"epoch": 0.13,
"grad_norm": 0.17877261040661208,
"learning_rate": 1e-05,
"loss": 0.0597,
"step": 260
},
{
"epoch": 0.1305,
"grad_norm": 0.1708210973205684,
"learning_rate": 1e-05,
"loss": 0.067,
"step": 261
},
{
"epoch": 0.131,
"grad_norm": 0.1850887764718648,
"learning_rate": 1e-05,
"loss": 0.0544,
"step": 262
},
{
"epoch": 0.1315,
"grad_norm": 0.2010060822058417,
"learning_rate": 1e-05,
"loss": 0.0696,
"step": 263
},
{
"epoch": 0.132,
"grad_norm": 0.18460835555899294,
"learning_rate": 1e-05,
"loss": 0.0607,
"step": 264
},
{
"epoch": 0.1325,
"grad_norm": 0.2264686856441524,
"learning_rate": 1e-05,
"loss": 0.0638,
"step": 265
},
{
"epoch": 0.133,
"grad_norm": 0.17252712599660533,
"learning_rate": 1e-05,
"loss": 0.055,
"step": 266
},
{
"epoch": 0.1335,
"grad_norm": 0.1668268871760919,
"learning_rate": 1e-05,
"loss": 0.0467,
"step": 267
},
{
"epoch": 0.134,
"grad_norm": 0.17711472127782535,
"learning_rate": 1e-05,
"loss": 0.0462,
"step": 268
},
{
"epoch": 0.1345,
"grad_norm": 0.16354851889499628,
"learning_rate": 1e-05,
"loss": 0.0687,
"step": 269
},
{
"epoch": 0.135,
"grad_norm": 0.17844367901102645,
"learning_rate": 1e-05,
"loss": 0.0455,
"step": 270
},
{
"epoch": 0.1355,
"grad_norm": 0.19303024902618696,
"learning_rate": 1e-05,
"loss": 0.0565,
"step": 271
},
{
"epoch": 0.136,
"grad_norm": 0.19049740233006035,
"learning_rate": 1e-05,
"loss": 0.0575,
"step": 272
},
{
"epoch": 0.1365,
"grad_norm": 0.18444934835307936,
"learning_rate": 1e-05,
"loss": 0.0495,
"step": 273
},
{
"epoch": 0.137,
"grad_norm": 0.2029153556589725,
"learning_rate": 1e-05,
"loss": 0.0632,
"step": 274
},
{
"epoch": 0.1375,
"grad_norm": 0.17742276348080663,
"learning_rate": 1e-05,
"loss": 0.0591,
"step": 275
},
{
"epoch": 0.138,
"grad_norm": 0.2086941032177491,
"learning_rate": 1e-05,
"loss": 0.0453,
"step": 276
},
{
"epoch": 0.1385,
"grad_norm": 0.1599025673655474,
"learning_rate": 1e-05,
"loss": 0.0346,
"step": 277
},
{
"epoch": 0.139,
"grad_norm": 0.20223917188200294,
"learning_rate": 1e-05,
"loss": 0.0516,
"step": 278
},
{
"epoch": 0.1395,
"grad_norm": 0.168572629555483,
"learning_rate": 1e-05,
"loss": 0.0502,
"step": 279
},
{
"epoch": 0.14,
"grad_norm": 0.21316857087336016,
"learning_rate": 1e-05,
"loss": 0.0585,
"step": 280
},
{
"epoch": 0.1405,
"grad_norm": 0.1533009935920478,
"learning_rate": 1e-05,
"loss": 0.0732,
"step": 281
},
{
"epoch": 0.141,
"grad_norm": 0.2226592493652288,
"learning_rate": 1e-05,
"loss": 0.0683,
"step": 282
},
{
"epoch": 0.1415,
"grad_norm": 0.2005052634299014,
"learning_rate": 1e-05,
"loss": 0.064,
"step": 283
},
{
"epoch": 0.142,
"grad_norm": 0.16977898059181232,
"learning_rate": 1e-05,
"loss": 0.0759,
"step": 284
},
{
"epoch": 0.1425,
"grad_norm": 0.17622048198257903,
"learning_rate": 1e-05,
"loss": 0.0546,
"step": 285
},
{
"epoch": 0.143,
"grad_norm": 0.15734345344681852,
"learning_rate": 1e-05,
"loss": 0.0371,
"step": 286
},
{
"epoch": 0.1435,
"grad_norm": 0.16097527322211574,
"learning_rate": 1e-05,
"loss": 0.0523,
"step": 287
},
{
"epoch": 0.144,
"grad_norm": 0.18490744056020517,
"learning_rate": 1e-05,
"loss": 0.0518,
"step": 288
},
{
"epoch": 0.144,
"eval_dev_acc": 0.5234375,
"eval_dev_token": 5945.451171875,
"eval_runtime": 402.5197,
"eval_samples_per_second": 0.159,
"eval_steps_per_second": 0.002,
"step": 288
},
{
"epoch": 0.1445,
"grad_norm": 0.16803932674344507,
"learning_rate": 1e-05,
"loss": 0.0564,
"step": 289
},
{
"epoch": 0.145,
"grad_norm": 0.18279683357896828,
"learning_rate": 1e-05,
"loss": 0.0604,
"step": 290
},
{
"epoch": 0.1455,
"grad_norm": 0.1773550481655194,
"learning_rate": 1e-05,
"loss": 0.0465,
"step": 291
},
{
"epoch": 0.146,
"grad_norm": 0.37855180864427374,
"learning_rate": 1e-05,
"loss": 0.0687,
"step": 292
},
{
"epoch": 0.1465,
"grad_norm": 0.2256173969538139,
"learning_rate": 1e-05,
"loss": 0.0597,
"step": 293
},
{
"epoch": 0.147,
"grad_norm": 0.19318315047474646,
"learning_rate": 1e-05,
"loss": 0.05,
"step": 294
},
{
"epoch": 0.1475,
"grad_norm": 0.1949573691249157,
"learning_rate": 1e-05,
"loss": 0.054,
"step": 295
},
{
"epoch": 0.148,
"grad_norm": 0.1784893290182381,
"learning_rate": 1e-05,
"loss": 0.0551,
"step": 296
},
{
"epoch": 0.1485,
"grad_norm": 0.1893518286787237,
"learning_rate": 1e-05,
"loss": 0.0582,
"step": 297
},
{
"epoch": 0.149,
"grad_norm": 0.16491423015511872,
"learning_rate": 1e-05,
"loss": 0.0393,
"step": 298
},
{
"epoch": 0.1495,
"grad_norm": 0.1762274160177828,
"learning_rate": 1e-05,
"loss": 0.0485,
"step": 299
},
{
"epoch": 0.15,
"grad_norm": 0.18398727773770782,
"learning_rate": 1e-05,
"loss": 0.0573,
"step": 300
},
{
"epoch": 0.1505,
"grad_norm": 0.18217281656040227,
"learning_rate": 1e-05,
"loss": 0.0485,
"step": 301
},
{
"epoch": 0.151,
"grad_norm": 0.17276763991718358,
"learning_rate": 1e-05,
"loss": 0.0571,
"step": 302
},
{
"epoch": 0.1515,
"grad_norm": 0.15668609882813492,
"learning_rate": 1e-05,
"loss": 0.0741,
"step": 303
},
{
"epoch": 0.152,
"grad_norm": 0.18965888700513778,
"learning_rate": 1e-05,
"loss": 0.0548,
"step": 304
},
{
"epoch": 0.1525,
"grad_norm": 0.1514861962476675,
"learning_rate": 1e-05,
"loss": 0.0408,
"step": 305
},
{
"epoch": 0.153,
"grad_norm": 0.21027930287961952,
"learning_rate": 1e-05,
"loss": 0.0473,
"step": 306
},
{
"epoch": 0.1535,
"grad_norm": 0.20086505297048218,
"learning_rate": 1e-05,
"loss": 0.0457,
"step": 307
},
{
"epoch": 0.154,
"grad_norm": 0.1834058060370301,
"learning_rate": 1e-05,
"loss": 0.0391,
"step": 308
},
{
"epoch": 0.1545,
"grad_norm": 0.1675035648173745,
"learning_rate": 1e-05,
"loss": 0.0346,
"step": 309
},
{
"epoch": 0.155,
"grad_norm": 0.19041217604042332,
"learning_rate": 1e-05,
"loss": 0.0447,
"step": 310
},
{
"epoch": 0.1555,
"grad_norm": 0.2063641120441124,
"learning_rate": 1e-05,
"loss": 0.0454,
"step": 311
},
{
"epoch": 0.156,
"grad_norm": 0.18205494058640856,
"learning_rate": 1e-05,
"loss": 0.0402,
"step": 312
},
{
"epoch": 0.1565,
"grad_norm": 0.18642039675473027,
"learning_rate": 1e-05,
"loss": 0.0397,
"step": 313
},
{
"epoch": 0.157,
"grad_norm": 0.16971154384699963,
"learning_rate": 1e-05,
"loss": 0.0426,
"step": 314
},
{
"epoch": 0.1575,
"grad_norm": 0.19779499600374534,
"learning_rate": 1e-05,
"loss": 0.042,
"step": 315
},
{
"epoch": 0.158,
"grad_norm": 0.1597506922805199,
"learning_rate": 1e-05,
"loss": 0.0561,
"step": 316
},
{
"epoch": 0.1585,
"grad_norm": 0.22455898943377722,
"learning_rate": 1e-05,
"loss": 0.0508,
"step": 317
},
{
"epoch": 0.159,
"grad_norm": 0.22816888070811367,
"learning_rate": 1e-05,
"loss": 0.0684,
"step": 318
},
{
"epoch": 0.1595,
"grad_norm": 0.20829314487846406,
"learning_rate": 1e-05,
"loss": 0.0558,
"step": 319
},
{
"epoch": 0.16,
"grad_norm": 0.1726376642892394,
"learning_rate": 1e-05,
"loss": 0.0382,
"step": 320
},
{
"epoch": 0.16,
"eval_dev_acc": 0.62109375,
"eval_dev_token": 5348.43359375,
"eval_runtime": 348.0694,
"eval_samples_per_second": 0.184,
"eval_steps_per_second": 0.003,
"step": 320
},
{
"epoch": 0.1605,
"grad_norm": 0.21533911997871472,
"learning_rate": 1e-05,
"loss": 0.0573,
"step": 321
},
{
"epoch": 0.161,
"grad_norm": 0.1945926926705701,
"learning_rate": 1e-05,
"loss": 0.0482,
"step": 322
},
{
"epoch": 0.1615,
"grad_norm": 0.16282045186801866,
"learning_rate": 1e-05,
"loss": 0.0395,
"step": 323
},
{
"epoch": 0.162,
"grad_norm": 0.21192465805738625,
"learning_rate": 1e-05,
"loss": 0.0534,
"step": 324
},
{
"epoch": 0.1625,
"grad_norm": 0.18412473332196624,
"learning_rate": 1e-05,
"loss": 0.0457,
"step": 325
},
{
"epoch": 0.163,
"grad_norm": 0.1760683716395308,
"learning_rate": 1e-05,
"loss": 0.0621,
"step": 326
},
{
"epoch": 0.1635,
"grad_norm": 0.1977120327808734,
"learning_rate": 1e-05,
"loss": 0.0443,
"step": 327
},
{
"epoch": 0.164,
"grad_norm": 0.17834928432327893,
"learning_rate": 1e-05,
"loss": 0.0474,
"step": 328
},
{
"epoch": 0.1645,
"grad_norm": 0.19810362807929732,
"learning_rate": 1e-05,
"loss": 0.0472,
"step": 329
},
{
"epoch": 0.165,
"grad_norm": 0.17306554655229037,
"learning_rate": 1e-05,
"loss": 0.0514,
"step": 330
},
{
"epoch": 0.1655,
"grad_norm": 0.1732660830513622,
"learning_rate": 1e-05,
"loss": 0.0487,
"step": 331
},
{
"epoch": 0.166,
"grad_norm": 0.15153749458139032,
"learning_rate": 1e-05,
"loss": 0.044,
"step": 332
},
{
"epoch": 0.1665,
"grad_norm": 0.20187085045913772,
"learning_rate": 1e-05,
"loss": 0.0551,
"step": 333
},
{
"epoch": 0.167,
"grad_norm": 0.16579582791974742,
"learning_rate": 1e-05,
"loss": 0.0497,
"step": 334
},
{
"epoch": 0.1675,
"grad_norm": 0.19316064563692958,
"learning_rate": 1e-05,
"loss": 0.0549,
"step": 335
},
{
"epoch": 0.168,
"grad_norm": 0.16491714800111232,
"learning_rate": 1e-05,
"loss": 0.0371,
"step": 336
},
{
"epoch": 0.1685,
"grad_norm": 0.17551519178449135,
"learning_rate": 1e-05,
"loss": 0.041,
"step": 337
},
{
"epoch": 0.169,
"grad_norm": 0.1734781492111704,
"learning_rate": 1e-05,
"loss": 0.0441,
"step": 338
},
{
"epoch": 0.1695,
"grad_norm": 0.18684315556716974,
"learning_rate": 1e-05,
"loss": 0.0488,
"step": 339
},
{
"epoch": 0.17,
"grad_norm": 0.16917376679576818,
"learning_rate": 1e-05,
"loss": 0.0437,
"step": 340
},
{
"epoch": 0.1705,
"grad_norm": 0.1501957333162884,
"learning_rate": 1e-05,
"loss": 0.0393,
"step": 341
},
{
"epoch": 0.171,
"grad_norm": 0.2205121105977978,
"learning_rate": 1e-05,
"loss": 0.0579,
"step": 342
},
{
"epoch": 0.1715,
"grad_norm": 0.18041963938373073,
"learning_rate": 1e-05,
"loss": 0.0492,
"step": 343
},
{
"epoch": 0.172,
"grad_norm": 0.17449103579952038,
"learning_rate": 1e-05,
"loss": 0.0401,
"step": 344
},
{
"epoch": 0.1725,
"grad_norm": 0.15528177663303303,
"learning_rate": 1e-05,
"loss": 0.0379,
"step": 345
},
{
"epoch": 0.173,
"grad_norm": 0.20473861699417553,
"learning_rate": 1e-05,
"loss": 0.0499,
"step": 346
},
{
"epoch": 0.1735,
"grad_norm": 0.22954222855804748,
"learning_rate": 1e-05,
"loss": 0.066,
"step": 347
},
{
"epoch": 0.174,
"grad_norm": 0.18144976711639327,
"learning_rate": 1e-05,
"loss": 0.0417,
"step": 348
},
{
"epoch": 0.1745,
"grad_norm": 0.1647885822020398,
"learning_rate": 1e-05,
"loss": 0.0406,
"step": 349
},
{
"epoch": 0.175,
"grad_norm": 0.13381930823784724,
"learning_rate": 1e-05,
"loss": 0.039,
"step": 350
},
{
"epoch": 0.1755,
"grad_norm": 0.17233840695200286,
"learning_rate": 1e-05,
"loss": 0.0504,
"step": 351
},
{
"epoch": 0.176,
"grad_norm": 0.18220085493590332,
"learning_rate": 1e-05,
"loss": 0.0411,
"step": 352
},
{
"epoch": 0.176,
"eval_dev_acc": 0.53515625,
"eval_dev_token": 5787.59375,
"eval_runtime": 364.0085,
"eval_samples_per_second": 0.176,
"eval_steps_per_second": 0.003,
"step": 352
},
{
"epoch": 0.1765,
"grad_norm": 0.20744185724682074,
"learning_rate": 1e-05,
"loss": 0.0606,
"step": 353
},
{
"epoch": 0.177,
"grad_norm": 0.18600570998480834,
"learning_rate": 1e-05,
"loss": 0.0416,
"step": 354
},
{
"epoch": 0.1775,
"grad_norm": 0.1776469292641811,
"learning_rate": 1e-05,
"loss": 0.042,
"step": 355
},
{
"epoch": 0.178,
"grad_norm": 0.1783568782453835,
"learning_rate": 1e-05,
"loss": 0.0457,
"step": 356
},
{
"epoch": 0.1785,
"grad_norm": 0.1981729227656145,
"learning_rate": 1e-05,
"loss": 0.0578,
"step": 357
},
{
"epoch": 0.179,
"grad_norm": 0.18984703197303243,
"learning_rate": 1e-05,
"loss": 0.0427,
"step": 358
},
{
"epoch": 0.1795,
"grad_norm": 0.21799846739281004,
"learning_rate": 1e-05,
"loss": 0.0492,
"step": 359
},
{
"epoch": 0.18,
"grad_norm": 0.2263278306065525,
"learning_rate": 1e-05,
"loss": 0.0708,
"step": 360
},
{
"epoch": 0.1805,
"grad_norm": 0.16612125798758726,
"learning_rate": 1e-05,
"loss": 0.0447,
"step": 361
},
{
"epoch": 0.181,
"grad_norm": 0.15311766872023147,
"learning_rate": 1e-05,
"loss": 0.0369,
"step": 362
},
{
"epoch": 0.1815,
"grad_norm": 0.18614757917185834,
"learning_rate": 1e-05,
"loss": 0.0484,
"step": 363
},
{
"epoch": 0.182,
"grad_norm": 0.18253431030668557,
"learning_rate": 1e-05,
"loss": 0.044,
"step": 364
},
{
"epoch": 0.1825,
"grad_norm": 0.19238661256236803,
"learning_rate": 1e-05,
"loss": 0.0615,
"step": 365
},
{
"epoch": 0.183,
"grad_norm": 0.1808155264273284,
"learning_rate": 1e-05,
"loss": 0.0444,
"step": 366
},
{
"epoch": 0.1835,
"grad_norm": 0.1743762662626829,
"learning_rate": 1e-05,
"loss": 0.0501,
"step": 367
},
{
"epoch": 0.184,
"grad_norm": 0.16508774246157967,
"learning_rate": 1e-05,
"loss": 0.0448,
"step": 368
},
{
"epoch": 0.1845,
"grad_norm": 0.1546243168773746,
"learning_rate": 1e-05,
"loss": 0.0366,
"step": 369
},
{
"epoch": 0.185,
"grad_norm": 0.1746189186464954,
"learning_rate": 1e-05,
"loss": 0.0471,
"step": 370
},
{
"epoch": 0.1855,
"grad_norm": 0.17995461422580256,
"learning_rate": 1e-05,
"loss": 0.0405,
"step": 371
},
{
"epoch": 0.186,
"grad_norm": 0.16745033647841967,
"learning_rate": 1e-05,
"loss": 0.0371,
"step": 372
},
{
"epoch": 0.1865,
"grad_norm": 0.14177227347565124,
"learning_rate": 1e-05,
"loss": 0.0336,
"step": 373
},
{
"epoch": 0.187,
"grad_norm": 0.19568633642105135,
"learning_rate": 1e-05,
"loss": 0.0419,
"step": 374
},
{
"epoch": 0.1875,
"grad_norm": 0.1694809590901385,
"learning_rate": 1e-05,
"loss": 0.0365,
"step": 375
},
{
"epoch": 0.188,
"grad_norm": 0.16086017791775223,
"learning_rate": 1e-05,
"loss": 0.0382,
"step": 376
},
{
"epoch": 0.1885,
"grad_norm": 0.14863922525565496,
"learning_rate": 1e-05,
"loss": 0.039,
"step": 377
},
{
"epoch": 0.189,
"grad_norm": 0.16084357735487792,
"learning_rate": 1e-05,
"loss": 0.0312,
"step": 378
},
{
"epoch": 0.1895,
"grad_norm": 0.19070881724879324,
"learning_rate": 1e-05,
"loss": 0.0491,
"step": 379
},
{
"epoch": 0.19,
"grad_norm": 0.17240390839318184,
"learning_rate": 1e-05,
"loss": 0.0455,
"step": 380
},
{
"epoch": 0.1905,
"grad_norm": 0.13331277326103189,
"learning_rate": 1e-05,
"loss": 0.0289,
"step": 381
},
{
"epoch": 0.191,
"grad_norm": 0.18756729894366522,
"learning_rate": 1e-05,
"loss": 0.0468,
"step": 382
},
{
"epoch": 0.1915,
"grad_norm": 0.1660248717735821,
"learning_rate": 1e-05,
"loss": 0.0424,
"step": 383
},
{
"epoch": 0.192,
"grad_norm": 0.16346974130070938,
"learning_rate": 1e-05,
"loss": 0.0311,
"step": 384
},
{
"epoch": 0.192,
"eval_dev_acc": 0.599609375,
"eval_dev_token": 5596.130859375,
"eval_runtime": 352.4793,
"eval_samples_per_second": 0.182,
"eval_steps_per_second": 0.003,
"step": 384
},
{
"epoch": 0.1925,
"grad_norm": 0.16858862798497806,
"learning_rate": 1e-05,
"loss": 0.0389,
"step": 385
},
{
"epoch": 0.193,
"grad_norm": 0.1484958580298565,
"learning_rate": 1e-05,
"loss": 0.0397,
"step": 386
},
{
"epoch": 0.1935,
"grad_norm": 0.17660261356555002,
"learning_rate": 1e-05,
"loss": 0.0515,
"step": 387
},
{
"epoch": 0.194,
"grad_norm": 0.1783517215939047,
"learning_rate": 1e-05,
"loss": 0.0431,
"step": 388
},
{
"epoch": 0.1945,
"grad_norm": 0.14136150090913457,
"learning_rate": 1e-05,
"loss": 0.0323,
"step": 389
},
{
"epoch": 0.195,
"grad_norm": 0.16595913921658337,
"learning_rate": 1e-05,
"loss": 0.0394,
"step": 390
},
{
"epoch": 0.1955,
"grad_norm": 0.17788297569443248,
"learning_rate": 1e-05,
"loss": 0.0698,
"step": 391
},
{
"epoch": 0.196,
"grad_norm": 0.14755167079389797,
"learning_rate": 1e-05,
"loss": 0.0308,
"step": 392
},
{
"epoch": 0.1965,
"grad_norm": 0.20681855290430337,
"learning_rate": 1e-05,
"loss": 0.0494,
"step": 393
},
{
"epoch": 0.197,
"grad_norm": 0.19060439020439998,
"learning_rate": 1e-05,
"loss": 0.0445,
"step": 394
},
{
"epoch": 0.1975,
"grad_norm": 0.17199443698076167,
"learning_rate": 1e-05,
"loss": 0.0414,
"step": 395
},
{
"epoch": 0.198,
"grad_norm": 0.15210077373082737,
"learning_rate": 1e-05,
"loss": 0.0296,
"step": 396
},
{
"epoch": 0.1985,
"grad_norm": 0.17482591540638856,
"learning_rate": 1e-05,
"loss": 0.044,
"step": 397
},
{
"epoch": 0.199,
"grad_norm": 0.15501601608099658,
"learning_rate": 1e-05,
"loss": 0.0376,
"step": 398
},
{
"epoch": 0.1995,
"grad_norm": 0.17142493205422682,
"learning_rate": 1e-05,
"loss": 0.0386,
"step": 399
},
{
"epoch": 0.2,
"grad_norm": 0.1921162644413309,
"learning_rate": 1e-05,
"loss": 0.0469,
"step": 400
},
{
"epoch": 0.2005,
"grad_norm": 0.15938080403417312,
"learning_rate": 1e-05,
"loss": 0.0496,
"step": 401
},
{
"epoch": 0.201,
"grad_norm": 0.14786848292294155,
"learning_rate": 1e-05,
"loss": 0.0426,
"step": 402
},
{
"epoch": 0.2015,
"grad_norm": 0.18628997533329272,
"learning_rate": 1e-05,
"loss": 0.0581,
"step": 403
},
{
"epoch": 0.202,
"grad_norm": 0.16058096254934043,
"learning_rate": 1e-05,
"loss": 0.0336,
"step": 404
},
{
"epoch": 0.2025,
"grad_norm": 0.19319024386507233,
"learning_rate": 1e-05,
"loss": 0.047,
"step": 405
},
{
"epoch": 0.203,
"grad_norm": 0.17328115011013,
"learning_rate": 1e-05,
"loss": 0.049,
"step": 406
},
{
"epoch": 0.2035,
"grad_norm": 0.13258378170371796,
"learning_rate": 1e-05,
"loss": 0.0286,
"step": 407
},
{
"epoch": 0.204,
"grad_norm": 0.17945245697241183,
"learning_rate": 1e-05,
"loss": 0.0518,
"step": 408
},
{
"epoch": 0.2045,
"grad_norm": 0.16689764407399071,
"learning_rate": 1e-05,
"loss": 0.0458,
"step": 409
},
{
"epoch": 0.205,
"grad_norm": 0.18446815699746041,
"learning_rate": 1e-05,
"loss": 0.0408,
"step": 410
},
{
"epoch": 0.2055,
"grad_norm": 0.1489326060726689,
"learning_rate": 1e-05,
"loss": 0.0656,
"step": 411
},
{
"epoch": 0.206,
"grad_norm": 0.14974593012017515,
"learning_rate": 1e-05,
"loss": 0.0297,
"step": 412
},
{
"epoch": 0.2065,
"grad_norm": 0.1918114395748189,
"learning_rate": 1e-05,
"loss": 0.0439,
"step": 413
},
{
"epoch": 0.207,
"grad_norm": 0.1689953495305046,
"learning_rate": 1e-05,
"loss": 0.04,
"step": 414
},
{
"epoch": 0.2075,
"grad_norm": 0.1403733317703667,
"learning_rate": 1e-05,
"loss": 0.0472,
"step": 415
},
{
"epoch": 0.208,
"grad_norm": 0.173982074128614,
"learning_rate": 1e-05,
"loss": 0.0395,
"step": 416
},
{
"epoch": 0.208,
"eval_dev_acc": 0.49609375,
"eval_dev_token": 5482.95703125,
"eval_runtime": 374.469,
"eval_samples_per_second": 0.171,
"eval_steps_per_second": 0.003,
"step": 416
},
{
"epoch": 0.2085,
"grad_norm": 0.16717306009352031,
"learning_rate": 1e-05,
"loss": 0.0369,
"step": 417
},
{
"epoch": 0.209,
"grad_norm": 0.17317803700896214,
"learning_rate": 1e-05,
"loss": 0.0581,
"step": 418
},
{
"epoch": 0.2095,
"grad_norm": 0.14729050118039705,
"learning_rate": 1e-05,
"loss": 0.0469,
"step": 419
},
{
"epoch": 0.21,
"grad_norm": 0.14599122830811173,
"learning_rate": 1e-05,
"loss": 0.0571,
"step": 420
},
{
"epoch": 0.2105,
"grad_norm": 0.16285142688584706,
"learning_rate": 1e-05,
"loss": 0.0291,
"step": 421
},
{
"epoch": 0.211,
"grad_norm": 0.19044973230329837,
"learning_rate": 1e-05,
"loss": 0.0599,
"step": 422
},
{
"epoch": 0.2115,
"grad_norm": 0.188861960333507,
"learning_rate": 1e-05,
"loss": 0.0471,
"step": 423
},
{
"epoch": 0.212,
"grad_norm": 0.19188548951756218,
"learning_rate": 1e-05,
"loss": 0.0529,
"step": 424
},
{
"epoch": 0.2125,
"grad_norm": 0.16267402517673002,
"learning_rate": 1e-05,
"loss": 0.0305,
"step": 425
},
{
"epoch": 0.213,
"grad_norm": 0.1447850696130614,
"learning_rate": 1e-05,
"loss": 0.0324,
"step": 426
},
{
"epoch": 0.2135,
"grad_norm": 0.15248164794588065,
"learning_rate": 1e-05,
"loss": 0.0388,
"step": 427
},
{
"epoch": 0.214,
"grad_norm": 0.1661241871100943,
"learning_rate": 1e-05,
"loss": 0.0328,
"step": 428
},
{
"epoch": 0.2145,
"grad_norm": 0.16566625624023265,
"learning_rate": 1e-05,
"loss": 0.031,
"step": 429
},
{
"epoch": 0.215,
"grad_norm": 0.15249287061514458,
"learning_rate": 1e-05,
"loss": 0.0319,
"step": 430
},
{
"epoch": 0.2155,
"grad_norm": 0.12995530917181783,
"learning_rate": 1e-05,
"loss": 0.0233,
"step": 431
},
{
"epoch": 0.216,
"grad_norm": 0.1704276552962093,
"learning_rate": 1e-05,
"loss": 0.0405,
"step": 432
},
{
"epoch": 0.2165,
"grad_norm": 0.17386329346754434,
"learning_rate": 1e-05,
"loss": 0.0336,
"step": 433
},
{
"epoch": 0.217,
"grad_norm": 0.15704760833763615,
"learning_rate": 1e-05,
"loss": 0.0325,
"step": 434
},
{
"epoch": 0.2175,
"grad_norm": 0.1495524799308763,
"learning_rate": 1e-05,
"loss": 0.0341,
"step": 435
},
{
"epoch": 0.218,
"grad_norm": 0.1686891909288217,
"learning_rate": 1e-05,
"loss": 0.0343,
"step": 436
},
{
"epoch": 0.2185,
"grad_norm": 0.13995459985426573,
"learning_rate": 1e-05,
"loss": 0.0398,
"step": 437
},
{
"epoch": 0.219,
"grad_norm": 0.15473569116081692,
"learning_rate": 1e-05,
"loss": 0.0412,
"step": 438
},
{
"epoch": 0.2195,
"grad_norm": 0.1801609077983992,
"learning_rate": 1e-05,
"loss": 0.0534,
"step": 439
},
{
"epoch": 0.22,
"grad_norm": 0.17809364795872226,
"learning_rate": 1e-05,
"loss": 0.0548,
"step": 440
},
{
"epoch": 0.2205,
"grad_norm": 0.1535032114151188,
"learning_rate": 1e-05,
"loss": 0.0593,
"step": 441
},
{
"epoch": 0.221,
"grad_norm": 0.20882248884544774,
"learning_rate": 1e-05,
"loss": 0.0402,
"step": 442
},
{
"epoch": 0.2215,
"grad_norm": 0.14517381058327564,
"learning_rate": 1e-05,
"loss": 0.0436,
"step": 443
},
{
"epoch": 0.222,
"grad_norm": 0.17014179155102424,
"learning_rate": 1e-05,
"loss": 0.0333,
"step": 444
},
{
"epoch": 0.2225,
"grad_norm": 0.1729306341614305,
"learning_rate": 1e-05,
"loss": 0.0301,
"step": 445
},
{
"epoch": 0.223,
"grad_norm": 0.1686712423851483,
"learning_rate": 1e-05,
"loss": 0.0406,
"step": 446
},
{
"epoch": 0.2235,
"grad_norm": 0.1535287640721648,
"learning_rate": 1e-05,
"loss": 0.0282,
"step": 447
},
{
"epoch": 0.224,
"grad_norm": 0.1406783148617548,
"learning_rate": 1e-05,
"loss": 0.0264,
"step": 448
},
{
"epoch": 0.224,
"eval_dev_acc": 0.576171875,
"eval_dev_token": 5738.91015625,
"eval_runtime": 360.4892,
"eval_samples_per_second": 0.178,
"eval_steps_per_second": 0.003,
"step": 448
},
{
"epoch": 0.2245,
"grad_norm": 0.15218760519443988,
"learning_rate": 1e-05,
"loss": 0.0227,
"step": 449
},
{
"epoch": 0.225,
"grad_norm": 0.13165646820927943,
"learning_rate": 1e-05,
"loss": 0.0244,
"step": 450
},
{
"epoch": 0.2255,
"grad_norm": 0.13658776890249372,
"learning_rate": 1e-05,
"loss": 0.0235,
"step": 451
},
{
"epoch": 0.226,
"grad_norm": 0.1595102880182028,
"learning_rate": 1e-05,
"loss": 0.0361,
"step": 452
},
{
"epoch": 0.2265,
"grad_norm": 0.18272272767076744,
"learning_rate": 1e-05,
"loss": 0.0341,
"step": 453
},
{
"epoch": 0.227,
"grad_norm": 0.15970250529787045,
"learning_rate": 1e-05,
"loss": 0.0277,
"step": 454
},
{
"epoch": 0.2275,
"grad_norm": 0.1641447638351716,
"learning_rate": 1e-05,
"loss": 0.0284,
"step": 455
},
{
"epoch": 0.228,
"grad_norm": 0.1294308434226962,
"learning_rate": 1e-05,
"loss": 0.022,
"step": 456
},
{
"epoch": 0.2285,
"grad_norm": 0.11954195360401737,
"learning_rate": 1e-05,
"loss": 0.0202,
"step": 457
},
{
"epoch": 0.229,
"grad_norm": 0.16068276912989043,
"learning_rate": 1e-05,
"loss": 0.0251,
"step": 458
},
{
"epoch": 0.2295,
"grad_norm": 0.15307414897001792,
"learning_rate": 1e-05,
"loss": 0.027,
"step": 459
},
{
"epoch": 0.23,
"grad_norm": 0.15979117725555442,
"learning_rate": 1e-05,
"loss": 0.0375,
"step": 460
},
{
"epoch": 0.2305,
"grad_norm": 0.2020059964338148,
"learning_rate": 1e-05,
"loss": 0.043,
"step": 461
},
{
"epoch": 0.231,
"grad_norm": 0.15074817260440432,
"learning_rate": 1e-05,
"loss": 0.0241,
"step": 462
},
{
"epoch": 0.2315,
"grad_norm": 0.16521962645814686,
"learning_rate": 1e-05,
"loss": 0.0447,
"step": 463
},
{
"epoch": 0.232,
"grad_norm": 0.14710027259702427,
"learning_rate": 1e-05,
"loss": 0.0322,
"step": 464
},
{
"epoch": 0.2325,
"grad_norm": 0.1879875645942318,
"learning_rate": 1e-05,
"loss": 0.0324,
"step": 465
},
{
"epoch": 0.233,
"grad_norm": 0.16624112738581265,
"learning_rate": 1e-05,
"loss": 0.0308,
"step": 466
},
{
"epoch": 0.2335,
"grad_norm": 0.19050875236463863,
"learning_rate": 1e-05,
"loss": 0.0342,
"step": 467
},
{
"epoch": 0.234,
"grad_norm": 0.11998525830480301,
"learning_rate": 1e-05,
"loss": 0.021,
"step": 468
},
{
"epoch": 0.2345,
"grad_norm": 0.1656701890014892,
"learning_rate": 1e-05,
"loss": 0.027,
"step": 469
},
{
"epoch": 0.235,
"grad_norm": 0.15307753326429366,
"learning_rate": 1e-05,
"loss": 0.0452,
"step": 470
},
{
"epoch": 0.2355,
"grad_norm": 0.1897260442883158,
"learning_rate": 1e-05,
"loss": 0.0245,
"step": 471
},
{
"epoch": 0.236,
"grad_norm": 0.11931107646228578,
"learning_rate": 1e-05,
"loss": 0.017,
"step": 472
},
{
"epoch": 0.2365,
"grad_norm": 0.18498950301005707,
"learning_rate": 1e-05,
"loss": 0.0342,
"step": 473
},
{
"epoch": 0.237,
"grad_norm": 0.14355239628818517,
"learning_rate": 1e-05,
"loss": 0.0221,
"step": 474
},
{
"epoch": 0.2375,
"grad_norm": 0.16525653932908532,
"learning_rate": 1e-05,
"loss": 0.029,
"step": 475
},
{
"epoch": 0.238,
"grad_norm": 0.1518153688638394,
"learning_rate": 1e-05,
"loss": 0.0267,
"step": 476
},
{
"epoch": 0.2385,
"grad_norm": 0.15987321641272437,
"learning_rate": 1e-05,
"loss": 0.0279,
"step": 477
},
{
"epoch": 0.239,
"grad_norm": 0.1442274823944727,
"learning_rate": 1e-05,
"loss": 0.0349,
"step": 478
},
{
"epoch": 0.2395,
"grad_norm": 0.11710766704672448,
"learning_rate": 1e-05,
"loss": 0.0179,
"step": 479
},
{
"epoch": 0.24,
"grad_norm": 0.15497604683020938,
"learning_rate": 1e-05,
"loss": 0.023,
"step": 480
},
{
"epoch": 0.24,
"eval_dev_acc": 0.544921875,
"eval_dev_token": 5840.61328125,
"eval_runtime": 373.4708,
"eval_samples_per_second": 0.171,
"eval_steps_per_second": 0.003,
"step": 480
},
{
"epoch": 0.2405,
"grad_norm": 0.13492229768745556,
"learning_rate": 1e-05,
"loss": 0.0205,
"step": 481
},
{
"epoch": 0.241,
"grad_norm": 0.1704648731314998,
"learning_rate": 1e-05,
"loss": 0.0353,
"step": 482
},
{
"epoch": 0.2415,
"grad_norm": 0.1491861836462168,
"learning_rate": 1e-05,
"loss": 0.0365,
"step": 483
},
{
"epoch": 0.242,
"grad_norm": 0.17050828891525746,
"learning_rate": 1e-05,
"loss": 0.0277,
"step": 484
},
{
"epoch": 0.2425,
"grad_norm": 0.17980691606220936,
"learning_rate": 1e-05,
"loss": 0.0335,
"step": 485
},
{
"epoch": 0.243,
"grad_norm": 0.16998825524724584,
"learning_rate": 1e-05,
"loss": 0.0362,
"step": 486
},
{
"epoch": 0.2435,
"grad_norm": 0.11641133365996917,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 487
},
{
"epoch": 0.244,
"grad_norm": 0.14362674831456992,
"learning_rate": 1e-05,
"loss": 0.0365,
"step": 488
},
{
"epoch": 0.2445,
"grad_norm": 0.14488123923452778,
"learning_rate": 1e-05,
"loss": 0.024,
"step": 489
},
{
"epoch": 0.245,
"grad_norm": 0.1517003378019991,
"learning_rate": 1e-05,
"loss": 0.0271,
"step": 490
},
{
"epoch": 0.2455,
"grad_norm": 0.14967074987714707,
"learning_rate": 1e-05,
"loss": 0.0294,
"step": 491
},
{
"epoch": 0.246,
"grad_norm": 0.15791993394836015,
"learning_rate": 1e-05,
"loss": 0.0283,
"step": 492
},
{
"epoch": 0.2465,
"grad_norm": 0.13495006239387555,
"learning_rate": 1e-05,
"loss": 0.0251,
"step": 493
},
{
"epoch": 0.247,
"grad_norm": 0.18930054102351096,
"learning_rate": 1e-05,
"loss": 0.0373,
"step": 494
},
{
"epoch": 0.2475,
"grad_norm": 0.13152234060084034,
"learning_rate": 1e-05,
"loss": 0.0233,
"step": 495
},
{
"epoch": 0.248,
"grad_norm": 0.1341531691510106,
"learning_rate": 1e-05,
"loss": 0.0269,
"step": 496
},
{
"epoch": 0.2485,
"grad_norm": 0.13741586371551992,
"learning_rate": 1e-05,
"loss": 0.0277,
"step": 497
},
{
"epoch": 0.249,
"grad_norm": 0.1554051684617337,
"learning_rate": 1e-05,
"loss": 0.0276,
"step": 498
},
{
"epoch": 0.2495,
"grad_norm": 0.14814577647609775,
"learning_rate": 1e-05,
"loss": 0.0235,
"step": 499
},
{
"epoch": 0.25,
"grad_norm": 0.14930720560226657,
"learning_rate": 1e-05,
"loss": 0.039,
"step": 500
},
{
"epoch": 0.2505,
"grad_norm": 0.1244942117603243,
"learning_rate": 1e-05,
"loss": 0.0202,
"step": 501
},
{
"epoch": 0.251,
"grad_norm": 0.14244145658079232,
"learning_rate": 1e-05,
"loss": 0.0209,
"step": 502
},
{
"epoch": 0.2515,
"grad_norm": 0.146145096145696,
"learning_rate": 1e-05,
"loss": 0.024,
"step": 503
},
{
"epoch": 0.252,
"grad_norm": 0.13594585715406687,
"learning_rate": 1e-05,
"loss": 0.0269,
"step": 504
},
{
"epoch": 0.2525,
"grad_norm": 0.1490412459954878,
"learning_rate": 1e-05,
"loss": 0.0345,
"step": 505
},
{
"epoch": 0.253,
"grad_norm": 0.11950170266380834,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 506
},
{
"epoch": 0.2535,
"grad_norm": 0.18548215823845707,
"learning_rate": 1e-05,
"loss": 0.0275,
"step": 507
},
{
"epoch": 0.254,
"grad_norm": 0.15108980653404058,
"learning_rate": 1e-05,
"loss": 0.0197,
"step": 508
},
{
"epoch": 0.2545,
"grad_norm": 0.16504836098536718,
"learning_rate": 1e-05,
"loss": 0.0289,
"step": 509
},
{
"epoch": 0.255,
"grad_norm": 0.15746261920489785,
"learning_rate": 1e-05,
"loss": 0.0253,
"step": 510
},
{
"epoch": 0.2555,
"grad_norm": 0.14071771991438595,
"learning_rate": 1e-05,
"loss": 0.0219,
"step": 511
},
{
"epoch": 0.256,
"grad_norm": 0.16079872072377113,
"learning_rate": 1e-05,
"loss": 0.0204,
"step": 512
},
{
"epoch": 0.256,
"eval_dev_acc": 0.56640625,
"eval_dev_token": 5634.1015625,
"eval_runtime": 361.9891,
"eval_samples_per_second": 0.177,
"eval_steps_per_second": 0.003,
"step": 512
},
{
"epoch": 0.2565,
"grad_norm": 0.13549471484008802,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 513
},
{
"epoch": 0.257,
"grad_norm": 0.12072963489745359,
"learning_rate": 1e-05,
"loss": 0.0302,
"step": 514
},
{
"epoch": 0.2575,
"grad_norm": 0.14026647684897994,
"learning_rate": 1e-05,
"loss": 0.0194,
"step": 515
},
{
"epoch": 0.258,
"grad_norm": 0.1634484411344168,
"learning_rate": 1e-05,
"loss": 0.0277,
"step": 516
},
{
"epoch": 0.2585,
"grad_norm": 0.15844211231505426,
"learning_rate": 1e-05,
"loss": 0.0214,
"step": 517
},
{
"epoch": 0.259,
"grad_norm": 0.1567910605652928,
"learning_rate": 1e-05,
"loss": 0.024,
"step": 518
},
{
"epoch": 0.2595,
"grad_norm": 0.17902606156745304,
"learning_rate": 1e-05,
"loss": 0.033,
"step": 519
},
{
"epoch": 0.26,
"grad_norm": 0.12339744538286439,
"learning_rate": 1e-05,
"loss": 0.0172,
"step": 520
},
{
"epoch": 0.2605,
"grad_norm": 0.13532209502494125,
"learning_rate": 1e-05,
"loss": 0.0206,
"step": 521
},
{
"epoch": 0.261,
"grad_norm": 0.15623082886780087,
"learning_rate": 1e-05,
"loss": 0.0213,
"step": 522
},
{
"epoch": 0.2615,
"grad_norm": 0.14428427308597647,
"learning_rate": 1e-05,
"loss": 0.0201,
"step": 523
},
{
"epoch": 0.262,
"grad_norm": 0.14835567545470982,
"learning_rate": 1e-05,
"loss": 0.0244,
"step": 524
},
{
"epoch": 0.2625,
"grad_norm": 0.14068070672711747,
"learning_rate": 1e-05,
"loss": 0.0239,
"step": 525
},
{
"epoch": 0.263,
"grad_norm": 0.1460843289248216,
"learning_rate": 1e-05,
"loss": 0.0223,
"step": 526
},
{
"epoch": 0.2635,
"grad_norm": 0.13777430449621855,
"learning_rate": 1e-05,
"loss": 0.0229,
"step": 527
},
{
"epoch": 0.264,
"grad_norm": 0.15161607294549337,
"learning_rate": 1e-05,
"loss": 0.0272,
"step": 528
},
{
"epoch": 0.2645,
"grad_norm": 0.13410519048089503,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 529
},
{
"epoch": 0.265,
"grad_norm": 0.15931617673254456,
"learning_rate": 1e-05,
"loss": 0.0244,
"step": 530
},
{
"epoch": 0.2655,
"grad_norm": 0.1410700523457689,
"learning_rate": 1e-05,
"loss": 0.0251,
"step": 531
},
{
"epoch": 0.266,
"grad_norm": 0.11388951846034073,
"learning_rate": 1e-05,
"loss": 0.0144,
"step": 532
},
{
"epoch": 0.2665,
"grad_norm": 0.12253780956369799,
"learning_rate": 1e-05,
"loss": 0.0177,
"step": 533
},
{
"epoch": 0.267,
"grad_norm": 0.15575473599510573,
"learning_rate": 1e-05,
"loss": 0.0192,
"step": 534
},
{
"epoch": 0.2675,
"grad_norm": 0.14690747155640696,
"learning_rate": 1e-05,
"loss": 0.0222,
"step": 535
},
{
"epoch": 0.268,
"grad_norm": 0.13584546405544728,
"learning_rate": 1e-05,
"loss": 0.0237,
"step": 536
},
{
"epoch": 0.2685,
"grad_norm": 0.13430763220790742,
"learning_rate": 1e-05,
"loss": 0.0291,
"step": 537
},
{
"epoch": 0.269,
"grad_norm": 0.14208572873353734,
"learning_rate": 1e-05,
"loss": 0.0187,
"step": 538
},
{
"epoch": 0.2695,
"grad_norm": 0.14058928149963162,
"learning_rate": 1e-05,
"loss": 0.0199,
"step": 539
},
{
"epoch": 0.27,
"grad_norm": 0.15100703501541832,
"learning_rate": 1e-05,
"loss": 0.0348,
"step": 540
},
{
"epoch": 0.2705,
"grad_norm": 0.12269452397268416,
"learning_rate": 1e-05,
"loss": 0.0247,
"step": 541
},
{
"epoch": 0.271,
"grad_norm": 0.1364796501674048,
"learning_rate": 1e-05,
"loss": 0.0227,
"step": 542
},
{
"epoch": 0.2715,
"grad_norm": 0.13163932605554884,
"learning_rate": 1e-05,
"loss": 0.0262,
"step": 543
},
{
"epoch": 0.272,
"grad_norm": 0.13497428740182482,
"learning_rate": 1e-05,
"loss": 0.0206,
"step": 544
},
{
"epoch": 0.272,
"eval_dev_acc": 0.59765625,
"eval_dev_token": 5639.98828125,
"eval_runtime": 359.2369,
"eval_samples_per_second": 0.178,
"eval_steps_per_second": 0.003,
"step": 544
},
{
"epoch": 0.2725,
"grad_norm": 0.13930735859181714,
"learning_rate": 1e-05,
"loss": 0.0234,
"step": 545
},
{
"epoch": 0.273,
"grad_norm": 0.11985280096835198,
"learning_rate": 1e-05,
"loss": 0.0269,
"step": 546
},
{
"epoch": 0.2735,
"grad_norm": 0.17031723198491708,
"learning_rate": 1e-05,
"loss": 0.028,
"step": 547
},
{
"epoch": 0.274,
"grad_norm": 0.17166197772315975,
"learning_rate": 1e-05,
"loss": 0.0281,
"step": 548
},
{
"epoch": 0.2745,
"grad_norm": 0.1167335581681914,
"learning_rate": 1e-05,
"loss": 0.022,
"step": 549
},
{
"epoch": 0.275,
"grad_norm": 0.1443441971157384,
"learning_rate": 1e-05,
"loss": 0.0198,
"step": 550
},
{
"epoch": 0.2755,
"grad_norm": 0.1268787923602722,
"learning_rate": 1e-05,
"loss": 0.017,
"step": 551
},
{
"epoch": 0.276,
"grad_norm": 0.11065296478824395,
"learning_rate": 1e-05,
"loss": 0.0157,
"step": 552
},
{
"epoch": 0.2765,
"grad_norm": 0.12047907824944362,
"learning_rate": 1e-05,
"loss": 0.016,
"step": 553
},
{
"epoch": 0.277,
"grad_norm": 0.13956303855472266,
"learning_rate": 1e-05,
"loss": 0.0216,
"step": 554
},
{
"epoch": 0.2775,
"grad_norm": 0.10533407777378404,
"learning_rate": 1e-05,
"loss": 0.0137,
"step": 555
},
{
"epoch": 0.278,
"grad_norm": 0.11532013491755984,
"learning_rate": 1e-05,
"loss": 0.0282,
"step": 556
},
{
"epoch": 0.2785,
"grad_norm": 0.11921463919727264,
"learning_rate": 1e-05,
"loss": 0.0163,
"step": 557
},
{
"epoch": 0.279,
"grad_norm": 0.15645731769207732,
"learning_rate": 1e-05,
"loss": 0.0241,
"step": 558
},
{
"epoch": 0.2795,
"grad_norm": 0.12096274696840706,
"learning_rate": 1e-05,
"loss": 0.0167,
"step": 559
},
{
"epoch": 0.28,
"grad_norm": 0.149157783124579,
"learning_rate": 1e-05,
"loss": 0.0193,
"step": 560
},
{
"epoch": 0.2805,
"grad_norm": 0.16982490839988412,
"learning_rate": 1e-05,
"loss": 0.0283,
"step": 561
},
{
"epoch": 0.281,
"grad_norm": 0.12038107977310454,
"learning_rate": 1e-05,
"loss": 0.0154,
"step": 562
},
{
"epoch": 0.2815,
"grad_norm": 0.16469919524412158,
"learning_rate": 1e-05,
"loss": 0.0214,
"step": 563
},
{
"epoch": 0.282,
"grad_norm": 0.15827423056846177,
"learning_rate": 1e-05,
"loss": 0.0216,
"step": 564
},
{
"epoch": 0.2825,
"grad_norm": 0.12058245559465251,
"learning_rate": 1e-05,
"loss": 0.0141,
"step": 565
},
{
"epoch": 0.283,
"grad_norm": 0.163789727088167,
"learning_rate": 1e-05,
"loss": 0.0241,
"step": 566
},
{
"epoch": 0.2835,
"grad_norm": 0.1390884369932456,
"learning_rate": 1e-05,
"loss": 0.0221,
"step": 567
},
{
"epoch": 0.284,
"grad_norm": 0.14472941005878595,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 568
},
{
"epoch": 0.2845,
"grad_norm": 0.15437454042645973,
"learning_rate": 1e-05,
"loss": 0.024,
"step": 569
},
{
"epoch": 0.285,
"grad_norm": 0.1207487307624573,
"learning_rate": 1e-05,
"loss": 0.0172,
"step": 570
},
{
"epoch": 0.2855,
"grad_norm": 0.1502409849611173,
"learning_rate": 1e-05,
"loss": 0.0298,
"step": 571
},
{
"epoch": 0.286,
"grad_norm": 0.16401355690597133,
"learning_rate": 1e-05,
"loss": 0.0225,
"step": 572
},
{
"epoch": 0.2865,
"grad_norm": 0.15181464752177645,
"learning_rate": 1e-05,
"loss": 0.0189,
"step": 573
},
{
"epoch": 0.287,
"grad_norm": 0.14560432645081878,
"learning_rate": 1e-05,
"loss": 0.0186,
"step": 574
},
{
"epoch": 0.2875,
"grad_norm": 0.12603042660981642,
"learning_rate": 1e-05,
"loss": 0.0155,
"step": 575
},
{
"epoch": 0.288,
"grad_norm": 0.11638577111126014,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 576
},
{
"epoch": 0.288,
"eval_dev_acc": 0.54296875,
"eval_dev_token": 5801.126953125,
"eval_runtime": 373.1703,
"eval_samples_per_second": 0.172,
"eval_steps_per_second": 0.003,
"step": 576
},
{
"epoch": 0.2885,
"grad_norm": 0.13420942083968396,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 577
},
{
"epoch": 0.289,
"grad_norm": 0.12122809121871923,
"learning_rate": 1e-05,
"loss": 0.0134,
"step": 578
},
{
"epoch": 0.2895,
"grad_norm": 0.13114866603642533,
"learning_rate": 1e-05,
"loss": 0.0174,
"step": 579
},
{
"epoch": 0.29,
"grad_norm": 0.1498609312158644,
"learning_rate": 1e-05,
"loss": 0.0214,
"step": 580
},
{
"epoch": 0.2905,
"grad_norm": 0.1527812218308566,
"learning_rate": 1e-05,
"loss": 0.0194,
"step": 581
},
{
"epoch": 0.291,
"grad_norm": 0.14711876695343454,
"learning_rate": 1e-05,
"loss": 0.018,
"step": 582
},
{
"epoch": 0.2915,
"grad_norm": 0.11529220604038168,
"learning_rate": 1e-05,
"loss": 0.0144,
"step": 583
},
{
"epoch": 0.292,
"grad_norm": 0.16180671831014115,
"learning_rate": 1e-05,
"loss": 0.0165,
"step": 584
},
{
"epoch": 0.2925,
"grad_norm": 0.13621545825638848,
"learning_rate": 1e-05,
"loss": 0.015,
"step": 585
},
{
"epoch": 0.293,
"grad_norm": 0.15473239935591382,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 586
},
{
"epoch": 0.2935,
"grad_norm": 0.15716799171541335,
"learning_rate": 1e-05,
"loss": 0.0194,
"step": 587
},
{
"epoch": 0.294,
"grad_norm": 0.1684941322847538,
"learning_rate": 1e-05,
"loss": 0.022,
"step": 588
},
{
"epoch": 0.2945,
"grad_norm": 0.15453918821249785,
"learning_rate": 1e-05,
"loss": 0.0188,
"step": 589
},
{
"epoch": 0.295,
"grad_norm": 0.140163345657633,
"learning_rate": 1e-05,
"loss": 0.0208,
"step": 590
},
{
"epoch": 0.2955,
"grad_norm": 0.15010258665645038,
"learning_rate": 1e-05,
"loss": 0.0215,
"step": 591
},
{
"epoch": 0.296,
"grad_norm": 0.14661643221841641,
"learning_rate": 1e-05,
"loss": 0.0191,
"step": 592
},
{
"epoch": 0.2965,
"grad_norm": 0.15435066476462508,
"learning_rate": 1e-05,
"loss": 0.0231,
"step": 593
},
{
"epoch": 0.297,
"grad_norm": 0.17094702806791945,
"learning_rate": 1e-05,
"loss": 0.0251,
"step": 594
},
{
"epoch": 0.2975,
"grad_norm": 0.1371139566901347,
"learning_rate": 1e-05,
"loss": 0.0186,
"step": 595
},
{
"epoch": 0.298,
"grad_norm": 0.11779673830033237,
"learning_rate": 1e-05,
"loss": 0.0269,
"step": 596
},
{
"epoch": 0.2985,
"grad_norm": 0.11853976704548681,
"learning_rate": 1e-05,
"loss": 0.0154,
"step": 597
},
{
"epoch": 0.299,
"grad_norm": 0.14881574569113099,
"learning_rate": 1e-05,
"loss": 0.0246,
"step": 598
},
{
"epoch": 0.2995,
"grad_norm": 0.11792409287393274,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 599
},
{
"epoch": 0.3,
"grad_norm": 0.13831559531762572,
"learning_rate": 1e-05,
"loss": 0.0257,
"step": 600
},
{
"epoch": 0.3005,
"grad_norm": 0.13756632720301187,
"learning_rate": 1e-05,
"loss": 0.0214,
"step": 601
},
{
"epoch": 0.301,
"grad_norm": 0.10998907458045305,
"learning_rate": 1e-05,
"loss": 0.015,
"step": 602
},
{
"epoch": 0.3015,
"grad_norm": 0.135955562101373,
"learning_rate": 1e-05,
"loss": 0.0211,
"step": 603
},
{
"epoch": 0.302,
"grad_norm": 0.1214956422000124,
"learning_rate": 1e-05,
"loss": 0.0219,
"step": 604
},
{
"epoch": 0.3025,
"grad_norm": 0.15757702522309963,
"learning_rate": 1e-05,
"loss": 0.018,
"step": 605
},
{
"epoch": 0.303,
"grad_norm": 0.1350858708023801,
"learning_rate": 1e-05,
"loss": 0.0211,
"step": 606
},
{
"epoch": 0.3035,
"grad_norm": 0.10610433140412452,
"learning_rate": 1e-05,
"loss": 0.0147,
"step": 607
},
{
"epoch": 0.304,
"grad_norm": 0.11514647079357257,
"learning_rate": 1e-05,
"loss": 0.0192,
"step": 608
},
{
"epoch": 0.304,
"eval_dev_acc": 0.62109375,
"eval_dev_token": 5367.916015625,
"eval_runtime": 354.8095,
"eval_samples_per_second": 0.18,
"eval_steps_per_second": 0.003,
"step": 608
},
{
"epoch": 0.3045,
"grad_norm": 0.12603567923188372,
"learning_rate": 1e-05,
"loss": 0.0188,
"step": 609
},
{
"epoch": 0.305,
"grad_norm": 0.14277125754270012,
"learning_rate": 1e-05,
"loss": 0.022,
"step": 610
},
{
"epoch": 0.3055,
"grad_norm": 0.12862855181841676,
"learning_rate": 1e-05,
"loss": 0.021,
"step": 611
},
{
"epoch": 0.306,
"grad_norm": 0.1227953424403543,
"learning_rate": 1e-05,
"loss": 0.018,
"step": 612
},
{
"epoch": 0.3065,
"grad_norm": 0.11646820367498804,
"learning_rate": 1e-05,
"loss": 0.0131,
"step": 613
},
{
"epoch": 0.307,
"grad_norm": 0.14701145754992329,
"learning_rate": 1e-05,
"loss": 0.0186,
"step": 614
},
{
"epoch": 0.3075,
"grad_norm": 0.1493073818813876,
"learning_rate": 1e-05,
"loss": 0.0254,
"step": 615
},
{
"epoch": 0.308,
"grad_norm": 0.1352952895732537,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 616
},
{
"epoch": 0.3085,
"grad_norm": 0.13007743097982305,
"learning_rate": 1e-05,
"loss": 0.0171,
"step": 617
},
{
"epoch": 0.309,
"grad_norm": 0.1665432351262121,
"learning_rate": 1e-05,
"loss": 0.0228,
"step": 618
},
{
"epoch": 0.3095,
"grad_norm": 0.16442931730443322,
"learning_rate": 1e-05,
"loss": 0.0238,
"step": 619
},
{
"epoch": 0.31,
"grad_norm": 0.16320986192220768,
"learning_rate": 1e-05,
"loss": 0.0238,
"step": 620
},
{
"epoch": 0.3105,
"grad_norm": 0.13880254871235365,
"learning_rate": 1e-05,
"loss": 0.0179,
"step": 621
},
{
"epoch": 0.311,
"grad_norm": 0.13609379700738453,
"learning_rate": 1e-05,
"loss": 0.0195,
"step": 622
},
{
"epoch": 0.3115,
"grad_norm": 0.1368415516519621,
"learning_rate": 1e-05,
"loss": 0.024,
"step": 623
},
{
"epoch": 0.312,
"grad_norm": 0.12821586481120512,
"learning_rate": 1e-05,
"loss": 0.0191,
"step": 624
},
{
"epoch": 0.3125,
"grad_norm": 0.13644927854222083,
"learning_rate": 1e-05,
"loss": 0.0154,
"step": 625
},
{
"epoch": 0.313,
"grad_norm": 0.15158164143556496,
"learning_rate": 1e-05,
"loss": 0.0318,
"step": 626
},
{
"epoch": 0.3135,
"grad_norm": 0.12404550422721679,
"learning_rate": 1e-05,
"loss": 0.0202,
"step": 627
},
{
"epoch": 0.314,
"grad_norm": 0.1235074023832298,
"learning_rate": 1e-05,
"loss": 0.0167,
"step": 628
},
{
"epoch": 0.3145,
"grad_norm": 0.16094487436899907,
"learning_rate": 1e-05,
"loss": 0.0227,
"step": 629
},
{
"epoch": 0.315,
"grad_norm": 0.11086598912590964,
"learning_rate": 1e-05,
"loss": 0.0158,
"step": 630
},
{
"epoch": 0.3155,
"grad_norm": 0.1147741974179167,
"learning_rate": 1e-05,
"loss": 0.0136,
"step": 631
},
{
"epoch": 0.316,
"grad_norm": 0.12346095617438974,
"learning_rate": 1e-05,
"loss": 0.017,
"step": 632
},
{
"epoch": 0.3165,
"grad_norm": 0.1235267138232638,
"learning_rate": 1e-05,
"loss": 0.0131,
"step": 633
},
{
"epoch": 0.317,
"grad_norm": 0.11979162262432065,
"learning_rate": 1e-05,
"loss": 0.0194,
"step": 634
},
{
"epoch": 0.3175,
"grad_norm": 0.12253729986288973,
"learning_rate": 1e-05,
"loss": 0.0205,
"step": 635
},
{
"epoch": 0.318,
"grad_norm": 0.1374736081434109,
"learning_rate": 1e-05,
"loss": 0.0207,
"step": 636
},
{
"epoch": 0.3185,
"grad_norm": 0.11667911740285354,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 637
},
{
"epoch": 0.319,
"grad_norm": 0.13725799823509804,
"learning_rate": 1e-05,
"loss": 0.0177,
"step": 638
},
{
"epoch": 0.3195,
"grad_norm": 0.1461325036101512,
"learning_rate": 1e-05,
"loss": 0.0233,
"step": 639
},
{
"epoch": 0.32,
"grad_norm": 0.1486586288765987,
"learning_rate": 1e-05,
"loss": 0.022,
"step": 640
},
{
"epoch": 0.32,
"eval_dev_acc": 0.52734375,
"eval_dev_token": 5585.80859375,
"eval_runtime": 360.5581,
"eval_samples_per_second": 0.178,
"eval_steps_per_second": 0.003,
"step": 640
},
{
"epoch": 0.3205,
"grad_norm": 0.13037027619165104,
"learning_rate": 1e-05,
"loss": 0.0181,
"step": 641
},
{
"epoch": 0.321,
"grad_norm": 0.1384777662387777,
"learning_rate": 1e-05,
"loss": 0.0284,
"step": 642
},
{
"epoch": 0.3215,
"grad_norm": 0.11731142200376247,
"learning_rate": 1e-05,
"loss": 0.0136,
"step": 643
},
{
"epoch": 0.322,
"grad_norm": 0.13199000719975476,
"learning_rate": 1e-05,
"loss": 0.0147,
"step": 644
},
{
"epoch": 0.3225,
"grad_norm": 0.145679314202878,
"learning_rate": 1e-05,
"loss": 0.0227,
"step": 645
},
{
"epoch": 0.323,
"grad_norm": 0.13813521110883425,
"learning_rate": 1e-05,
"loss": 0.0173,
"step": 646
},
{
"epoch": 0.3235,
"grad_norm": 0.11216370610734963,
"learning_rate": 1e-05,
"loss": 0.0125,
"step": 647
},
{
"epoch": 0.324,
"grad_norm": 0.09898218700430327,
"learning_rate": 1e-05,
"loss": 0.0165,
"step": 648
},
{
"epoch": 0.3245,
"grad_norm": 0.13555813085878698,
"learning_rate": 1e-05,
"loss": 0.0163,
"step": 649
},
{
"epoch": 0.325,
"grad_norm": 0.11552480540546263,
"learning_rate": 1e-05,
"loss": 0.0161,
"step": 650
},
{
"epoch": 0.3255,
"grad_norm": 0.12028253181729011,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 651
},
{
"epoch": 0.326,
"grad_norm": 0.11773491790468957,
"learning_rate": 1e-05,
"loss": 0.0143,
"step": 652
},
{
"epoch": 0.3265,
"grad_norm": 0.11052002031571376,
"learning_rate": 1e-05,
"loss": 0.0177,
"step": 653
},
{
"epoch": 0.327,
"grad_norm": 0.1251016102612686,
"learning_rate": 1e-05,
"loss": 0.0134,
"step": 654
},
{
"epoch": 0.3275,
"grad_norm": 0.13329112874236815,
"learning_rate": 1e-05,
"loss": 0.0148,
"step": 655
},
{
"epoch": 0.328,
"grad_norm": 0.11382501473889628,
"learning_rate": 1e-05,
"loss": 0.0145,
"step": 656
},
{
"epoch": 0.3285,
"grad_norm": 0.1319495587226548,
"learning_rate": 1e-05,
"loss": 0.0139,
"step": 657
},
{
"epoch": 0.329,
"grad_norm": 0.12070785233460224,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 658
},
{
"epoch": 0.3295,
"grad_norm": 0.12142680030465443,
"learning_rate": 1e-05,
"loss": 0.0159,
"step": 659
},
{
"epoch": 0.33,
"grad_norm": 0.12182666413282645,
"learning_rate": 1e-05,
"loss": 0.0114,
"step": 660
},
{
"epoch": 0.3305,
"grad_norm": 0.13190056217824564,
"learning_rate": 1e-05,
"loss": 0.0163,
"step": 661
},
{
"epoch": 0.331,
"grad_norm": 0.12804397999806036,
"learning_rate": 1e-05,
"loss": 0.016,
"step": 662
},
{
"epoch": 0.3315,
"grad_norm": 0.1461414693983946,
"learning_rate": 1e-05,
"loss": 0.0179,
"step": 663
},
{
"epoch": 0.332,
"grad_norm": 0.1467248720157808,
"learning_rate": 1e-05,
"loss": 0.0171,
"step": 664
},
{
"epoch": 0.3325,
"grad_norm": 0.1393703645756879,
"learning_rate": 1e-05,
"loss": 0.017,
"step": 665
},
{
"epoch": 0.333,
"grad_norm": 0.1677851073130961,
"learning_rate": 1e-05,
"loss": 0.033,
"step": 666
},
{
"epoch": 0.3335,
"grad_norm": 0.13245384270982133,
"learning_rate": 1e-05,
"loss": 0.0163,
"step": 667
},
{
"epoch": 0.334,
"grad_norm": 0.12289205645457782,
"learning_rate": 1e-05,
"loss": 0.015,
"step": 668
},
{
"epoch": 0.3345,
"grad_norm": 0.13290806824539741,
"learning_rate": 1e-05,
"loss": 0.0131,
"step": 669
},
{
"epoch": 0.335,
"grad_norm": 0.12256221521384754,
"learning_rate": 1e-05,
"loss": 0.0138,
"step": 670
},
{
"epoch": 0.3355,
"grad_norm": 0.11871572464334247,
"learning_rate": 1e-05,
"loss": 0.0156,
"step": 671
},
{
"epoch": 0.336,
"grad_norm": 0.1425273432040885,
"learning_rate": 1e-05,
"loss": 0.0132,
"step": 672
},
{
"epoch": 0.336,
"eval_dev_acc": 0.5625,
"eval_dev_token": 5492.04296875,
"eval_runtime": 348.043,
"eval_samples_per_second": 0.184,
"eval_steps_per_second": 0.003,
"step": 672
},
{
"epoch": 0.3365,
"grad_norm": 0.12612838576017849,
"learning_rate": 1e-05,
"loss": 0.0127,
"step": 673
},
{
"epoch": 0.337,
"grad_norm": 0.1294118778329942,
"learning_rate": 1e-05,
"loss": 0.0161,
"step": 674
},
{
"epoch": 0.3375,
"grad_norm": 0.11771619437889824,
"learning_rate": 1e-05,
"loss": 0.0142,
"step": 675
},
{
"epoch": 0.338,
"grad_norm": 0.13434389137713848,
"learning_rate": 1e-05,
"loss": 0.0157,
"step": 676
},
{
"epoch": 0.3385,
"grad_norm": 0.2579148520419769,
"learning_rate": 1e-05,
"loss": 0.0322,
"step": 677
},
{
"epoch": 0.339,
"grad_norm": 0.12876645046050272,
"learning_rate": 1e-05,
"loss": 0.0127,
"step": 678
},
{
"epoch": 0.3395,
"grad_norm": 0.12336988858308351,
"learning_rate": 1e-05,
"loss": 0.0114,
"step": 679
},
{
"epoch": 0.34,
"grad_norm": 0.11038801232074134,
"learning_rate": 1e-05,
"loss": 0.0112,
"step": 680
},
{
"epoch": 0.3405,
"grad_norm": 0.13782079916676085,
"learning_rate": 1e-05,
"loss": 0.0156,
"step": 681
},
{
"epoch": 0.341,
"grad_norm": 0.12593807949317973,
"learning_rate": 1e-05,
"loss": 0.0179,
"step": 682
},
{
"epoch": 0.3415,
"grad_norm": 0.13416291611922937,
"learning_rate": 1e-05,
"loss": 0.0167,
"step": 683
},
{
"epoch": 0.342,
"grad_norm": 0.14107509427243767,
"learning_rate": 1e-05,
"loss": 0.0182,
"step": 684
},
{
"epoch": 0.3425,
"grad_norm": 0.13829290958101634,
"learning_rate": 1e-05,
"loss": 0.0133,
"step": 685
},
{
"epoch": 0.343,
"grad_norm": 0.12072602797225344,
"learning_rate": 1e-05,
"loss": 0.0135,
"step": 686
},
{
"epoch": 0.3435,
"grad_norm": 0.09808985286323638,
"learning_rate": 1e-05,
"loss": 0.0109,
"step": 687
},
{
"epoch": 0.344,
"grad_norm": 0.1237155662701831,
"learning_rate": 1e-05,
"loss": 0.0137,
"step": 688
},
{
"epoch": 0.3445,
"grad_norm": 0.11512509686864711,
"learning_rate": 1e-05,
"loss": 0.0131,
"step": 689
},
{
"epoch": 0.345,
"grad_norm": 0.10310189822258317,
"learning_rate": 1e-05,
"loss": 0.011,
"step": 690
},
{
"epoch": 0.3455,
"grad_norm": 0.17008589258309467,
"learning_rate": 1e-05,
"loss": 0.0219,
"step": 691
},
{
"epoch": 0.346,
"grad_norm": 0.12175425158539896,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 692
},
{
"epoch": 0.3465,
"grad_norm": 0.11242731302801981,
"learning_rate": 1e-05,
"loss": 0.0142,
"step": 693
},
{
"epoch": 0.347,
"grad_norm": 0.13975508336225442,
"learning_rate": 1e-05,
"loss": 0.0149,
"step": 694
},
{
"epoch": 0.3475,
"grad_norm": 0.08820258618918772,
"learning_rate": 1e-05,
"loss": 0.008,
"step": 695
},
{
"epoch": 0.348,
"grad_norm": 0.1343553646574964,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 696
},
{
"epoch": 0.3485,
"grad_norm": 0.13274844121268298,
"learning_rate": 1e-05,
"loss": 0.0143,
"step": 697
},
{
"epoch": 0.349,
"grad_norm": 0.1412865550899799,
"learning_rate": 1e-05,
"loss": 0.0164,
"step": 698
},
{
"epoch": 0.3495,
"grad_norm": 0.1471608978417448,
"learning_rate": 1e-05,
"loss": 0.0144,
"step": 699
},
{
"epoch": 0.35,
"grad_norm": 0.13861936123681107,
"learning_rate": 1e-05,
"loss": 0.0151,
"step": 700
},
{
"epoch": 0.3505,
"grad_norm": 0.11790173787959472,
"learning_rate": 1e-05,
"loss": 0.0252,
"step": 701
},
{
"epoch": 0.351,
"grad_norm": 0.12166501219045824,
"learning_rate": 1e-05,
"loss": 0.0145,
"step": 702
},
{
"epoch": 0.3515,
"grad_norm": 0.1082871078184047,
"learning_rate": 1e-05,
"loss": 0.0095,
"step": 703
},
{
"epoch": 0.352,
"grad_norm": 0.12226846848831563,
"learning_rate": 1e-05,
"loss": 0.0141,
"step": 704
},
{
"epoch": 0.352,
"eval_dev_acc": 0.52734375,
"eval_dev_token": 5760.470703125,
"eval_runtime": 408.3746,
"eval_samples_per_second": 0.157,
"eval_steps_per_second": 0.002,
"step": 704
},
{
"epoch": 0.3525,
"grad_norm": 0.09666366634628189,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 705
},
{
"epoch": 0.353,
"grad_norm": 0.11283217042776503,
"learning_rate": 1e-05,
"loss": 0.0143,
"step": 706
},
{
"epoch": 0.3535,
"grad_norm": 0.12289011751838193,
"learning_rate": 1e-05,
"loss": 0.0173,
"step": 707
},
{
"epoch": 0.354,
"grad_norm": 0.11335657340497375,
"learning_rate": 1e-05,
"loss": 0.0193,
"step": 708
},
{
"epoch": 0.3545,
"grad_norm": 0.12878402333619293,
"learning_rate": 1e-05,
"loss": 0.0198,
"step": 709
},
{
"epoch": 0.355,
"grad_norm": 0.09768170830344952,
"learning_rate": 1e-05,
"loss": 0.0123,
"step": 710
},
{
"epoch": 0.3555,
"grad_norm": 0.11142711749521704,
"learning_rate": 1e-05,
"loss": 0.0142,
"step": 711
},
{
"epoch": 0.356,
"grad_norm": 0.12263582497303413,
"learning_rate": 1e-05,
"loss": 0.0125,
"step": 712
},
{
"epoch": 0.3565,
"grad_norm": 0.10602173709981424,
"learning_rate": 1e-05,
"loss": 0.013,
"step": 713
},
{
"epoch": 0.357,
"grad_norm": 0.09797092044857854,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 714
},
{
"epoch": 0.3575,
"grad_norm": 0.10267028455310084,
"learning_rate": 1e-05,
"loss": 0.0153,
"step": 715
},
{
"epoch": 0.358,
"grad_norm": 0.11247269902696218,
"learning_rate": 1e-05,
"loss": 0.0124,
"step": 716
},
{
"epoch": 0.3585,
"grad_norm": 0.10192090643527008,
"learning_rate": 1e-05,
"loss": 0.01,
"step": 717
},
{
"epoch": 0.359,
"grad_norm": 0.15447030723754146,
"learning_rate": 1e-05,
"loss": 0.0191,
"step": 718
},
{
"epoch": 0.3595,
"grad_norm": 0.12653173296866044,
"learning_rate": 1e-05,
"loss": 0.0144,
"step": 719
},
{
"epoch": 0.36,
"grad_norm": 0.09798115770372441,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 720
},
{
"epoch": 0.3605,
"grad_norm": 0.13673705245890774,
"learning_rate": 1e-05,
"loss": 0.0176,
"step": 721
},
{
"epoch": 0.361,
"grad_norm": 0.10484924187213118,
"learning_rate": 1e-05,
"loss": 0.011,
"step": 722
},
{
"epoch": 0.3615,
"grad_norm": 0.12026390551992476,
"learning_rate": 1e-05,
"loss": 0.0161,
"step": 723
},
{
"epoch": 0.362,
"grad_norm": 0.12248845158519388,
"learning_rate": 1e-05,
"loss": 0.0135,
"step": 724
},
{
"epoch": 0.3625,
"grad_norm": 0.08242254382606763,
"learning_rate": 1e-05,
"loss": 0.0103,
"step": 725
},
{
"epoch": 0.363,
"grad_norm": 0.1310711546429888,
"learning_rate": 1e-05,
"loss": 0.0213,
"step": 726
},
{
"epoch": 0.3635,
"grad_norm": 0.11504216894290854,
"learning_rate": 1e-05,
"loss": 0.0122,
"step": 727
},
{
"epoch": 0.364,
"grad_norm": 0.12368075631500317,
"learning_rate": 1e-05,
"loss": 0.0184,
"step": 728
},
{
"epoch": 0.3645,
"grad_norm": 0.12437061544598803,
"learning_rate": 1e-05,
"loss": 0.0152,
"step": 729
},
{
"epoch": 0.365,
"grad_norm": 0.13397286445240938,
"learning_rate": 1e-05,
"loss": 0.0158,
"step": 730
},
{
"epoch": 0.3655,
"grad_norm": 0.0982171426550068,
"learning_rate": 1e-05,
"loss": 0.0096,
"step": 731
},
{
"epoch": 0.366,
"grad_norm": 0.10415731829601732,
"learning_rate": 1e-05,
"loss": 0.0105,
"step": 732
},
{
"epoch": 0.3665,
"grad_norm": 0.13524982542339864,
"learning_rate": 1e-05,
"loss": 0.0183,
"step": 733
},
{
"epoch": 0.367,
"grad_norm": 0.09664339538491498,
"learning_rate": 1e-05,
"loss": 0.0115,
"step": 734
},
{
"epoch": 0.3675,
"grad_norm": 0.0851015567710096,
"learning_rate": 1e-05,
"loss": 0.0112,
"step": 735
},
{
"epoch": 0.368,
"grad_norm": 0.10062979284816816,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 736
},
{
"epoch": 0.368,
"eval_dev_acc": 0.53515625,
"eval_dev_token": 5898.23046875,
"eval_runtime": 365.57,
"eval_samples_per_second": 0.175,
"eval_steps_per_second": 0.003,
"step": 736
},
{
"epoch": 0.3685,
"grad_norm": 0.11264161967062039,
"learning_rate": 1e-05,
"loss": 0.0108,
"step": 737
},
{
"epoch": 0.369,
"grad_norm": 0.11222433927286389,
"learning_rate": 1e-05,
"loss": 0.0104,
"step": 738
},
{
"epoch": 0.3695,
"grad_norm": 0.20055167671089585,
"learning_rate": 1e-05,
"loss": 0.013,
"step": 739
},
{
"epoch": 0.37,
"grad_norm": 0.10491538007846005,
"learning_rate": 1e-05,
"loss": 0.0131,
"step": 740
},
{
"epoch": 0.3705,
"grad_norm": 0.09174967069041229,
"learning_rate": 1e-05,
"loss": 0.0095,
"step": 741
},
{
"epoch": 0.371,
"grad_norm": 0.10582756451393364,
"learning_rate": 1e-05,
"loss": 0.012,
"step": 742
},
{
"epoch": 0.3715,
"grad_norm": 0.09194993895839906,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 743
},
{
"epoch": 0.372,
"grad_norm": 0.08858743590625234,
"learning_rate": 1e-05,
"loss": 0.01,
"step": 744
},
{
"epoch": 0.3725,
"grad_norm": 0.10826764845042156,
"learning_rate": 1e-05,
"loss": 0.0145,
"step": 745
},
{
"epoch": 0.373,
"grad_norm": 0.11033032167695773,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 746
},
{
"epoch": 0.3735,
"grad_norm": 0.15107498232603372,
"learning_rate": 1e-05,
"loss": 0.0157,
"step": 747
},
{
"epoch": 0.374,
"grad_norm": 0.11411131491498813,
"learning_rate": 1e-05,
"loss": 0.0114,
"step": 748
},
{
"epoch": 0.3745,
"grad_norm": 0.12425138196940645,
"learning_rate": 1e-05,
"loss": 0.0138,
"step": 749
},
{
"epoch": 0.375,
"grad_norm": 0.10647093971928946,
"learning_rate": 1e-05,
"loss": 0.0107,
"step": 750
},
{
"epoch": 0.3755,
"grad_norm": 0.12413448325714936,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 751
},
{
"epoch": 0.376,
"grad_norm": 0.105340917878943,
"learning_rate": 1e-05,
"loss": 0.0103,
"step": 752
},
{
"epoch": 0.3765,
"grad_norm": 0.11802541191692037,
"learning_rate": 1e-05,
"loss": 0.0127,
"step": 753
},
{
"epoch": 0.377,
"grad_norm": 0.09864810460683521,
"learning_rate": 1e-05,
"loss": 0.0128,
"step": 754
},
{
"epoch": 0.3775,
"grad_norm": 0.12491513055109611,
"learning_rate": 1e-05,
"loss": 0.0102,
"step": 755
},
{
"epoch": 0.378,
"grad_norm": 0.11778790720208179,
"learning_rate": 1e-05,
"loss": 0.0093,
"step": 756
},
{
"epoch": 0.3785,
"grad_norm": 0.11902239371415295,
"learning_rate": 1e-05,
"loss": 0.0115,
"step": 757
},
{
"epoch": 0.379,
"grad_norm": 0.1025498591924567,
"learning_rate": 1e-05,
"loss": 0.0099,
"step": 758
},
{
"epoch": 0.3795,
"grad_norm": 0.14120840112868438,
"learning_rate": 1e-05,
"loss": 0.0123,
"step": 759
},
{
"epoch": 0.38,
"grad_norm": 0.08964665828244849,
"learning_rate": 1e-05,
"loss": 0.0116,
"step": 760
},
{
"epoch": 0.3805,
"grad_norm": 0.13533754381134197,
"learning_rate": 1e-05,
"loss": 0.0142,
"step": 761
},
{
"epoch": 0.381,
"grad_norm": 0.11151852649444051,
"learning_rate": 1e-05,
"loss": 0.0122,
"step": 762
},
{
"epoch": 0.3815,
"grad_norm": 0.12448616858204287,
"learning_rate": 1e-05,
"loss": 0.0148,
"step": 763
},
{
"epoch": 0.382,
"grad_norm": 0.10160526390201502,
"learning_rate": 1e-05,
"loss": 0.0102,
"step": 764
},
{
"epoch": 0.3825,
"grad_norm": 0.12329039775788013,
"learning_rate": 1e-05,
"loss": 0.0226,
"step": 765
},
{
"epoch": 0.383,
"grad_norm": 0.1706851101549876,
"learning_rate": 1e-05,
"loss": 0.013,
"step": 766
},
{
"epoch": 0.3835,
"grad_norm": 0.11518698920716465,
"learning_rate": 1e-05,
"loss": 0.0136,
"step": 767
},
{
"epoch": 0.384,
"grad_norm": 0.1130084278865893,
"learning_rate": 1e-05,
"loss": 0.0151,
"step": 768
},
{
"epoch": 0.384,
"eval_dev_acc": 0.505859375,
"eval_dev_token": 5660.353515625,
"eval_runtime": 362.2613,
"eval_samples_per_second": 0.177,
"eval_steps_per_second": 0.003,
"step": 768
},
{
"epoch": 0.3845,
"grad_norm": 0.12092852490034717,
"learning_rate": 1e-05,
"loss": 0.0111,
"step": 769
},
{
"epoch": 0.385,
"grad_norm": 0.12041159291779986,
"learning_rate": 1e-05,
"loss": 0.0157,
"step": 770
},
{
"epoch": 0.3855,
"grad_norm": 0.13135896730332378,
"learning_rate": 1e-05,
"loss": 0.0149,
"step": 771
},
{
"epoch": 0.386,
"grad_norm": 0.09838622926936438,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 772
},
{
"epoch": 0.3865,
"grad_norm": 0.1118823935585986,
"learning_rate": 1e-05,
"loss": 0.016,
"step": 773
},
{
"epoch": 0.387,
"grad_norm": 0.10549287704509497,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 774
},
{
"epoch": 0.3875,
"grad_norm": 0.1232106501038667,
"learning_rate": 1e-05,
"loss": 0.013,
"step": 775
},
{
"epoch": 0.388,
"grad_norm": 0.12975165037101072,
"learning_rate": 1e-05,
"loss": 0.0121,
"step": 776
},
{
"epoch": 0.3885,
"grad_norm": 0.13595262265606586,
"learning_rate": 1e-05,
"loss": 0.0126,
"step": 777
},
{
"epoch": 0.389,
"grad_norm": 0.1371181387315353,
"learning_rate": 1e-05,
"loss": 0.0123,
"step": 778
},
{
"epoch": 0.3895,
"grad_norm": 0.13359565264106144,
"learning_rate": 1e-05,
"loss": 0.0153,
"step": 779
},
{
"epoch": 0.39,
"grad_norm": 0.11480062436980976,
"learning_rate": 1e-05,
"loss": 0.0107,
"step": 780
},
{
"epoch": 0.3905,
"grad_norm": 0.09947922362782227,
"learning_rate": 1e-05,
"loss": 0.0086,
"step": 781
},
{
"epoch": 0.391,
"grad_norm": 0.08734624126570777,
"learning_rate": 1e-05,
"loss": 0.0076,
"step": 782
},
{
"epoch": 0.3915,
"grad_norm": 0.10267346744796824,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 783
},
{
"epoch": 0.392,
"grad_norm": 0.09576192367034056,
"learning_rate": 1e-05,
"loss": 0.0083,
"step": 784
},
{
"epoch": 0.3925,
"grad_norm": 0.14141610877259214,
"learning_rate": 1e-05,
"loss": 0.0139,
"step": 785
},
{
"epoch": 0.393,
"grad_norm": 0.12071117391984963,
"learning_rate": 1e-05,
"loss": 0.0115,
"step": 786
},
{
"epoch": 0.3935,
"grad_norm": 0.13172358788988653,
"learning_rate": 1e-05,
"loss": 0.0129,
"step": 787
},
{
"epoch": 0.394,
"grad_norm": 0.1345424657278948,
"learning_rate": 1e-05,
"loss": 0.014,
"step": 788
},
{
"epoch": 0.3945,
"grad_norm": 0.1166885723717584,
"learning_rate": 1e-05,
"loss": 0.0125,
"step": 789
},
{
"epoch": 0.395,
"grad_norm": 0.1513194581227132,
"learning_rate": 1e-05,
"loss": 0.0145,
"step": 790
},
{
"epoch": 0.3955,
"grad_norm": 0.11105132556458801,
"learning_rate": 1e-05,
"loss": 0.011,
"step": 791
},
{
"epoch": 0.396,
"grad_norm": 0.0989302250949553,
"learning_rate": 1e-05,
"loss": 0.0086,
"step": 792
},
{
"epoch": 0.3965,
"grad_norm": 0.1384321782028591,
"learning_rate": 1e-05,
"loss": 0.0122,
"step": 793
},
{
"epoch": 0.397,
"grad_norm": 0.1439582094147518,
"learning_rate": 1e-05,
"loss": 0.0142,
"step": 794
},
{
"epoch": 0.3975,
"grad_norm": 0.11905534302900607,
"learning_rate": 1e-05,
"loss": 0.0102,
"step": 795
},
{
"epoch": 0.398,
"grad_norm": 0.1353943654211476,
"learning_rate": 1e-05,
"loss": 0.015,
"step": 796
},
{
"epoch": 0.3985,
"grad_norm": 0.1227008050574365,
"learning_rate": 1e-05,
"loss": 0.0159,
"step": 797
},
{
"epoch": 0.399,
"grad_norm": 0.10084479670557596,
"learning_rate": 1e-05,
"loss": 0.0103,
"step": 798
},
{
"epoch": 0.3995,
"grad_norm": 0.11430101131806415,
"learning_rate": 1e-05,
"loss": 0.0128,
"step": 799
},
{
"epoch": 0.4,
"grad_norm": 0.12720193634725216,
"learning_rate": 1e-05,
"loss": 0.0166,
"step": 800
},
{
"epoch": 0.4,
"eval_dev_acc": 0.49609375,
"eval_dev_token": 5933.87109375,
"eval_runtime": 371.4205,
"eval_samples_per_second": 0.172,
"eval_steps_per_second": 0.003,
"step": 800
},
{
"epoch": 0.4005,
"grad_norm": 0.10747534966369286,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 801
},
{
"epoch": 0.401,
"grad_norm": 0.09755157150747987,
"learning_rate": 1e-05,
"loss": 0.0094,
"step": 802
},
{
"epoch": 0.4015,
"grad_norm": 0.1093754763997552,
"learning_rate": 1e-05,
"loss": 0.0136,
"step": 803
},
{
"epoch": 0.402,
"grad_norm": 0.12186788180512477,
"learning_rate": 1e-05,
"loss": 0.0112,
"step": 804
},
{
"epoch": 0.4025,
"grad_norm": 0.12341274893925132,
"learning_rate": 1e-05,
"loss": 0.0126,
"step": 805
},
{
"epoch": 0.403,
"grad_norm": 0.1000051489528446,
"learning_rate": 1e-05,
"loss": 0.0171,
"step": 806
},
{
"epoch": 0.4035,
"grad_norm": 0.09210699513663904,
"learning_rate": 1e-05,
"loss": 0.0097,
"step": 807
},
{
"epoch": 0.404,
"grad_norm": 0.12629053026155362,
"learning_rate": 1e-05,
"loss": 0.0126,
"step": 808
},
{
"epoch": 0.4045,
"grad_norm": 0.11979707428750866,
"learning_rate": 1e-05,
"loss": 0.0129,
"step": 809
},
{
"epoch": 0.405,
"grad_norm": 0.13240620090939892,
"learning_rate": 1e-05,
"loss": 0.0173,
"step": 810
},
{
"epoch": 0.4055,
"grad_norm": 0.11509826968311627,
"learning_rate": 1e-05,
"loss": 0.0133,
"step": 811
},
{
"epoch": 0.406,
"grad_norm": 0.09930674216692448,
"learning_rate": 1e-05,
"loss": 0.0103,
"step": 812
},
{
"epoch": 0.4065,
"grad_norm": 0.12486374889735856,
"learning_rate": 1e-05,
"loss": 0.0152,
"step": 813
},
{
"epoch": 0.407,
"grad_norm": 0.09439307719243419,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 814
},
{
"epoch": 0.4075,
"grad_norm": 0.09010069214916468,
"learning_rate": 1e-05,
"loss": 0.0101,
"step": 815
},
{
"epoch": 0.408,
"grad_norm": 0.10807091465236611,
"learning_rate": 1e-05,
"loss": 0.0136,
"step": 816
},
{
"epoch": 0.4085,
"grad_norm": 0.09320554728801374,
"learning_rate": 1e-05,
"loss": 0.0102,
"step": 817
},
{
"epoch": 0.409,
"grad_norm": 0.09578310039513883,
"learning_rate": 1e-05,
"loss": 0.0105,
"step": 818
},
{
"epoch": 0.4095,
"grad_norm": 0.10795646602867415,
"learning_rate": 1e-05,
"loss": 0.0141,
"step": 819
},
{
"epoch": 0.41,
"grad_norm": 0.11260765818079684,
"learning_rate": 1e-05,
"loss": 0.0123,
"step": 820
},
{
"epoch": 0.4105,
"grad_norm": 0.13162780397133028,
"learning_rate": 1e-05,
"loss": 0.0141,
"step": 821
},
{
"epoch": 0.411,
"grad_norm": 0.10884500486087925,
"learning_rate": 1e-05,
"loss": 0.0122,
"step": 822
},
{
"epoch": 0.4115,
"grad_norm": 0.12472839162847292,
"learning_rate": 1e-05,
"loss": 0.0135,
"step": 823
},
{
"epoch": 0.412,
"grad_norm": 0.13289735991638021,
"learning_rate": 1e-05,
"loss": 0.0112,
"step": 824
},
{
"epoch": 0.4125,
"grad_norm": 0.12509974441211302,
"learning_rate": 1e-05,
"loss": 0.0256,
"step": 825
},
{
"epoch": 0.413,
"grad_norm": 0.12014632147622897,
"learning_rate": 1e-05,
"loss": 0.0113,
"step": 826
},
{
"epoch": 0.4135,
"grad_norm": 0.09172916013688245,
"learning_rate": 1e-05,
"loss": 0.0082,
"step": 827
},
{
"epoch": 0.414,
"grad_norm": 0.09305774811224422,
"learning_rate": 1e-05,
"loss": 0.0074,
"step": 828
},
{
"epoch": 0.4145,
"grad_norm": 0.12720506111352092,
"learning_rate": 1e-05,
"loss": 0.0091,
"step": 829
},
{
"epoch": 0.415,
"grad_norm": 0.0815065287803298,
"learning_rate": 1e-05,
"loss": 0.0061,
"step": 830
},
{
"epoch": 0.4155,
"grad_norm": 0.09730425117259746,
"learning_rate": 1e-05,
"loss": 0.0079,
"step": 831
},
{
"epoch": 0.416,
"grad_norm": 0.09579694307116488,
"learning_rate": 1e-05,
"loss": 0.0098,
"step": 832
},
{
"epoch": 0.416,
"eval_dev_acc": 0.513671875,
"eval_dev_token": 4446.53125,
"eval_runtime": 329.208,
"eval_samples_per_second": 0.194,
"eval_steps_per_second": 0.003,
"step": 832
}
],
"logging_steps": 1.0,
"max_steps": 2000,
"num_input_tokens_seen": 0,
"num_train_epochs": 9223372036854775807,
"save_steps": 32,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 224786117410816.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}