PEFT
PyTorch
Safetensors
llama
Generated from Trainer
mtasic85's picture
Upload folder using huggingface_hub
69cbba5 verified
raw
history blame
44.1 kB
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 1.0,
"eval_steps": 82,
"global_step": 245,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.004081632653061225,
"grad_norm": 0.03618604317307472,
"learning_rate": 2e-05,
"loss": 1.1867,
"step": 1
},
{
"epoch": 0.004081632653061225,
"eval_loss": 1.2217298746109009,
"eval_runtime": 117.7701,
"eval_samples_per_second": 20.735,
"eval_steps_per_second": 10.368,
"step": 1
},
{
"epoch": 0.00816326530612245,
"grad_norm": 0.03361086547374725,
"learning_rate": 4e-05,
"loss": 1.1923,
"step": 2
},
{
"epoch": 0.012244897959183673,
"grad_norm": 0.034457892179489136,
"learning_rate": 6e-05,
"loss": 1.0744,
"step": 3
},
{
"epoch": 0.0163265306122449,
"grad_norm": 0.03408154472708702,
"learning_rate": 8e-05,
"loss": 1.2796,
"step": 4
},
{
"epoch": 0.02040816326530612,
"grad_norm": 0.03475559875369072,
"learning_rate": 0.0001,
"loss": 1.2709,
"step": 5
},
{
"epoch": 0.024489795918367346,
"grad_norm": 0.03768599033355713,
"learning_rate": 0.00012,
"loss": 1.0895,
"step": 6
},
{
"epoch": 0.02857142857142857,
"grad_norm": 0.03434902802109718,
"learning_rate": 0.00014,
"loss": 1.2194,
"step": 7
},
{
"epoch": 0.0326530612244898,
"grad_norm": 0.029295403510332108,
"learning_rate": 0.00016,
"loss": 1.1522,
"step": 8
},
{
"epoch": 0.036734693877551024,
"grad_norm": 0.027247965335845947,
"learning_rate": 0.00018,
"loss": 1.1636,
"step": 9
},
{
"epoch": 0.04081632653061224,
"grad_norm": 0.027813177555799484,
"learning_rate": 0.0002,
"loss": 1.1848,
"step": 10
},
{
"epoch": 0.044897959183673466,
"grad_norm": 0.027455640956759453,
"learning_rate": 0.00019999906115681734,
"loss": 1.2151,
"step": 11
},
{
"epoch": 0.04897959183673469,
"grad_norm": 0.026825131848454475,
"learning_rate": 0.0001999962446448979,
"loss": 1.1751,
"step": 12
},
{
"epoch": 0.053061224489795916,
"grad_norm": 0.026261834427714348,
"learning_rate": 0.0001999915505171269,
"loss": 1.0573,
"step": 13
},
{
"epoch": 0.05714285714285714,
"grad_norm": 0.022048471495509148,
"learning_rate": 0.0001999849788616454,
"loss": 1.1997,
"step": 14
},
{
"epoch": 0.061224489795918366,
"grad_norm": 0.01966056413948536,
"learning_rate": 0.00019997652980184843,
"loss": 1.1469,
"step": 15
},
{
"epoch": 0.0653061224489796,
"grad_norm": 0.021213103085756302,
"learning_rate": 0.00019996620349638285,
"loss": 1.15,
"step": 16
},
{
"epoch": 0.06938775510204082,
"grad_norm": 0.02238585613667965,
"learning_rate": 0.00019995400013914427,
"loss": 1.206,
"step": 17
},
{
"epoch": 0.07346938775510205,
"grad_norm": 0.02359834685921669,
"learning_rate": 0.0001999399199592735,
"loss": 1.3413,
"step": 18
},
{
"epoch": 0.07755102040816327,
"grad_norm": 0.022565221413969994,
"learning_rate": 0.00019992396322115213,
"loss": 1.1798,
"step": 19
},
{
"epoch": 0.08163265306122448,
"grad_norm": 0.02135898545384407,
"learning_rate": 0.0001999061302243977,
"loss": 1.1602,
"step": 20
},
{
"epoch": 0.08571428571428572,
"grad_norm": 0.02203250862658024,
"learning_rate": 0.00019988642130385788,
"loss": 1.0915,
"step": 21
},
{
"epoch": 0.08979591836734693,
"grad_norm": 0.01716572232544422,
"learning_rate": 0.00019986483682960445,
"loss": 1.1858,
"step": 22
},
{
"epoch": 0.09387755102040816,
"grad_norm": 0.01753074862062931,
"learning_rate": 0.00019984137720692612,
"loss": 1.1849,
"step": 23
},
{
"epoch": 0.09795918367346938,
"grad_norm": 0.017932750284671783,
"learning_rate": 0.00019981604287632102,
"loss": 1.1579,
"step": 24
},
{
"epoch": 0.10204081632653061,
"grad_norm": 0.017208363860845566,
"learning_rate": 0.00019978883431348845,
"loss": 1.1223,
"step": 25
},
{
"epoch": 0.10612244897959183,
"grad_norm": 0.01637221872806549,
"learning_rate": 0.00019975975202931982,
"loss": 1.1809,
"step": 26
},
{
"epoch": 0.11020408163265306,
"grad_norm": 0.016254756599664688,
"learning_rate": 0.0001997287965698893,
"loss": 1.1978,
"step": 27
},
{
"epoch": 0.11428571428571428,
"grad_norm": 0.016743820160627365,
"learning_rate": 0.00019969596851644327,
"loss": 1.1273,
"step": 28
},
{
"epoch": 0.11836734693877551,
"grad_norm": 0.020556606352329254,
"learning_rate": 0.0001996612684853896,
"loss": 1.1638,
"step": 29
},
{
"epoch": 0.12244897959183673,
"grad_norm": 0.019542187452316284,
"learning_rate": 0.00019962469712828614,
"loss": 1.2558,
"step": 30
},
{
"epoch": 0.12653061224489795,
"grad_norm": 0.01658390648663044,
"learning_rate": 0.00019958625513182815,
"loss": 1.2176,
"step": 31
},
{
"epoch": 0.1306122448979592,
"grad_norm": 0.017420461401343346,
"learning_rate": 0.00019954594321783584,
"loss": 1.2671,
"step": 32
},
{
"epoch": 0.1346938775510204,
"grad_norm": 0.01890096440911293,
"learning_rate": 0.00019950376214324052,
"loss": 1.1448,
"step": 33
},
{
"epoch": 0.13877551020408163,
"grad_norm": 0.016807271167635918,
"learning_rate": 0.00019945971270007043,
"loss": 1.2085,
"step": 34
},
{
"epoch": 0.14285714285714285,
"grad_norm": 0.015759294852614403,
"learning_rate": 0.00019941379571543596,
"loss": 1.2207,
"step": 35
},
{
"epoch": 0.1469387755102041,
"grad_norm": 0.016418293118476868,
"learning_rate": 0.00019936601205151414,
"loss": 1.1031,
"step": 36
},
{
"epoch": 0.1510204081632653,
"grad_norm": 0.017639920115470886,
"learning_rate": 0.00019931636260553224,
"loss": 1.1439,
"step": 37
},
{
"epoch": 0.15510204081632653,
"grad_norm": 0.01651432178914547,
"learning_rate": 0.00019926484830975113,
"loss": 1.1111,
"step": 38
},
{
"epoch": 0.15918367346938775,
"grad_norm": 0.016843752935528755,
"learning_rate": 0.0001992114701314478,
"loss": 1.2519,
"step": 39
},
{
"epoch": 0.16326530612244897,
"grad_norm": 0.016024339944124222,
"learning_rate": 0.00019915622907289694,
"loss": 1.2122,
"step": 40
},
{
"epoch": 0.1673469387755102,
"grad_norm": 0.017350338399410248,
"learning_rate": 0.00019909912617135244,
"loss": 1.2104,
"step": 41
},
{
"epoch": 0.17142857142857143,
"grad_norm": 0.01698286086320877,
"learning_rate": 0.00019904016249902763,
"loss": 1.2258,
"step": 42
},
{
"epoch": 0.17551020408163265,
"grad_norm": 0.016692234203219414,
"learning_rate": 0.00019897933916307543,
"loss": 1.1614,
"step": 43
},
{
"epoch": 0.17959183673469387,
"grad_norm": 0.01875966228544712,
"learning_rate": 0.00019891665730556725,
"loss": 1.2314,
"step": 44
},
{
"epoch": 0.1836734693877551,
"grad_norm": 0.016782566905021667,
"learning_rate": 0.00019885211810347184,
"loss": 1.2096,
"step": 45
},
{
"epoch": 0.18775510204081633,
"grad_norm": 0.017428990453481674,
"learning_rate": 0.00019878572276863294,
"loss": 1.0537,
"step": 46
},
{
"epoch": 0.19183673469387755,
"grad_norm": 0.016738982871174812,
"learning_rate": 0.00019871747254774673,
"loss": 1.2018,
"step": 47
},
{
"epoch": 0.19591836734693877,
"grad_norm": 0.018019411712884903,
"learning_rate": 0.0001986473687223383,
"loss": 1.2504,
"step": 48
},
{
"epoch": 0.2,
"grad_norm": 0.017473606392741203,
"learning_rate": 0.0001985754126087376,
"loss": 1.0481,
"step": 49
},
{
"epoch": 0.20408163265306123,
"grad_norm": 0.017300540581345558,
"learning_rate": 0.00019850160555805486,
"loss": 1.2109,
"step": 50
},
{
"epoch": 0.20816326530612245,
"grad_norm": 0.0182975921779871,
"learning_rate": 0.00019842594895615488,
"loss": 1.1598,
"step": 51
},
{
"epoch": 0.21224489795918366,
"grad_norm": 0.020277326926589012,
"learning_rate": 0.00019834844422363142,
"loss": 1.1177,
"step": 52
},
{
"epoch": 0.2163265306122449,
"grad_norm": 0.018494602292776108,
"learning_rate": 0.00019826909281578026,
"loss": 1.1037,
"step": 53
},
{
"epoch": 0.22040816326530613,
"grad_norm": 0.017682479694485664,
"learning_rate": 0.00019818789622257196,
"loss": 1.1358,
"step": 54
},
{
"epoch": 0.22448979591836735,
"grad_norm": 0.0186194758862257,
"learning_rate": 0.00019810485596862392,
"loss": 1.1403,
"step": 55
},
{
"epoch": 0.22857142857142856,
"grad_norm": 0.018025796860456467,
"learning_rate": 0.00019801997361317163,
"loss": 1.1959,
"step": 56
},
{
"epoch": 0.23265306122448978,
"grad_norm": 0.01819712296128273,
"learning_rate": 0.0001979332507500395,
"loss": 1.1181,
"step": 57
},
{
"epoch": 0.23673469387755103,
"grad_norm": 0.018991166725754738,
"learning_rate": 0.00019784468900761095,
"loss": 1.1393,
"step": 58
},
{
"epoch": 0.24081632653061225,
"grad_norm": 0.01914682239294052,
"learning_rate": 0.0001977542900487977,
"loss": 1.182,
"step": 59
},
{
"epoch": 0.24489795918367346,
"grad_norm": 0.01992950588464737,
"learning_rate": 0.00019766205557100868,
"loss": 1.1629,
"step": 60
},
{
"epoch": 0.24897959183673468,
"grad_norm": 0.024829212576150894,
"learning_rate": 0.00019756798730611813,
"loss": 1.0753,
"step": 61
},
{
"epoch": 0.2530612244897959,
"grad_norm": 0.02457556687295437,
"learning_rate": 0.00019747208702043296,
"loss": 1.1718,
"step": 62
},
{
"epoch": 0.2571428571428571,
"grad_norm": 0.019101744517683983,
"learning_rate": 0.0001973743565146599,
"loss": 1.1343,
"step": 63
},
{
"epoch": 0.2612244897959184,
"grad_norm": 0.020323360338807106,
"learning_rate": 0.00019727479762387116,
"loss": 1.1689,
"step": 64
},
{
"epoch": 0.2653061224489796,
"grad_norm": 0.018925843760371208,
"learning_rate": 0.00019717341221747056,
"loss": 1.2098,
"step": 65
},
{
"epoch": 0.2693877551020408,
"grad_norm": 0.019229738041758537,
"learning_rate": 0.00019707020219915806,
"loss": 1.1542,
"step": 66
},
{
"epoch": 0.27346938775510204,
"grad_norm": 0.019665885716676712,
"learning_rate": 0.00019696516950689404,
"loss": 1.1154,
"step": 67
},
{
"epoch": 0.27755102040816326,
"grad_norm": 0.019163204357028008,
"learning_rate": 0.0001968583161128631,
"loss": 1.1726,
"step": 68
},
{
"epoch": 0.2816326530612245,
"grad_norm": 0.019880875945091248,
"learning_rate": 0.00019674964402343684,
"loss": 1.1441,
"step": 69
},
{
"epoch": 0.2857142857142857,
"grad_norm": 0.019096847623586655,
"learning_rate": 0.00019663915527913625,
"loss": 1.2021,
"step": 70
},
{
"epoch": 0.2897959183673469,
"grad_norm": 0.019253911450505257,
"learning_rate": 0.00019652685195459344,
"loss": 1.1237,
"step": 71
},
{
"epoch": 0.2938775510204082,
"grad_norm": 0.020909370854496956,
"learning_rate": 0.00019641273615851257,
"loss": 1.0575,
"step": 72
},
{
"epoch": 0.2979591836734694,
"grad_norm": 0.019176874309778214,
"learning_rate": 0.00019629681003363044,
"loss": 1.0377,
"step": 73
},
{
"epoch": 0.3020408163265306,
"grad_norm": 0.019969960674643517,
"learning_rate": 0.00019617907575667602,
"loss": 1.1471,
"step": 74
},
{
"epoch": 0.30612244897959184,
"grad_norm": 0.019450997933745384,
"learning_rate": 0.00019605953553832988,
"loss": 1.2022,
"step": 75
},
{
"epoch": 0.31020408163265306,
"grad_norm": 0.02082456275820732,
"learning_rate": 0.00019593819162318232,
"loss": 1.0932,
"step": 76
},
{
"epoch": 0.3142857142857143,
"grad_norm": 0.020122263580560684,
"learning_rate": 0.00019581504628969154,
"loss": 1.0826,
"step": 77
},
{
"epoch": 0.3183673469387755,
"grad_norm": 0.020104922354221344,
"learning_rate": 0.00019569010185014062,
"loss": 1.1404,
"step": 78
},
{
"epoch": 0.3224489795918367,
"grad_norm": 0.021288864314556122,
"learning_rate": 0.00019556336065059432,
"loss": 1.0951,
"step": 79
},
{
"epoch": 0.32653061224489793,
"grad_norm": 0.02072795107960701,
"learning_rate": 0.00019543482507085482,
"loss": 1.0649,
"step": 80
},
{
"epoch": 0.3306122448979592,
"grad_norm": 0.020123451948165894,
"learning_rate": 0.00019530449752441718,
"loss": 1.1333,
"step": 81
},
{
"epoch": 0.3346938775510204,
"grad_norm": 0.020374584943056107,
"learning_rate": 0.00019517238045842404,
"loss": 1.147,
"step": 82
},
{
"epoch": 0.3346938775510204,
"eval_loss": 1.1397618055343628,
"eval_runtime": 120.3925,
"eval_samples_per_second": 20.284,
"eval_steps_per_second": 10.142,
"step": 82
},
{
"epoch": 0.33877551020408164,
"grad_norm": 0.0197721179574728,
"learning_rate": 0.00019503847635361944,
"loss": 1.0954,
"step": 83
},
{
"epoch": 0.34285714285714286,
"grad_norm": 0.021181972697377205,
"learning_rate": 0.00019490278772430256,
"loss": 1.1145,
"step": 84
},
{
"epoch": 0.3469387755102041,
"grad_norm": 0.021057790145277977,
"learning_rate": 0.00019476531711828027,
"loss": 1.2255,
"step": 85
},
{
"epoch": 0.3510204081632653,
"grad_norm": 0.0207724217325449,
"learning_rate": 0.00019462606711681936,
"loss": 1.0366,
"step": 86
},
{
"epoch": 0.3551020408163265,
"grad_norm": 0.021172018721699715,
"learning_rate": 0.00019448504033459818,
"loss": 1.0835,
"step": 87
},
{
"epoch": 0.35918367346938773,
"grad_norm": 0.02109033428132534,
"learning_rate": 0.00019434223941965738,
"loss": 1.0902,
"step": 88
},
{
"epoch": 0.363265306122449,
"grad_norm": 0.02078443393111229,
"learning_rate": 0.00019419766705335026,
"loss": 1.087,
"step": 89
},
{
"epoch": 0.3673469387755102,
"grad_norm": 0.020614784210920334,
"learning_rate": 0.0001940513259502924,
"loss": 1.0834,
"step": 90
},
{
"epoch": 0.37142857142857144,
"grad_norm": 0.020187893882393837,
"learning_rate": 0.0001939032188583108,
"loss": 1.1847,
"step": 91
},
{
"epoch": 0.37551020408163266,
"grad_norm": 0.020622489973902702,
"learning_rate": 0.0001937533485583921,
"loss": 1.1767,
"step": 92
},
{
"epoch": 0.3795918367346939,
"grad_norm": 0.02118833176791668,
"learning_rate": 0.00019360171786463043,
"loss": 1.1491,
"step": 93
},
{
"epoch": 0.3836734693877551,
"grad_norm": 0.021402837708592415,
"learning_rate": 0.00019344832962417475,
"loss": 1.1547,
"step": 94
},
{
"epoch": 0.3877551020408163,
"grad_norm": 0.02148307114839554,
"learning_rate": 0.0001932931867171751,
"loss": 1.2062,
"step": 95
},
{
"epoch": 0.39183673469387753,
"grad_norm": 0.022203955799341202,
"learning_rate": 0.00019313629205672868,
"loss": 1.1271,
"step": 96
},
{
"epoch": 0.39591836734693875,
"grad_norm": 0.02081882208585739,
"learning_rate": 0.00019297764858882514,
"loss": 1.0965,
"step": 97
},
{
"epoch": 0.4,
"grad_norm": 0.021749386563897133,
"learning_rate": 0.00019281725929229127,
"loss": 1.11,
"step": 98
},
{
"epoch": 0.40408163265306124,
"grad_norm": 0.02319493517279625,
"learning_rate": 0.00019265512717873498,
"loss": 1.0196,
"step": 99
},
{
"epoch": 0.40816326530612246,
"grad_norm": 0.021901512518525124,
"learning_rate": 0.0001924912552924889,
"loss": 1.1552,
"step": 100
},
{
"epoch": 0.4122448979591837,
"grad_norm": 0.02368365228176117,
"learning_rate": 0.00019232564671055306,
"loss": 1.2047,
"step": 101
},
{
"epoch": 0.4163265306122449,
"grad_norm": 0.022438503801822662,
"learning_rate": 0.00019215830454253724,
"loss": 1.1429,
"step": 102
},
{
"epoch": 0.4204081632653061,
"grad_norm": 0.022685807198286057,
"learning_rate": 0.00019198923193060254,
"loss": 1.0717,
"step": 103
},
{
"epoch": 0.42448979591836733,
"grad_norm": 0.023902015760540962,
"learning_rate": 0.00019181843204940232,
"loss": 1.1139,
"step": 104
},
{
"epoch": 0.42857142857142855,
"grad_norm": 0.022928839549422264,
"learning_rate": 0.00019164590810602262,
"loss": 1.1515,
"step": 105
},
{
"epoch": 0.4326530612244898,
"grad_norm": 0.02210794948041439,
"learning_rate": 0.00019147166333992205,
"loss": 1.1762,
"step": 106
},
{
"epoch": 0.43673469387755104,
"grad_norm": 0.023219434544444084,
"learning_rate": 0.00019129570102287082,
"loss": 1.1925,
"step": 107
},
{
"epoch": 0.44081632653061226,
"grad_norm": 0.023004446178674698,
"learning_rate": 0.00019111802445888936,
"loss": 1.1449,
"step": 108
},
{
"epoch": 0.4448979591836735,
"grad_norm": 0.023740626871585846,
"learning_rate": 0.00019093863698418627,
"loss": 1.2179,
"step": 109
},
{
"epoch": 0.4489795918367347,
"grad_norm": 0.06276971846818924,
"learning_rate": 0.00019075754196709572,
"loss": 1.2497,
"step": 110
},
{
"epoch": 0.4530612244897959,
"grad_norm": 0.023355931043624878,
"learning_rate": 0.00019057474280801415,
"loss": 1.1625,
"step": 111
},
{
"epoch": 0.45714285714285713,
"grad_norm": 0.03020629473030567,
"learning_rate": 0.00019039024293933645,
"loss": 1.0821,
"step": 112
},
{
"epoch": 0.46122448979591835,
"grad_norm": 0.022392934188246727,
"learning_rate": 0.00019020404582539152,
"loss": 1.1074,
"step": 113
},
{
"epoch": 0.46530612244897956,
"grad_norm": 0.023577727377414703,
"learning_rate": 0.00019001615496237712,
"loss": 1.1128,
"step": 114
},
{
"epoch": 0.46938775510204084,
"grad_norm": 0.023507297039031982,
"learning_rate": 0.00018982657387829445,
"loss": 1.1345,
"step": 115
},
{
"epoch": 0.47346938775510206,
"grad_norm": 0.023036476224660873,
"learning_rate": 0.0001896353061328816,
"loss": 1.128,
"step": 116
},
{
"epoch": 0.4775510204081633,
"grad_norm": 0.023640332743525505,
"learning_rate": 0.00018944235531754698,
"loss": 1.062,
"step": 117
},
{
"epoch": 0.4816326530612245,
"grad_norm": 0.022972116246819496,
"learning_rate": 0.00018924772505530174,
"loss": 1.1024,
"step": 118
},
{
"epoch": 0.4857142857142857,
"grad_norm": 0.023795459419488907,
"learning_rate": 0.00018905141900069178,
"loss": 1.1595,
"step": 119
},
{
"epoch": 0.4897959183673469,
"grad_norm": 0.02379484474658966,
"learning_rate": 0.00018885344083972914,
"loss": 1.1492,
"step": 120
},
{
"epoch": 0.49387755102040815,
"grad_norm": 0.023673737421631813,
"learning_rate": 0.00018865379428982271,
"loss": 1.1571,
"step": 121
},
{
"epoch": 0.49795918367346936,
"grad_norm": 0.02365570329129696,
"learning_rate": 0.00018845248309970854,
"loss": 1.1216,
"step": 122
},
{
"epoch": 0.5020408163265306,
"grad_norm": 0.025217821821570396,
"learning_rate": 0.00018824951104937947,
"loss": 1.1422,
"step": 123
},
{
"epoch": 0.5061224489795918,
"grad_norm": 0.02457055076956749,
"learning_rate": 0.00018804488195001392,
"loss": 1.119,
"step": 124
},
{
"epoch": 0.5102040816326531,
"grad_norm": 0.023103831335902214,
"learning_rate": 0.00018783859964390464,
"loss": 1.153,
"step": 125
},
{
"epoch": 0.5142857142857142,
"grad_norm": 0.02386569045484066,
"learning_rate": 0.00018763066800438636,
"loss": 1.137,
"step": 126
},
{
"epoch": 0.5183673469387755,
"grad_norm": 0.024395650252699852,
"learning_rate": 0.00018742109093576313,
"loss": 1.1592,
"step": 127
},
{
"epoch": 0.5224489795918368,
"grad_norm": 0.024944225326180458,
"learning_rate": 0.000187209872373235,
"loss": 1.1395,
"step": 128
},
{
"epoch": 0.5265306122448979,
"grad_norm": 0.023946771398186684,
"learning_rate": 0.00018699701628282407,
"loss": 1.1779,
"step": 129
},
{
"epoch": 0.5306122448979592,
"grad_norm": 0.02391199767589569,
"learning_rate": 0.00018678252666130013,
"loss": 1.1299,
"step": 130
},
{
"epoch": 0.5346938775510204,
"grad_norm": 0.0230008065700531,
"learning_rate": 0.00018656640753610563,
"loss": 1.1087,
"step": 131
},
{
"epoch": 0.5387755102040817,
"grad_norm": 0.023263461887836456,
"learning_rate": 0.0001863486629652799,
"loss": 1.0835,
"step": 132
},
{
"epoch": 0.5428571428571428,
"grad_norm": 0.024622686207294464,
"learning_rate": 0.000186129297037383,
"loss": 1.0911,
"step": 133
},
{
"epoch": 0.5469387755102041,
"grad_norm": 0.023301225155591965,
"learning_rate": 0.0001859083138714191,
"loss": 1.1345,
"step": 134
},
{
"epoch": 0.5510204081632653,
"grad_norm": 0.023047855123877525,
"learning_rate": 0.00018568571761675893,
"loss": 1.1353,
"step": 135
},
{
"epoch": 0.5551020408163265,
"grad_norm": 0.025225356221199036,
"learning_rate": 0.00018546151245306205,
"loss": 1.149,
"step": 136
},
{
"epoch": 0.5591836734693878,
"grad_norm": 0.024374982342123985,
"learning_rate": 0.00018523570259019827,
"loss": 1.0618,
"step": 137
},
{
"epoch": 0.563265306122449,
"grad_norm": 0.024517694488167763,
"learning_rate": 0.00018500829226816853,
"loss": 1.148,
"step": 138
},
{
"epoch": 0.5673469387755102,
"grad_norm": 0.022940896451473236,
"learning_rate": 0.0001847792857570255,
"loss": 1.0824,
"step": 139
},
{
"epoch": 0.5714285714285714,
"grad_norm": 0.025163279846310616,
"learning_rate": 0.0001845486873567932,
"loss": 1.112,
"step": 140
},
{
"epoch": 0.5755102040816327,
"grad_norm": 0.024664800614118576,
"learning_rate": 0.00018431650139738633,
"loss": 1.1398,
"step": 141
},
{
"epoch": 0.5795918367346938,
"grad_norm": 0.02429143153131008,
"learning_rate": 0.000184082732238529,
"loss": 1.2003,
"step": 142
},
{
"epoch": 0.5836734693877551,
"grad_norm": 0.026087280362844467,
"learning_rate": 0.00018384738426967283,
"loss": 1.1151,
"step": 143
},
{
"epoch": 0.5877551020408164,
"grad_norm": 0.025691909715533257,
"learning_rate": 0.00018361046190991455,
"loss": 1.1526,
"step": 144
},
{
"epoch": 0.5918367346938775,
"grad_norm": 0.024672340601682663,
"learning_rate": 0.00018337196960791302,
"loss": 1.1781,
"step": 145
},
{
"epoch": 0.5959183673469388,
"grad_norm": 0.025797029957175255,
"learning_rate": 0.00018313191184180568,
"loss": 1.2053,
"step": 146
},
{
"epoch": 0.6,
"grad_norm": 0.025542639195919037,
"learning_rate": 0.0001828902931191244,
"loss": 1.0819,
"step": 147
},
{
"epoch": 0.6040816326530613,
"grad_norm": 0.025126414373517036,
"learning_rate": 0.0001826471179767111,
"loss": 1.1935,
"step": 148
},
{
"epoch": 0.6081632653061224,
"grad_norm": 0.02403743751347065,
"learning_rate": 0.0001824023909806322,
"loss": 1.1151,
"step": 149
},
{
"epoch": 0.6122448979591837,
"grad_norm": 0.025347478687763214,
"learning_rate": 0.00018215611672609317,
"loss": 1.0863,
"step": 150
},
{
"epoch": 0.6163265306122448,
"grad_norm": 0.024234246462583542,
"learning_rate": 0.00018190829983735207,
"loss": 0.9933,
"step": 151
},
{
"epoch": 0.6204081632653061,
"grad_norm": 0.02541242353618145,
"learning_rate": 0.00018165894496763286,
"loss": 1.016,
"step": 152
},
{
"epoch": 0.6244897959183674,
"grad_norm": 0.025475289672613144,
"learning_rate": 0.00018140805679903795,
"loss": 1.1199,
"step": 153
},
{
"epoch": 0.6285714285714286,
"grad_norm": 0.025047162547707558,
"learning_rate": 0.00018115564004246023,
"loss": 1.1866,
"step": 154
},
{
"epoch": 0.6326530612244898,
"grad_norm": 0.024512339383363724,
"learning_rate": 0.00018090169943749476,
"loss": 1.0407,
"step": 155
},
{
"epoch": 0.636734693877551,
"grad_norm": 0.027875879779458046,
"learning_rate": 0.0001806462397523496,
"loss": 1.1148,
"step": 156
},
{
"epoch": 0.6408163265306123,
"grad_norm": 0.027647798880934715,
"learning_rate": 0.00018038926578375653,
"loss": 1.1748,
"step": 157
},
{
"epoch": 0.6448979591836734,
"grad_norm": 0.0258516576141119,
"learning_rate": 0.0001801307823568806,
"loss": 1.156,
"step": 158
},
{
"epoch": 0.6489795918367347,
"grad_norm": 0.02649206854403019,
"learning_rate": 0.00017987079432522996,
"loss": 1.0625,
"step": 159
},
{
"epoch": 0.6530612244897959,
"grad_norm": 0.027242302894592285,
"learning_rate": 0.00017960930657056438,
"loss": 1.0624,
"step": 160
},
{
"epoch": 0.6571428571428571,
"grad_norm": 0.026996418833732605,
"learning_rate": 0.0001793463240028038,
"loss": 1.0979,
"step": 161
},
{
"epoch": 0.6612244897959184,
"grad_norm": 0.025911618024110794,
"learning_rate": 0.00017908185155993605,
"loss": 1.0767,
"step": 162
},
{
"epoch": 0.6653061224489796,
"grad_norm": 0.02710540033876896,
"learning_rate": 0.0001788158942079241,
"loss": 0.9389,
"step": 163
},
{
"epoch": 0.6693877551020408,
"grad_norm": 0.026086492463946342,
"learning_rate": 0.00017854845694061292,
"loss": 1.1475,
"step": 164
},
{
"epoch": 0.6693877551020408,
"eval_loss": 1.1236058473587036,
"eval_runtime": 120.3474,
"eval_samples_per_second": 20.291,
"eval_steps_per_second": 10.146,
"step": 164
},
{
"epoch": 0.673469387755102,
"grad_norm": 0.02641923353075981,
"learning_rate": 0.00017827954477963557,
"loss": 1.1128,
"step": 165
},
{
"epoch": 0.6775510204081633,
"grad_norm": 0.026196755468845367,
"learning_rate": 0.00017800916277431908,
"loss": 1.059,
"step": 166
},
{
"epoch": 0.6816326530612244,
"grad_norm": 0.025523358955979347,
"learning_rate": 0.00017773731600158947,
"loss": 1.1618,
"step": 167
},
{
"epoch": 0.6857142857142857,
"grad_norm": 0.026307355612516403,
"learning_rate": 0.00017746400956587653,
"loss": 1.0585,
"step": 168
},
{
"epoch": 0.689795918367347,
"grad_norm": 0.026362139731645584,
"learning_rate": 0.00017718924859901793,
"loss": 1.0543,
"step": 169
},
{
"epoch": 0.6938775510204082,
"grad_norm": 0.02530169114470482,
"learning_rate": 0.0001769130382601629,
"loss": 1.1009,
"step": 170
},
{
"epoch": 0.6979591836734694,
"grad_norm": 0.026273014023900032,
"learning_rate": 0.0001766353837356753,
"loss": 1.1963,
"step": 171
},
{
"epoch": 0.7020408163265306,
"grad_norm": 0.026570703834295273,
"learning_rate": 0.00017635629023903627,
"loss": 1.1098,
"step": 172
},
{
"epoch": 0.7061224489795919,
"grad_norm": 0.02533114328980446,
"learning_rate": 0.0001760757630107464,
"loss": 1.2321,
"step": 173
},
{
"epoch": 0.710204081632653,
"grad_norm": 0.02558620274066925,
"learning_rate": 0.00017579380731822712,
"loss": 1.1174,
"step": 174
},
{
"epoch": 0.7142857142857143,
"grad_norm": 0.028141306713223457,
"learning_rate": 0.00017551042845572208,
"loss": 1.0609,
"step": 175
},
{
"epoch": 0.7183673469387755,
"grad_norm": 0.028484290465712547,
"learning_rate": 0.00017522563174419753,
"loss": 1.0779,
"step": 176
},
{
"epoch": 0.7224489795918367,
"grad_norm": 0.0262393057346344,
"learning_rate": 0.00017493942253124248,
"loss": 1.0889,
"step": 177
},
{
"epoch": 0.726530612244898,
"grad_norm": 0.029188042506575584,
"learning_rate": 0.00017465180619096832,
"loss": 1.1816,
"step": 178
},
{
"epoch": 0.7306122448979592,
"grad_norm": 0.02816024050116539,
"learning_rate": 0.00017436278812390786,
"loss": 1.155,
"step": 179
},
{
"epoch": 0.7346938775510204,
"grad_norm": 0.026962874457240105,
"learning_rate": 0.00017407237375691392,
"loss": 1.0784,
"step": 180
},
{
"epoch": 0.7387755102040816,
"grad_norm": 0.027746373787522316,
"learning_rate": 0.00017378056854305747,
"loss": 1.1896,
"step": 181
},
{
"epoch": 0.7428571428571429,
"grad_norm": 0.02716743014752865,
"learning_rate": 0.00017348737796152522,
"loss": 1.0722,
"step": 182
},
{
"epoch": 0.746938775510204,
"grad_norm": 0.02938067726790905,
"learning_rate": 0.0001731928075175168,
"loss": 1.0179,
"step": 183
},
{
"epoch": 0.7510204081632653,
"grad_norm": 0.03580696880817413,
"learning_rate": 0.00017289686274214118,
"loss": 1.0951,
"step": 184
},
{
"epoch": 0.7551020408163265,
"grad_norm": 0.02849019691348076,
"learning_rate": 0.0001725995491923131,
"loss": 1.1006,
"step": 185
},
{
"epoch": 0.7591836734693878,
"grad_norm": 0.026822512969374657,
"learning_rate": 0.00017230087245064858,
"loss": 1.1535,
"step": 186
},
{
"epoch": 0.763265306122449,
"grad_norm": 0.027751388028264046,
"learning_rate": 0.00017200083812536,
"loss": 1.0975,
"step": 187
},
{
"epoch": 0.7673469387755102,
"grad_norm": 0.027201151475310326,
"learning_rate": 0.00017169945185015106,
"loss": 1.1022,
"step": 188
},
{
"epoch": 0.7714285714285715,
"grad_norm": 0.026958104223012924,
"learning_rate": 0.00017139671928411072,
"loss": 1.14,
"step": 189
},
{
"epoch": 0.7755102040816326,
"grad_norm": 0.026919234544038773,
"learning_rate": 0.00017109264611160708,
"loss": 1.1056,
"step": 190
},
{
"epoch": 0.7795918367346939,
"grad_norm": 0.027088504284620285,
"learning_rate": 0.00017078723804218066,
"loss": 1.2057,
"step": 191
},
{
"epoch": 0.7836734693877551,
"grad_norm": 0.027464529499411583,
"learning_rate": 0.00017048050081043713,
"loss": 1.1709,
"step": 192
},
{
"epoch": 0.7877551020408163,
"grad_norm": 0.028090914711356163,
"learning_rate": 0.0001701724401759397,
"loss": 1.1489,
"step": 193
},
{
"epoch": 0.7918367346938775,
"grad_norm": 0.026381971314549446,
"learning_rate": 0.00016986306192310084,
"loss": 1.0724,
"step": 194
},
{
"epoch": 0.7959183673469388,
"grad_norm": 0.027733104303479195,
"learning_rate": 0.00016955237186107387,
"loss": 1.1281,
"step": 195
},
{
"epoch": 0.8,
"grad_norm": 0.026476366445422173,
"learning_rate": 0.0001692403758236437,
"loss": 1.19,
"step": 196
},
{
"epoch": 0.8040816326530612,
"grad_norm": 0.02931978367269039,
"learning_rate": 0.0001689270796691174,
"loss": 1.0904,
"step": 197
},
{
"epoch": 0.8081632653061225,
"grad_norm": 0.02770647406578064,
"learning_rate": 0.00016861248928021411,
"loss": 1.1499,
"step": 198
},
{
"epoch": 0.8122448979591836,
"grad_norm": 0.02689889818429947,
"learning_rate": 0.00016829661056395474,
"loss": 1.1525,
"step": 199
},
{
"epoch": 0.8163265306122449,
"grad_norm": 0.02684464119374752,
"learning_rate": 0.0001679794494515508,
"loss": 1.0428,
"step": 200
},
{
"epoch": 0.8204081632653061,
"grad_norm": 0.026796666905283928,
"learning_rate": 0.0001676610118982933,
"loss": 1.1829,
"step": 201
},
{
"epoch": 0.8244897959183674,
"grad_norm": 0.028474919497966766,
"learning_rate": 0.00016734130388344073,
"loss": 1.0853,
"step": 202
},
{
"epoch": 0.8285714285714286,
"grad_norm": 0.027959391474723816,
"learning_rate": 0.00016702033141010694,
"loss": 1.0554,
"step": 203
},
{
"epoch": 0.8326530612244898,
"grad_norm": 0.027617380023002625,
"learning_rate": 0.00016669810050514827,
"loss": 1.2237,
"step": 204
},
{
"epoch": 0.8367346938775511,
"grad_norm": 0.028105339035391808,
"learning_rate": 0.00016637461721905045,
"loss": 1.0778,
"step": 205
},
{
"epoch": 0.8408163265306122,
"grad_norm": 0.028216930106282234,
"learning_rate": 0.00016604988762581512,
"loss": 1.1507,
"step": 206
},
{
"epoch": 0.8448979591836735,
"grad_norm": 0.026569265872240067,
"learning_rate": 0.00016572391782284547,
"loss": 1.0134,
"step": 207
},
{
"epoch": 0.8489795918367347,
"grad_norm": 0.027188677340745926,
"learning_rate": 0.00016539671393083215,
"loss": 1.0784,
"step": 208
},
{
"epoch": 0.8530612244897959,
"grad_norm": 0.02793797291815281,
"learning_rate": 0.00016506828209363796,
"loss": 1.1477,
"step": 209
},
{
"epoch": 0.8571428571428571,
"grad_norm": 0.027332162484526634,
"learning_rate": 0.00016473862847818277,
"loss": 1.1653,
"step": 210
},
{
"epoch": 0.8612244897959184,
"grad_norm": 0.02860257215797901,
"learning_rate": 0.00016440775927432753,
"loss": 1.1599,
"step": 211
},
{
"epoch": 0.8653061224489796,
"grad_norm": 0.028649616986513138,
"learning_rate": 0.0001640756806947582,
"loss": 1.1796,
"step": 212
},
{
"epoch": 0.8693877551020408,
"grad_norm": 0.027630403637886047,
"learning_rate": 0.000163742398974869,
"loss": 1.1315,
"step": 213
},
{
"epoch": 0.8734693877551021,
"grad_norm": 0.027829712256789207,
"learning_rate": 0.00016340792037264527,
"loss": 1.0573,
"step": 214
},
{
"epoch": 0.8775510204081632,
"grad_norm": 0.027864307165145874,
"learning_rate": 0.00016307225116854622,
"loss": 1.083,
"step": 215
},
{
"epoch": 0.8816326530612245,
"grad_norm": 0.028005510568618774,
"learning_rate": 0.0001627353976653866,
"loss": 1.238,
"step": 216
},
{
"epoch": 0.8857142857142857,
"grad_norm": 0.02755396068096161,
"learning_rate": 0.00016239736618821885,
"loss": 1.1124,
"step": 217
},
{
"epoch": 0.889795918367347,
"grad_norm": 0.026409147307276726,
"learning_rate": 0.00016205816308421386,
"loss": 1.1627,
"step": 218
},
{
"epoch": 0.8938775510204081,
"grad_norm": 0.02800268866121769,
"learning_rate": 0.00016171779472254206,
"loss": 1.0921,
"step": 219
},
{
"epoch": 0.8979591836734694,
"grad_norm": 0.02667434886097908,
"learning_rate": 0.00016137626749425377,
"loss": 1.207,
"step": 220
},
{
"epoch": 0.9020408163265307,
"grad_norm": 0.027823835611343384,
"learning_rate": 0.0001610335878121592,
"loss": 1.0776,
"step": 221
},
{
"epoch": 0.9061224489795918,
"grad_norm": 0.02820996195077896,
"learning_rate": 0.000160689762110708,
"loss": 1.0903,
"step": 222
},
{
"epoch": 0.9102040816326531,
"grad_norm": 0.02725161425769329,
"learning_rate": 0.00016034479684586854,
"loss": 1.1587,
"step": 223
},
{
"epoch": 0.9142857142857143,
"grad_norm": 0.027408665046095848,
"learning_rate": 0.0001599986984950065,
"loss": 1.1058,
"step": 224
},
{
"epoch": 0.9183673469387755,
"grad_norm": 0.027829701080918312,
"learning_rate": 0.00015965147355676343,
"loss": 1.1037,
"step": 225
},
{
"epoch": 0.9224489795918367,
"grad_norm": 0.027295254170894623,
"learning_rate": 0.0001593031285509347,
"loss": 1.0681,
"step": 226
},
{
"epoch": 0.926530612244898,
"grad_norm": 0.028630707412958145,
"learning_rate": 0.00015895367001834694,
"loss": 1.1282,
"step": 227
},
{
"epoch": 0.9306122448979591,
"grad_norm": 0.02818943001329899,
"learning_rate": 0.0001586031045207354,
"loss": 1.1101,
"step": 228
},
{
"epoch": 0.9346938775510204,
"grad_norm": 0.028520913794636726,
"learning_rate": 0.0001582514386406206,
"loss": 1.1716,
"step": 229
},
{
"epoch": 0.9387755102040817,
"grad_norm": 0.027190426364541054,
"learning_rate": 0.0001578986789811849,
"loss": 1.0615,
"step": 230
},
{
"epoch": 0.9428571428571428,
"grad_norm": 0.028909413143992424,
"learning_rate": 0.00015754483216614821,
"loss": 1.0757,
"step": 231
},
{
"epoch": 0.9469387755102041,
"grad_norm": 0.028231430798768997,
"learning_rate": 0.000157189904839644,
"loss": 1.14,
"step": 232
},
{
"epoch": 0.9510204081632653,
"grad_norm": 0.02774875983595848,
"learning_rate": 0.00015683390366609426,
"loss": 1.0429,
"step": 233
},
{
"epoch": 0.9551020408163265,
"grad_norm": 0.030419837683439255,
"learning_rate": 0.00015647683533008455,
"loss": 1.113,
"step": 234
},
{
"epoch": 0.9591836734693877,
"grad_norm": 0.028292929753661156,
"learning_rate": 0.00015611870653623825,
"loss": 1.1127,
"step": 235
},
{
"epoch": 0.963265306122449,
"grad_norm": 0.028285473585128784,
"learning_rate": 0.00015575952400909092,
"loss": 1.0493,
"step": 236
},
{
"epoch": 0.9673469387755103,
"grad_norm": 0.028861958533525467,
"learning_rate": 0.00015539929449296392,
"loss": 1.0879,
"step": 237
},
{
"epoch": 0.9714285714285714,
"grad_norm": 0.02901526540517807,
"learning_rate": 0.00015503802475183773,
"loss": 1.21,
"step": 238
},
{
"epoch": 0.9755102040816327,
"grad_norm": 0.02809295803308487,
"learning_rate": 0.00015467572156922503,
"loss": 1.1949,
"step": 239
},
{
"epoch": 0.9795918367346939,
"grad_norm": 0.03164355456829071,
"learning_rate": 0.00015431239174804328,
"loss": 1.068,
"step": 240
},
{
"epoch": 0.9836734693877551,
"grad_norm": 0.029855625703930855,
"learning_rate": 0.00015394804211048702,
"loss": 1.1485,
"step": 241
},
{
"epoch": 0.9877551020408163,
"grad_norm": 0.027486370876431465,
"learning_rate": 0.00015358267949789966,
"loss": 1.0449,
"step": 242
},
{
"epoch": 0.9918367346938776,
"grad_norm": 0.02765297144651413,
"learning_rate": 0.0001532163107706452,
"loss": 1.0341,
"step": 243
},
{
"epoch": 0.9959183673469387,
"grad_norm": 0.02840602956712246,
"learning_rate": 0.0001528489428079793,
"loss": 1.2001,
"step": 244
},
{
"epoch": 1.0,
"grad_norm": 0.028239021077752113,
"learning_rate": 0.00015248058250792008,
"loss": 1.1528,
"step": 245
}
],
"logging_steps": 1,
"max_steps": 735,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 245,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 2.954959055015117e+17,
"train_batch_size": 2,
"trial_name": null,
"trial_params": null
}