Qwen-2.5VL-7B-Cold-Start / trainer_state.json
zli12321
Upload folder using huggingface_hub
0bc6e76 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 2.988610478359909,
"eval_steps": 500,
"global_step": 657,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 0.04555808656036447,
"grad_norm": 16.628382336058554,
"learning_rate": 1.5151515151515152e-06,
"loss": 1.5883,
"step": 10
},
{
"epoch": 0.09111617312072894,
"grad_norm": 3.221028223725061,
"learning_rate": 3.0303030303030305e-06,
"loss": 0.9317,
"step": 20
},
{
"epoch": 0.1366742596810934,
"grad_norm": 1.5103642702449496,
"learning_rate": 4.5454545454545455e-06,
"loss": 0.5336,
"step": 30
},
{
"epoch": 0.18223234624145787,
"grad_norm": 1.0681798496560502,
"learning_rate": 6.060606060606061e-06,
"loss": 0.4417,
"step": 40
},
{
"epoch": 0.22779043280182232,
"grad_norm": 0.9679921213671909,
"learning_rate": 7.5757575757575764e-06,
"loss": 0.3979,
"step": 50
},
{
"epoch": 0.2733485193621868,
"grad_norm": 0.926628742270625,
"learning_rate": 9.090909090909091e-06,
"loss": 0.3744,
"step": 60
},
{
"epoch": 0.31890660592255127,
"grad_norm": 0.9491261260284587,
"learning_rate": 9.998869765883566e-06,
"loss": 0.3486,
"step": 70
},
{
"epoch": 0.36446469248291574,
"grad_norm": 0.84920311370356,
"learning_rate": 9.986160499534318e-06,
"loss": 0.3453,
"step": 80
},
{
"epoch": 0.41002277904328016,
"grad_norm": 0.9789183018412179,
"learning_rate": 9.959365197965824e-06,
"loss": 0.3572,
"step": 90
},
{
"epoch": 0.45558086560364464,
"grad_norm": 0.9102455500916922,
"learning_rate": 9.918559558613344e-06,
"loss": 0.3356,
"step": 100
},
{
"epoch": 0.5011389521640092,
"grad_norm": 0.9258850899143908,
"learning_rate": 9.863858858486736e-06,
"loss": 0.3278,
"step": 110
},
{
"epoch": 0.5466970387243736,
"grad_norm": 0.8922762696366542,
"learning_rate": 9.795417628509857e-06,
"loss": 0.3268,
"step": 120
},
{
"epoch": 0.592255125284738,
"grad_norm": 0.9396732926347776,
"learning_rate": 9.713429216966624e-06,
"loss": 0.3108,
"step": 130
},
{
"epoch": 0.6378132118451025,
"grad_norm": 0.845601467481683,
"learning_rate": 9.618125243286989e-06,
"loss": 0.3164,
"step": 140
},
{
"epoch": 0.683371298405467,
"grad_norm": 0.8159514975969553,
"learning_rate": 9.50977494371594e-06,
"loss": 0.3127,
"step": 150
},
{
"epoch": 0.7289293849658315,
"grad_norm": 0.8778654186332652,
"learning_rate": 9.388684410713977e-06,
"loss": 0.3118,
"step": 160
},
{
"epoch": 0.7744874715261959,
"grad_norm": 0.8158041504461108,
"learning_rate": 9.255195728237837e-06,
"loss": 0.3033,
"step": 170
},
{
"epoch": 0.8200455580865603,
"grad_norm": 0.8241477933179293,
"learning_rate": 9.109686005344258e-06,
"loss": 0.298,
"step": 180
},
{
"epoch": 0.8656036446469249,
"grad_norm": 0.7836442953404367,
"learning_rate": 8.952566310846931e-06,
"loss": 0.2967,
"step": 190
},
{
"epoch": 0.9111617312072893,
"grad_norm": 0.8589198296868927,
"learning_rate": 8.784280512036235e-06,
"loss": 0.2908,
"step": 200
},
{
"epoch": 0.9567198177676538,
"grad_norm": 0.8976108880329798,
"learning_rate": 8.60530402074241e-06,
"loss": 0.288,
"step": 210
},
{
"epoch": 1.0,
"grad_norm": 0.8244261812910136,
"learning_rate": 8.416142450284565e-06,
"loss": 0.274,
"step": 220
},
{
"epoch": 1.0455580865603644,
"grad_norm": 0.7709445009431932,
"learning_rate": 8.217330187099689e-06,
"loss": 0.178,
"step": 230
},
{
"epoch": 1.0911161731207288,
"grad_norm": 0.7371669649380324,
"learning_rate": 8.009428881086836e-06,
"loss": 0.1706,
"step": 240
},
{
"epoch": 1.1366742596810935,
"grad_norm": 0.8325359027964357,
"learning_rate": 7.793025858931317e-06,
"loss": 0.1717,
"step": 250
},
{
"epoch": 1.182232346241458,
"grad_norm": 0.7810927662905283,
"learning_rate": 7.568732464891293e-06,
"loss": 0.1768,
"step": 260
},
{
"epoch": 1.2277904328018223,
"grad_norm": 0.7328567227225008,
"learning_rate": 7.33718233373407e-06,
"loss": 0.174,
"step": 270
},
{
"epoch": 1.2733485193621867,
"grad_norm": 0.829615067889057,
"learning_rate": 7.099029600701144e-06,
"loss": 0.1721,
"step": 280
},
{
"epoch": 1.3189066059225514,
"grad_norm": 0.7764413190939856,
"learning_rate": 6.854947053558849e-06,
"loss": 0.169,
"step": 290
},
{
"epoch": 1.3644646924829158,
"grad_norm": 0.7592152186492857,
"learning_rate": 6.6056242319551315e-06,
"loss": 0.1667,
"step": 300
},
{
"epoch": 1.4100227790432802,
"grad_norm": 0.7583979219386041,
"learning_rate": 6.3517654794518156e-06,
"loss": 0.1644,
"step": 310
},
{
"epoch": 1.4555808656036446,
"grad_norm": 0.7329938588396943,
"learning_rate": 6.094087953735423e-06,
"loss": 0.168,
"step": 320
},
{
"epoch": 1.501138952164009,
"grad_norm": 0.8196364714199265,
"learning_rate": 5.8333196006277536e-06,
"loss": 0.1617,
"step": 330
},
{
"epoch": 1.5466970387243735,
"grad_norm": 0.7565763296431312,
"learning_rate": 5.570197097619688e-06,
"loss": 0.1611,
"step": 340
},
{
"epoch": 1.592255125284738,
"grad_norm": 0.8868392512031471,
"learning_rate": 5.305463772737812e-06,
"loss": 0.1609,
"step": 350
},
{
"epoch": 1.6378132118451025,
"grad_norm": 0.8812463705246544,
"learning_rate": 5.039867504623084e-06,
"loss": 0.1677,
"step": 360
},
{
"epoch": 1.683371298405467,
"grad_norm": 0.7633777742059934,
"learning_rate": 4.774158609753908e-06,
"loss": 0.1512,
"step": 370
},
{
"epoch": 1.7289293849658316,
"grad_norm": 0.7820693250794272,
"learning_rate": 4.5090877227822424e-06,
"loss": 0.1572,
"step": 380
},
{
"epoch": 1.774487471526196,
"grad_norm": 0.7688524374447184,
"learning_rate": 4.245403675970877e-06,
"loss": 0.1617,
"step": 390
},
{
"epoch": 1.8200455580865604,
"grad_norm": 0.7254423021253176,
"learning_rate": 3.9838513837224814e-06,
"loss": 0.1519,
"step": 400
},
{
"epoch": 1.8656036446469249,
"grad_norm": 0.8076268890579671,
"learning_rate": 3.7251697381767373e-06,
"loss": 0.154,
"step": 410
},
{
"epoch": 1.9111617312072893,
"grad_norm": 0.7615743261772087,
"learning_rate": 3.4700895218205026e-06,
"loss": 0.1468,
"step": 420
},
{
"epoch": 1.9567198177676537,
"grad_norm": 0.7511849885826963,
"learning_rate": 3.2193313430079737e-06,
"loss": 0.147,
"step": 430
},
{
"epoch": 2.0,
"grad_norm": 0.7584213495047252,
"learning_rate": 2.9736036002230332e-06,
"loss": 0.1429,
"step": 440
},
{
"epoch": 2.0455580865603644,
"grad_norm": 0.6815225387343684,
"learning_rate": 2.7336004808348094e-06,
"loss": 0.0751,
"step": 450
},
{
"epoch": 2.091116173120729,
"grad_norm": 0.6174750652979677,
"learning_rate": 2.5000000000000015e-06,
"loss": 0.0711,
"step": 460
},
{
"epoch": 2.1366742596810933,
"grad_norm": 0.6236626143150734,
"learning_rate": 2.273462085252146e-06,
"loss": 0.0722,
"step": 470
},
{
"epoch": 2.1822323462414577,
"grad_norm": 0.5971939289865311,
"learning_rate": 2.0546267121888863e-06,
"loss": 0.0693,
"step": 480
},
{
"epoch": 2.2277904328018225,
"grad_norm": 0.631697145877715,
"learning_rate": 1.8441120965239912e-06,
"loss": 0.0713,
"step": 490
},
{
"epoch": 2.273348519362187,
"grad_norm": 0.6817300872785331,
"learning_rate": 1.642512947611622e-06,
"loss": 0.0729,
"step": 500
},
{
"epoch": 2.3189066059225514,
"grad_norm": 0.5682490177071967,
"learning_rate": 1.4503987883766857e-06,
"loss": 0.0656,
"step": 510
},
{
"epoch": 2.364464692482916,
"grad_norm": 0.5912819077013511,
"learning_rate": 1.2683123463975144e-06,
"loss": 0.0658,
"step": 520
},
{
"epoch": 2.41002277904328,
"grad_norm": 0.606352905873965,
"learning_rate": 1.0967680206861198e-06,
"loss": 0.0694,
"step": 530
},
{
"epoch": 2.4555808656036446,
"grad_norm": 0.6353412192229968,
"learning_rate": 9.362504284973683e-07,
"loss": 0.0701,
"step": 540
},
{
"epoch": 2.501138952164009,
"grad_norm": 0.6607014617898831,
"learning_rate": 7.872130362724422e-07,
"loss": 0.0691,
"step": 550
},
{
"epoch": 2.5466970387243735,
"grad_norm": 0.6527059303311876,
"learning_rate": 6.500768785841482e-07,
"loss": 0.0647,
"step": 560
},
{
"epoch": 2.592255125284738,
"grad_norm": 0.6203144834959747,
"learning_rate": 5.252293687031196e-07,
"loss": 0.0669,
"step": 570
},
{
"epoch": 2.6378132118451028,
"grad_norm": 0.6069382959627451,
"learning_rate": 4.130232041450866e-07,
"loss": 0.0624,
"step": 580
},
{
"epoch": 2.6833712984054667,
"grad_norm": 0.6304879856190249,
"learning_rate": 3.1377537029107174e-07,
"loss": 0.0655,
"step": 590
},
{
"epoch": 2.7289293849658316,
"grad_norm": 0.6325528536254185,
"learning_rate": 2.2776624489530664e-07,
"loss": 0.0667,
"step": 600
},
{
"epoch": 2.774487471526196,
"grad_norm": 0.6179538898003908,
"learning_rate": 1.55238806010668e-07,
"loss": 0.0644,
"step": 610
},
{
"epoch": 2.8200455580865604,
"grad_norm": 0.5965673740089573,
"learning_rate": 9.639794556925041e-08,
"loss": 0.0659,
"step": 620
},
{
"epoch": 2.865603644646925,
"grad_norm": 0.6382075048871372,
"learning_rate": 5.1409890557246876e-08,
"loss": 0.0645,
"step": 630
},
{
"epoch": 2.9111617312072893,
"grad_norm": 0.612194769673645,
"learning_rate": 2.0401733419315727e-08,
"loss": 0.063,
"step": 640
},
{
"epoch": 2.9567198177676537,
"grad_norm": 0.6571824905728761,
"learning_rate": 3.4610730190648423e-09,
"loss": 0.0647,
"step": 650
}
],
"logging_steps": 10,
"max_steps": 657,
"num_input_tokens_seen": 0,
"num_train_epochs": 3,
"save_steps": 500,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": true
},
"attributes": {}
}
},
"total_flos": 418097876697088.0,
"train_batch_size": 8,
"trial_name": null,
"trial_params": null
}