smollm2-whisper-turbo-uvx-7200 / trainer_state.json
reach-vb's picture
reach-vb HF staff
Upload folder using huggingface_hub
ee32550 verified
{
"best_metric": null,
"best_model_checkpoint": null,
"epoch": 0.49996528018887576,
"eval_steps": 1000,
"global_step": 7200,
"is_hyper_param_search": false,
"is_local_process_zero": true,
"is_world_process_zero": true,
"log_history": [
{
"epoch": 6.943962224845496e-05,
"grad_norm": 9.8125,
"learning_rate": 2e-06,
"loss": 0.7188,
"step": 1
},
{
"epoch": 0.006943962224845497,
"grad_norm": 0.1376953125,
"learning_rate": 0.0002,
"loss": 0.3907,
"step": 100
},
{
"epoch": 0.013887924449690994,
"grad_norm": 0.0849609375,
"learning_rate": 0.0004,
"loss": 0.2341,
"step": 200
},
{
"epoch": 0.02083188667453649,
"grad_norm": 0.146484375,
"learning_rate": 0.0006,
"loss": 0.2181,
"step": 300
},
{
"epoch": 0.027775848899381988,
"grad_norm": 0.10546875,
"learning_rate": 0.0008,
"loss": 0.197,
"step": 400
},
{
"epoch": 0.034719811124227486,
"grad_norm": 0.11376953125,
"learning_rate": 0.001,
"loss": 0.1695,
"step": 500
},
{
"epoch": 0.04166377334907298,
"grad_norm": 0.0849609375,
"learning_rate": 0.0012,
"loss": 0.14,
"step": 600
},
{
"epoch": 0.04860773557391848,
"grad_norm": 0.078125,
"learning_rate": 0.0014,
"loss": 0.1231,
"step": 700
},
{
"epoch": 0.055551697798763976,
"grad_norm": 0.0712890625,
"learning_rate": 0.0016,
"loss": 0.1141,
"step": 800
},
{
"epoch": 0.06249566002360947,
"grad_norm": 0.06298828125,
"learning_rate": 0.0018000000000000002,
"loss": 0.1077,
"step": 900
},
{
"epoch": 0.06943962224845497,
"grad_norm": 0.0615234375,
"learning_rate": 0.002,
"loss": 0.1045,
"step": 1000
},
{
"epoch": 0.06943962224845497,
"eval_covost2-en-de_loss": 1.4858413934707642,
"eval_covost2-en-de_runtime": 32.1589,
"eval_covost2-en-de_samples_per_second": 1.99,
"eval_covost2-en-de_steps_per_second": 0.062,
"step": 1000
},
{
"epoch": 0.06943962224845497,
"eval_covost2-zh-en_loss": 2.7152516841888428,
"eval_covost2-zh-en_runtime": 31.3842,
"eval_covost2-zh-en_samples_per_second": 2.039,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 1000
},
{
"epoch": 0.06943962224845497,
"eval_peoplespeech-clean-transcription_loss": 2.0398874282836914,
"eval_peoplespeech-clean-transcription_runtime": 32.088,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.995,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
"step": 1000
},
{
"epoch": 0.07638358447330046,
"grad_norm": 0.053955078125,
"learning_rate": 0.001999725185109816,
"loss": 0.101,
"step": 1100
},
{
"epoch": 0.08332754669814596,
"grad_norm": 0.0517578125,
"learning_rate": 0.0019989008914857113,
"loss": 0.0956,
"step": 1200
},
{
"epoch": 0.09027150892299146,
"grad_norm": 0.04443359375,
"learning_rate": 0.00199752757218401,
"loss": 0.0936,
"step": 1300
},
{
"epoch": 0.09721547114783696,
"grad_norm": 0.0390625,
"learning_rate": 0.001995605982021898,
"loss": 0.0917,
"step": 1400
},
{
"epoch": 0.10415943337268245,
"grad_norm": 0.0517578125,
"learning_rate": 0.0019931371771625545,
"loss": 0.0894,
"step": 1500
},
{
"epoch": 0.11110339559752795,
"grad_norm": 0.0419921875,
"learning_rate": 0.001990122514534651,
"loss": 0.0868,
"step": 1600
},
{
"epoch": 0.11804735782237345,
"grad_norm": 0.039306640625,
"learning_rate": 0.0019865636510865464,
"loss": 0.0861,
"step": 1700
},
{
"epoch": 0.12499132004721894,
"grad_norm": 0.047119140625,
"learning_rate": 0.001982462542875576,
"loss": 0.0854,
"step": 1800
},
{
"epoch": 0.13193528227206444,
"grad_norm": 0.0390625,
"learning_rate": 0.001977821443992945,
"loss": 0.0837,
"step": 1900
},
{
"epoch": 0.13887924449690994,
"grad_norm": 0.04052734375,
"learning_rate": 0.001972642905324813,
"loss": 0.0818,
"step": 2000
},
{
"epoch": 0.13887924449690994,
"eval_covost2-en-de_loss": 1.4137890338897705,
"eval_covost2-en-de_runtime": 32.5714,
"eval_covost2-en-de_samples_per_second": 1.965,
"eval_covost2-en-de_steps_per_second": 0.061,
"step": 2000
},
{
"epoch": 0.13887924449690994,
"eval_covost2-zh-en_loss": 2.667837381362915,
"eval_covost2-zh-en_runtime": 31.1685,
"eval_covost2-zh-en_samples_per_second": 2.053,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 2000
},
{
"epoch": 0.13887924449690994,
"eval_peoplespeech-clean-transcription_loss": 1.835880160331726,
"eval_peoplespeech-clean-transcription_runtime": 32.0265,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.998,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
"step": 2000
},
{
"epoch": 0.14582320672175544,
"grad_norm": 0.039794921875,
"learning_rate": 0.0019669297731502505,
"loss": 0.0813,
"step": 2100
},
{
"epoch": 0.15276716894660092,
"grad_norm": 0.03515625,
"learning_rate": 0.00196068518757684,
"loss": 0.0811,
"step": 2200
},
{
"epoch": 0.15971113117144642,
"grad_norm": 0.04443359375,
"learning_rate": 0.001953912580814779,
"loss": 0.0793,
"step": 2300
},
{
"epoch": 0.16665509339629192,
"grad_norm": 0.037841796875,
"learning_rate": 0.0019466156752904343,
"loss": 0.0788,
"step": 2400
},
{
"epoch": 0.17359905562113742,
"grad_norm": 0.04052734375,
"learning_rate": 0.0019387984816003866,
"loss": 0.0783,
"step": 2500
},
{
"epoch": 0.18054301784598292,
"grad_norm": 0.03466796875,
"learning_rate": 0.0019304652963070869,
"loss": 0.0772,
"step": 2600
},
{
"epoch": 0.18748698007082842,
"grad_norm": 0.036376953125,
"learning_rate": 0.0019216206995773372,
"loss": 0.0771,
"step": 2700
},
{
"epoch": 0.19443094229567393,
"grad_norm": 0.0400390625,
"learning_rate": 0.0019122695526648968,
"loss": 0.0766,
"step": 2800
},
{
"epoch": 0.2013749045205194,
"grad_norm": 0.0361328125,
"learning_rate": 0.0019024169952385887,
"loss": 0.0753,
"step": 2900
},
{
"epoch": 0.2083188667453649,
"grad_norm": 0.03125,
"learning_rate": 0.0018920684425573864,
"loss": 0.075,
"step": 3000
},
{
"epoch": 0.2083188667453649,
"eval_covost2-en-de_loss": 1.3828500509262085,
"eval_covost2-en-de_runtime": 32.2972,
"eval_covost2-en-de_samples_per_second": 1.982,
"eval_covost2-en-de_steps_per_second": 0.062,
"step": 3000
},
{
"epoch": 0.2083188667453649,
"eval_covost2-zh-en_loss": 2.646721839904785,
"eval_covost2-zh-en_runtime": 31.0128,
"eval_covost2-zh-en_samples_per_second": 2.064,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 3000
},
{
"epoch": 0.2083188667453649,
"eval_peoplespeech-clean-transcription_loss": 1.773127555847168,
"eval_peoplespeech-clean-transcription_runtime": 32.5192,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.968,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
"step": 3000
},
{
"epoch": 0.2152628289702104,
"grad_norm": 0.03369140625,
"learning_rate": 0.0018812295824940284,
"loss": 0.0743,
"step": 3100
},
{
"epoch": 0.2222067911950559,
"grad_norm": 0.034912109375,
"learning_rate": 0.0018699063724087904,
"loss": 0.074,
"step": 3200
},
{
"epoch": 0.2291507534199014,
"grad_norm": 0.033203125,
"learning_rate": 0.0018581050358751443,
"loss": 0.0742,
"step": 3300
},
{
"epoch": 0.2360947156447469,
"grad_norm": 0.03857421875,
"learning_rate": 0.0018458320592590974,
"loss": 0.0742,
"step": 3400
},
{
"epoch": 0.24303867786959238,
"grad_norm": 0.033935546875,
"learning_rate": 0.0018330941881540914,
"loss": 0.0728,
"step": 3500
},
{
"epoch": 0.24998264009443788,
"grad_norm": 0.031005859375,
"learning_rate": 0.0018198984236734246,
"loss": 0.0728,
"step": 3600
},
{
"epoch": 0.2569266023192834,
"grad_norm": 0.030029296875,
"learning_rate": 0.0018062520186022297,
"loss": 0.0714,
"step": 3700
},
{
"epoch": 0.2638705645441289,
"grad_norm": 0.02734375,
"learning_rate": 0.0017921624734111292,
"loss": 0.071,
"step": 3800
},
{
"epoch": 0.2708145267689744,
"grad_norm": 0.03271484375,
"learning_rate": 0.001777637532133752,
"loss": 0.0705,
"step": 3900
},
{
"epoch": 0.2777584889938199,
"grad_norm": 0.035400390625,
"learning_rate": 0.0017626851781103819,
"loss": 0.0714,
"step": 4000
},
{
"epoch": 0.2777584889938199,
"eval_covost2-en-de_loss": 1.3778624534606934,
"eval_covost2-en-de_runtime": 32.6181,
"eval_covost2-en-de_samples_per_second": 1.962,
"eval_covost2-en-de_steps_per_second": 0.061,
"step": 4000
},
{
"epoch": 0.2777584889938199,
"eval_covost2-zh-en_loss": 2.6438870429992676,
"eval_covost2-zh-en_runtime": 31.4603,
"eval_covost2-zh-en_samples_per_second": 2.034,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 4000
},
{
"epoch": 0.2777584889938199,
"eval_peoplespeech-clean-transcription_loss": 1.7361584901809692,
"eval_peoplespeech-clean-transcription_runtime": 32.442,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.973,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.062,
"step": 4000
},
{
"epoch": 0.2847024512186654,
"grad_norm": 0.0281982421875,
"learning_rate": 0.001747313629600077,
"loss": 0.0713,
"step": 4100
},
{
"epoch": 0.2916464134435109,
"grad_norm": 0.028076171875,
"learning_rate": 0.001731531335263669,
"loss": 0.0699,
"step": 4200
},
{
"epoch": 0.2985903756683564,
"grad_norm": 0.0277099609375,
"learning_rate": 0.0017153469695201276,
"loss": 0.0702,
"step": 4300
},
{
"epoch": 0.30553433789320184,
"grad_norm": 0.031982421875,
"learning_rate": 0.0016987694277788418,
"loss": 0.0692,
"step": 4400
},
{
"epoch": 0.31247830011804734,
"grad_norm": 0.02880859375,
"learning_rate": 0.001681807821550438,
"loss": 0.0686,
"step": 4500
},
{
"epoch": 0.31942226234289284,
"grad_norm": 0.0289306640625,
"learning_rate": 0.0016644714734388218,
"loss": 0.0698,
"step": 4600
},
{
"epoch": 0.32636622456773834,
"grad_norm": 0.029541015625,
"learning_rate": 0.0016467699120171987,
"loss": 0.0683,
"step": 4700
},
{
"epoch": 0.33331018679258384,
"grad_norm": 0.034423828125,
"learning_rate": 0.001628712866590885,
"loss": 0.0687,
"step": 4800
},
{
"epoch": 0.34025414901742934,
"grad_norm": 0.0289306640625,
"learning_rate": 0.0016103102618497923,
"loss": 0.0684,
"step": 4900
},
{
"epoch": 0.34719811124227484,
"grad_norm": 0.0263671875,
"learning_rate": 0.0015915722124135226,
"loss": 0.0681,
"step": 5000
},
{
"epoch": 0.34719811124227484,
"eval_covost2-en-de_loss": 1.3711879253387451,
"eval_covost2-en-de_runtime": 32.6293,
"eval_covost2-en-de_samples_per_second": 1.961,
"eval_covost2-en-de_steps_per_second": 0.061,
"step": 5000
},
{
"epoch": 0.34719811124227484,
"eval_covost2-zh-en_loss": 2.6346511840820312,
"eval_covost2-zh-en_runtime": 32.1513,
"eval_covost2-zh-en_samples_per_second": 1.991,
"eval_covost2-zh-en_steps_per_second": 0.062,
"step": 5000
},
{
"epoch": 0.34719811124227484,
"eval_peoplespeech-clean-transcription_loss": 1.7350472211837769,
"eval_peoplespeech-clean-transcription_runtime": 32.5813,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.964,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
"step": 5000
},
{
"epoch": 0.35414207346712034,
"grad_norm": 0.028076171875,
"learning_rate": 0.001572509017272072,
"loss": 0.0693,
"step": 5100
},
{
"epoch": 0.36108603569196585,
"grad_norm": 0.030517578125,
"learning_rate": 0.0015531311541251993,
"loss": 0.0683,
"step": 5200
},
{
"epoch": 0.36802999791681135,
"grad_norm": 0.031494140625,
"learning_rate": 0.0015334492736235703,
"loss": 0.0677,
"step": 5300
},
{
"epoch": 0.37497396014165685,
"grad_norm": 0.0284423828125,
"learning_rate": 0.0015134741935148419,
"loss": 0.0669,
"step": 5400
},
{
"epoch": 0.38191792236650235,
"grad_norm": 0.0302734375,
"learning_rate": 0.0014932168926979072,
"loss": 0.0669,
"step": 5500
},
{
"epoch": 0.38886188459134785,
"grad_norm": 0.0260009765625,
"learning_rate": 0.0014726885051885652,
"loss": 0.0666,
"step": 5600
},
{
"epoch": 0.3958058468161933,
"grad_norm": 0.033447265625,
"learning_rate": 0.0014519003139999338,
"loss": 0.0659,
"step": 5700
},
{
"epoch": 0.4027498090410388,
"grad_norm": 0.0283203125,
"learning_rate": 0.0014308637449409706,
"loss": 0.0653,
"step": 5800
},
{
"epoch": 0.4096937712658843,
"grad_norm": 0.0283203125,
"learning_rate": 0.0014095903603365066,
"loss": 0.0662,
"step": 5900
},
{
"epoch": 0.4166377334907298,
"grad_norm": 0.0267333984375,
"learning_rate": 0.0013880918526722496,
"loss": 0.0665,
"step": 6000
},
{
"epoch": 0.4166377334907298,
"eval_covost2-en-de_loss": 1.3651559352874756,
"eval_covost2-en-de_runtime": 32.5621,
"eval_covost2-en-de_samples_per_second": 1.965,
"eval_covost2-en-de_steps_per_second": 0.061,
"step": 6000
},
{
"epoch": 0.4166377334907298,
"eval_covost2-zh-en_loss": 2.6372551918029785,
"eval_covost2-zh-en_runtime": 31.028,
"eval_covost2-zh-en_samples_per_second": 2.063,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 6000
},
{
"epoch": 0.4166377334907298,
"eval_peoplespeech-clean-transcription_loss": 1.7209596633911133,
"eval_peoplespeech-clean-transcription_runtime": 32.6773,
"eval_peoplespeech-clean-transcription_samples_per_second": 1.959,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.061,
"step": 6000
},
{
"epoch": 0.4235816957155753,
"grad_norm": 0.028564453125,
"learning_rate": 0.0013663800381682463,
"loss": 0.0658,
"step": 6100
},
{
"epoch": 0.4305256579404208,
"grad_norm": 0.0299072265625,
"learning_rate": 0.0013444668502843329,
"loss": 0.0657,
"step": 6200
},
{
"epoch": 0.4374696201652663,
"grad_norm": 0.0296630859375,
"learning_rate": 0.0013223643331611537,
"loss": 0.0655,
"step": 6300
},
{
"epoch": 0.4444135823901118,
"grad_norm": 0.0286865234375,
"learning_rate": 0.001300084635000341,
"loss": 0.0654,
"step": 6400
},
{
"epoch": 0.4513575446149573,
"grad_norm": 0.028564453125,
"learning_rate": 0.0012776400013875004,
"loss": 0.0655,
"step": 6500
},
{
"epoch": 0.4583015068398028,
"grad_norm": 0.030029296875,
"learning_rate": 0.0012550427685616766,
"loss": 0.0648,
"step": 6600
},
{
"epoch": 0.4652454690646483,
"grad_norm": 0.037109375,
"learning_rate": 0.0012323053566349834,
"loss": 0.0654,
"step": 6700
},
{
"epoch": 0.4721894312894938,
"grad_norm": 0.029296875,
"learning_rate": 0.0012094402627661448,
"loss": 0.0643,
"step": 6800
},
{
"epoch": 0.47913339351433926,
"grad_norm": 0.030517578125,
"learning_rate": 0.0011864600542916813,
"loss": 0.0646,
"step": 6900
},
{
"epoch": 0.48607735573918476,
"grad_norm": 0.037353515625,
"learning_rate": 0.0011633773618185302,
"loss": 0.0642,
"step": 7000
},
{
"epoch": 0.48607735573918476,
"eval_covost2-en-de_loss": 1.3594859838485718,
"eval_covost2-en-de_runtime": 32.6807,
"eval_covost2-en-de_samples_per_second": 1.958,
"eval_covost2-en-de_steps_per_second": 0.061,
"step": 7000
},
{
"epoch": 0.48607735573918476,
"eval_covost2-zh-en_loss": 2.626713514328003,
"eval_covost2-zh-en_runtime": 31.0228,
"eval_covost2-zh-en_samples_per_second": 2.063,
"eval_covost2-zh-en_steps_per_second": 0.064,
"step": 7000
},
{
"epoch": 0.48607735573918476,
"eval_peoplespeech-clean-transcription_loss": 1.693739652633667,
"eval_peoplespeech-clean-transcription_runtime": 31.9776,
"eval_peoplespeech-clean-transcription_samples_per_second": 2.001,
"eval_peoplespeech-clean-transcription_steps_per_second": 0.063,
"step": 7000
},
{
"epoch": 0.49302131796403026,
"grad_norm": 0.02978515625,
"learning_rate": 0.0011402048722818862,
"loss": 0.0656,
"step": 7100
},
{
"epoch": 0.49996528018887576,
"grad_norm": 0.0281982421875,
"learning_rate": 0.0011169553219720827,
"loss": 0.064,
"step": 7200
}
],
"logging_steps": 100,
"max_steps": 14400,
"num_input_tokens_seen": 0,
"num_train_epochs": 1,
"save_steps": 3600,
"stateful_callbacks": {
"TrainerControl": {
"args": {
"should_epoch_stop": false,
"should_evaluate": false,
"should_log": false,
"should_save": true,
"should_training_stop": false
},
"attributes": {}
}
},
"total_flos": 1.8942929465072026e+18,
"train_batch_size": 96,
"trial_name": null,
"trial_params": null
}