|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 2.0, |
|
"eval_steps": 500, |
|
"global_step": 958, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 0.0020876826722338203, |
|
"grad_norm": 0.8805100452450126, |
|
"learning_rate": 1.3793103448275863e-05, |
|
"loss": 3.4375, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0041753653444676405, |
|
"grad_norm": 0.8962056177448029, |
|
"learning_rate": 2.7586206896551727e-05, |
|
"loss": 3.4531, |
|
"step": 2 |
|
}, |
|
{ |
|
"epoch": 0.006263048016701462, |
|
"grad_norm": 0.8975533488683494, |
|
"learning_rate": 4.1379310344827587e-05, |
|
"loss": 3.375, |
|
"step": 3 |
|
}, |
|
{ |
|
"epoch": 0.008350730688935281, |
|
"grad_norm": 0.8562154080719621, |
|
"learning_rate": 5.517241379310345e-05, |
|
"loss": 3.0312, |
|
"step": 4 |
|
}, |
|
{ |
|
"epoch": 0.010438413361169102, |
|
"grad_norm": 0.7722844092130382, |
|
"learning_rate": 6.896551724137931e-05, |
|
"loss": 2.0938, |
|
"step": 5 |
|
}, |
|
{ |
|
"epoch": 0.012526096033402923, |
|
"grad_norm": 0.5469423252625637, |
|
"learning_rate": 8.275862068965517e-05, |
|
"loss": 1.5391, |
|
"step": 6 |
|
}, |
|
{ |
|
"epoch": 0.014613778705636743, |
|
"grad_norm": 0.269527877565377, |
|
"learning_rate": 9.655172413793105e-05, |
|
"loss": 1.3281, |
|
"step": 7 |
|
}, |
|
{ |
|
"epoch": 0.016701461377870562, |
|
"grad_norm": 0.2129561647095634, |
|
"learning_rate": 0.0001103448275862069, |
|
"loss": 1.0547, |
|
"step": 8 |
|
}, |
|
{ |
|
"epoch": 0.018789144050104383, |
|
"grad_norm": 0.209174835903919, |
|
"learning_rate": 0.00012413793103448277, |
|
"loss": 0.8672, |
|
"step": 9 |
|
}, |
|
{ |
|
"epoch": 0.020876826722338204, |
|
"grad_norm": 0.16723166688980518, |
|
"learning_rate": 0.00013793103448275863, |
|
"loss": 0.7461, |
|
"step": 10 |
|
}, |
|
{ |
|
"epoch": 0.022964509394572025, |
|
"grad_norm": 0.13523071310549678, |
|
"learning_rate": 0.00015172413793103449, |
|
"loss": 0.75, |
|
"step": 11 |
|
}, |
|
{ |
|
"epoch": 0.025052192066805846, |
|
"grad_norm": 0.13783786374844356, |
|
"learning_rate": 0.00016551724137931035, |
|
"loss": 0.7578, |
|
"step": 12 |
|
}, |
|
{ |
|
"epoch": 0.027139874739039668, |
|
"grad_norm": 0.11159117473743674, |
|
"learning_rate": 0.0001793103448275862, |
|
"loss": 0.6914, |
|
"step": 13 |
|
}, |
|
{ |
|
"epoch": 0.029227557411273485, |
|
"grad_norm": 0.10486826556036052, |
|
"learning_rate": 0.0001931034482758621, |
|
"loss": 0.6484, |
|
"step": 14 |
|
}, |
|
{ |
|
"epoch": 0.031315240083507306, |
|
"grad_norm": 0.1012538060823961, |
|
"learning_rate": 0.00020689655172413795, |
|
"loss": 0.6562, |
|
"step": 15 |
|
}, |
|
{ |
|
"epoch": 0.033402922755741124, |
|
"grad_norm": 0.10310591282634196, |
|
"learning_rate": 0.0002206896551724138, |
|
"loss": 0.6602, |
|
"step": 16 |
|
}, |
|
{ |
|
"epoch": 0.03549060542797495, |
|
"grad_norm": 0.07838991965173675, |
|
"learning_rate": 0.00023448275862068965, |
|
"loss": 0.5469, |
|
"step": 17 |
|
}, |
|
{ |
|
"epoch": 0.037578288100208766, |
|
"grad_norm": 0.07804122247550843, |
|
"learning_rate": 0.00024827586206896553, |
|
"loss": 0.543, |
|
"step": 18 |
|
}, |
|
{ |
|
"epoch": 0.03966597077244259, |
|
"grad_norm": 0.08565464666867292, |
|
"learning_rate": 0.00026206896551724137, |
|
"loss": 0.5625, |
|
"step": 19 |
|
}, |
|
{ |
|
"epoch": 0.04175365344467641, |
|
"grad_norm": 0.07879295016594862, |
|
"learning_rate": 0.00027586206896551725, |
|
"loss": 0.4551, |
|
"step": 20 |
|
}, |
|
{ |
|
"epoch": 0.04384133611691023, |
|
"grad_norm": 0.07549627256843491, |
|
"learning_rate": 0.00028965517241379314, |
|
"loss": 0.4863, |
|
"step": 21 |
|
}, |
|
{ |
|
"epoch": 0.04592901878914405, |
|
"grad_norm": 0.07816061161931821, |
|
"learning_rate": 0.00030344827586206897, |
|
"loss": 0.459, |
|
"step": 22 |
|
}, |
|
{ |
|
"epoch": 0.04801670146137787, |
|
"grad_norm": 0.07796400588496231, |
|
"learning_rate": 0.00031724137931034486, |
|
"loss": 0.5156, |
|
"step": 23 |
|
}, |
|
{ |
|
"epoch": 0.05010438413361169, |
|
"grad_norm": 0.06699113066481945, |
|
"learning_rate": 0.0003310344827586207, |
|
"loss": 0.5508, |
|
"step": 24 |
|
}, |
|
{ |
|
"epoch": 0.05219206680584551, |
|
"grad_norm": 0.07261576114010894, |
|
"learning_rate": 0.0003448275862068965, |
|
"loss": 0.5469, |
|
"step": 25 |
|
}, |
|
{ |
|
"epoch": 0.054279749478079335, |
|
"grad_norm": 0.07464032217874282, |
|
"learning_rate": 0.0003586206896551724, |
|
"loss": 0.4805, |
|
"step": 26 |
|
}, |
|
{ |
|
"epoch": 0.05636743215031315, |
|
"grad_norm": 0.06913377612077472, |
|
"learning_rate": 0.0003724137931034483, |
|
"loss": 0.5117, |
|
"step": 27 |
|
}, |
|
{ |
|
"epoch": 0.05845511482254697, |
|
"grad_norm": 0.061903733751750956, |
|
"learning_rate": 0.0003862068965517242, |
|
"loss": 0.4844, |
|
"step": 28 |
|
}, |
|
{ |
|
"epoch": 0.060542797494780795, |
|
"grad_norm": 0.06908123825325603, |
|
"learning_rate": 0.0004, |
|
"loss": 0.4434, |
|
"step": 29 |
|
}, |
|
{ |
|
"epoch": 0.06263048016701461, |
|
"grad_norm": 0.05444037852658766, |
|
"learning_rate": 0.00039999885641643965, |
|
"loss": 0.4336, |
|
"step": 30 |
|
}, |
|
{ |
|
"epoch": 0.06471816283924843, |
|
"grad_norm": 0.06329218054783699, |
|
"learning_rate": 0.00039999542567883625, |
|
"loss": 0.5078, |
|
"step": 31 |
|
}, |
|
{ |
|
"epoch": 0.06680584551148225, |
|
"grad_norm": 0.05617311135086726, |
|
"learning_rate": 0.0003999897078264233, |
|
"loss": 0.3418, |
|
"step": 32 |
|
}, |
|
{ |
|
"epoch": 0.06889352818371608, |
|
"grad_norm": 0.0631466065976358, |
|
"learning_rate": 0.00039998170292458913, |
|
"loss": 0.4512, |
|
"step": 33 |
|
}, |
|
{ |
|
"epoch": 0.0709812108559499, |
|
"grad_norm": 0.05939776609982831, |
|
"learning_rate": 0.0003999714110648765, |
|
"loss": 0.4727, |
|
"step": 34 |
|
}, |
|
{ |
|
"epoch": 0.07306889352818371, |
|
"grad_norm": 0.07015310870913878, |
|
"learning_rate": 0.0003999588323649815, |
|
"loss": 0.4316, |
|
"step": 35 |
|
}, |
|
{ |
|
"epoch": 0.07515657620041753, |
|
"grad_norm": 0.05541574151493628, |
|
"learning_rate": 0.00039994396696875197, |
|
"loss": 0.3691, |
|
"step": 36 |
|
}, |
|
{ |
|
"epoch": 0.07724425887265135, |
|
"grad_norm": 0.04781737127842468, |
|
"learning_rate": 0.0003999268150461862, |
|
"loss": 0.3711, |
|
"step": 37 |
|
}, |
|
{ |
|
"epoch": 0.07933194154488518, |
|
"grad_norm": 0.052081481582784105, |
|
"learning_rate": 0.00039990737679343073, |
|
"loss": 0.3633, |
|
"step": 38 |
|
}, |
|
{ |
|
"epoch": 0.081419624217119, |
|
"grad_norm": 0.04824424451620144, |
|
"learning_rate": 0.0003998856524327782, |
|
"loss": 0.3867, |
|
"step": 39 |
|
}, |
|
{ |
|
"epoch": 0.08350730688935282, |
|
"grad_norm": 0.048331445388276904, |
|
"learning_rate": 0.0003998616422126649, |
|
"loss": 0.4199, |
|
"step": 40 |
|
}, |
|
{ |
|
"epoch": 0.08559498956158663, |
|
"grad_norm": 0.06796823383551855, |
|
"learning_rate": 0.00039983534640766766, |
|
"loss": 0.4629, |
|
"step": 41 |
|
}, |
|
{ |
|
"epoch": 0.08768267223382047, |
|
"grad_norm": 0.04720979824092969, |
|
"learning_rate": 0.0003998067653185011, |
|
"loss": 0.3594, |
|
"step": 42 |
|
}, |
|
{ |
|
"epoch": 0.08977035490605428, |
|
"grad_norm": 0.05374730136315078, |
|
"learning_rate": 0.00039977589927201373, |
|
"loss": 0.4414, |
|
"step": 43 |
|
}, |
|
{ |
|
"epoch": 0.0918580375782881, |
|
"grad_norm": 0.050894908072710146, |
|
"learning_rate": 0.0003997427486211847, |
|
"loss": 0.4043, |
|
"step": 44 |
|
}, |
|
{ |
|
"epoch": 0.09394572025052192, |
|
"grad_norm": 0.04383506588833311, |
|
"learning_rate": 0.0003997073137451194, |
|
"loss": 0.3281, |
|
"step": 45 |
|
}, |
|
{ |
|
"epoch": 0.09603340292275574, |
|
"grad_norm": 0.055686594769476794, |
|
"learning_rate": 0.00039966959504904515, |
|
"loss": 0.4062, |
|
"step": 46 |
|
}, |
|
{ |
|
"epoch": 0.09812108559498957, |
|
"grad_norm": 0.05244693300757331, |
|
"learning_rate": 0.00039962959296430676, |
|
"loss": 0.3652, |
|
"step": 47 |
|
}, |
|
{ |
|
"epoch": 0.10020876826722339, |
|
"grad_norm": 0.05926741920735249, |
|
"learning_rate": 0.0003995873079483616, |
|
"loss": 0.4844, |
|
"step": 48 |
|
}, |
|
{ |
|
"epoch": 0.1022964509394572, |
|
"grad_norm": 0.04551719284385615, |
|
"learning_rate": 0.00039954274048477405, |
|
"loss": 0.3223, |
|
"step": 49 |
|
}, |
|
{ |
|
"epoch": 0.10438413361169102, |
|
"grad_norm": 0.04385484814869444, |
|
"learning_rate": 0.0003994958910832104, |
|
"loss": 0.4004, |
|
"step": 50 |
|
}, |
|
{ |
|
"epoch": 0.10647181628392484, |
|
"grad_norm": 0.04780597607586747, |
|
"learning_rate": 0.0003994467602794327, |
|
"loss": 0.3125, |
|
"step": 51 |
|
}, |
|
{ |
|
"epoch": 0.10855949895615867, |
|
"grad_norm": 0.08970784709982177, |
|
"learning_rate": 0.00039939534863529265, |
|
"loss": 0.3359, |
|
"step": 52 |
|
}, |
|
{ |
|
"epoch": 0.11064718162839249, |
|
"grad_norm": 0.055961833035488864, |
|
"learning_rate": 0.00039934165673872536, |
|
"loss": 0.4023, |
|
"step": 53 |
|
}, |
|
{ |
|
"epoch": 0.1127348643006263, |
|
"grad_norm": 0.051294633986921585, |
|
"learning_rate": 0.00039928568520374263, |
|
"loss": 0.3848, |
|
"step": 54 |
|
}, |
|
{ |
|
"epoch": 0.11482254697286012, |
|
"grad_norm": 0.05439699876212045, |
|
"learning_rate": 0.0003992274346704257, |
|
"loss": 0.4785, |
|
"step": 55 |
|
}, |
|
{ |
|
"epoch": 0.11691022964509394, |
|
"grad_norm": 0.04990096696869215, |
|
"learning_rate": 0.00039916690580491806, |
|
"loss": 0.377, |
|
"step": 56 |
|
}, |
|
{ |
|
"epoch": 0.11899791231732777, |
|
"grad_norm": 0.04880913842131259, |
|
"learning_rate": 0.00039910409929941787, |
|
"loss": 0.3809, |
|
"step": 57 |
|
}, |
|
{ |
|
"epoch": 0.12108559498956159, |
|
"grad_norm": 0.048110591694635114, |
|
"learning_rate": 0.0003990390158721701, |
|
"loss": 0.375, |
|
"step": 58 |
|
}, |
|
{ |
|
"epoch": 0.12317327766179541, |
|
"grad_norm": 0.05425382131888079, |
|
"learning_rate": 0.0003989716562674579, |
|
"loss": 0.4297, |
|
"step": 59 |
|
}, |
|
{ |
|
"epoch": 0.12526096033402923, |
|
"grad_norm": 0.04140427495861459, |
|
"learning_rate": 0.0003989020212555948, |
|
"loss": 0.3164, |
|
"step": 60 |
|
}, |
|
{ |
|
"epoch": 0.12734864300626306, |
|
"grad_norm": 0.055270560900777276, |
|
"learning_rate": 0.00039883011163291545, |
|
"loss": 0.418, |
|
"step": 61 |
|
}, |
|
{ |
|
"epoch": 0.12943632567849686, |
|
"grad_norm": 0.04546442173482473, |
|
"learning_rate": 0.0003987559282217662, |
|
"loss": 0.3125, |
|
"step": 62 |
|
}, |
|
{ |
|
"epoch": 0.1315240083507307, |
|
"grad_norm": 0.05835477454974449, |
|
"learning_rate": 0.00039867947187049654, |
|
"loss": 0.4199, |
|
"step": 63 |
|
}, |
|
{ |
|
"epoch": 0.1336116910229645, |
|
"grad_norm": 0.06512707863408194, |
|
"learning_rate": 0.00039860074345344875, |
|
"loss": 0.4238, |
|
"step": 64 |
|
}, |
|
{ |
|
"epoch": 0.13569937369519833, |
|
"grad_norm": 0.04896390157005381, |
|
"learning_rate": 0.0003985197438709479, |
|
"loss": 0.2598, |
|
"step": 65 |
|
}, |
|
{ |
|
"epoch": 0.13778705636743216, |
|
"grad_norm": 0.048613271081049014, |
|
"learning_rate": 0.00039843647404929207, |
|
"loss": 0.3633, |
|
"step": 66 |
|
}, |
|
{ |
|
"epoch": 0.13987473903966596, |
|
"grad_norm": 0.050233778334495785, |
|
"learning_rate": 0.0003983509349407412, |
|
"loss": 0.3926, |
|
"step": 67 |
|
}, |
|
{ |
|
"epoch": 0.1419624217118998, |
|
"grad_norm": 0.051442044526522436, |
|
"learning_rate": 0.0003982631275235064, |
|
"loss": 0.4531, |
|
"step": 68 |
|
}, |
|
{ |
|
"epoch": 0.1440501043841336, |
|
"grad_norm": 0.04624056126157752, |
|
"learning_rate": 0.00039817305280173893, |
|
"loss": 0.373, |
|
"step": 69 |
|
}, |
|
{ |
|
"epoch": 0.14613778705636743, |
|
"grad_norm": 0.04350984771291989, |
|
"learning_rate": 0.0003980807118055185, |
|
"loss": 0.377, |
|
"step": 70 |
|
}, |
|
{ |
|
"epoch": 0.14822546972860126, |
|
"grad_norm": 0.04395725697208839, |
|
"learning_rate": 0.0003979861055908415, |
|
"loss": 0.3633, |
|
"step": 71 |
|
}, |
|
{ |
|
"epoch": 0.15031315240083507, |
|
"grad_norm": 0.04706297130636851, |
|
"learning_rate": 0.0003978892352396091, |
|
"loss": 0.4199, |
|
"step": 72 |
|
}, |
|
{ |
|
"epoch": 0.1524008350730689, |
|
"grad_norm": 0.044497773696428565, |
|
"learning_rate": 0.00039779010185961473, |
|
"loss": 0.4707, |
|
"step": 73 |
|
}, |
|
{ |
|
"epoch": 0.1544885177453027, |
|
"grad_norm": 0.0411155822363719, |
|
"learning_rate": 0.0003976887065845314, |
|
"loss": 0.3262, |
|
"step": 74 |
|
}, |
|
{ |
|
"epoch": 0.15657620041753653, |
|
"grad_norm": 0.043812847498547546, |
|
"learning_rate": 0.0003975850505738988, |
|
"loss": 0.3359, |
|
"step": 75 |
|
}, |
|
{ |
|
"epoch": 0.15866388308977036, |
|
"grad_norm": 0.0554940446413714, |
|
"learning_rate": 0.0003974791350131101, |
|
"loss": 0.3867, |
|
"step": 76 |
|
}, |
|
{ |
|
"epoch": 0.16075156576200417, |
|
"grad_norm": 0.04898321080003146, |
|
"learning_rate": 0.00039737096111339804, |
|
"loss": 0.4648, |
|
"step": 77 |
|
}, |
|
{ |
|
"epoch": 0.162839248434238, |
|
"grad_norm": 0.04703455829390756, |
|
"learning_rate": 0.00039726053011182176, |
|
"loss": 0.3945, |
|
"step": 78 |
|
}, |
|
{ |
|
"epoch": 0.1649269311064718, |
|
"grad_norm": 0.048291736323549404, |
|
"learning_rate": 0.00039714784327125196, |
|
"loss": 0.3555, |
|
"step": 79 |
|
}, |
|
{ |
|
"epoch": 0.16701461377870563, |
|
"grad_norm": 0.048158996103635526, |
|
"learning_rate": 0.00039703290188035687, |
|
"loss": 0.3848, |
|
"step": 80 |
|
}, |
|
{ |
|
"epoch": 0.16910229645093947, |
|
"grad_norm": 0.041070665599436396, |
|
"learning_rate": 0.00039691570725358726, |
|
"loss": 0.3672, |
|
"step": 81 |
|
}, |
|
{ |
|
"epoch": 0.17118997912317327, |
|
"grad_norm": 0.04514739709583794, |
|
"learning_rate": 0.0003967962607311617, |
|
"loss": 0.416, |
|
"step": 82 |
|
}, |
|
{ |
|
"epoch": 0.1732776617954071, |
|
"grad_norm": 0.05018154319088866, |
|
"learning_rate": 0.0003966745636790509, |
|
"loss": 0.4219, |
|
"step": 83 |
|
}, |
|
{ |
|
"epoch": 0.17536534446764093, |
|
"grad_norm": 0.04615841699661321, |
|
"learning_rate": 0.00039655061748896243, |
|
"loss": 0.3594, |
|
"step": 84 |
|
}, |
|
{ |
|
"epoch": 0.17745302713987474, |
|
"grad_norm": 0.04770173243618234, |
|
"learning_rate": 0.00039642442357832447, |
|
"loss": 0.3789, |
|
"step": 85 |
|
}, |
|
{ |
|
"epoch": 0.17954070981210857, |
|
"grad_norm": 0.0529492248920703, |
|
"learning_rate": 0.00039629598339026984, |
|
"loss": 0.4102, |
|
"step": 86 |
|
}, |
|
{ |
|
"epoch": 0.18162839248434237, |
|
"grad_norm": 0.04108643805111346, |
|
"learning_rate": 0.00039616529839361946, |
|
"loss": 0.3691, |
|
"step": 87 |
|
}, |
|
{ |
|
"epoch": 0.1837160751565762, |
|
"grad_norm": 0.04230844261748932, |
|
"learning_rate": 0.00039603237008286544, |
|
"loss": 0.3672, |
|
"step": 88 |
|
}, |
|
{ |
|
"epoch": 0.18580375782881003, |
|
"grad_norm": 0.057536412046600406, |
|
"learning_rate": 0.00039589719997815413, |
|
"loss": 0.3574, |
|
"step": 89 |
|
}, |
|
{ |
|
"epoch": 0.18789144050104384, |
|
"grad_norm": 0.03943187889859641, |
|
"learning_rate": 0.0003957597896252685, |
|
"loss": 0.332, |
|
"step": 90 |
|
}, |
|
{ |
|
"epoch": 0.18997912317327767, |
|
"grad_norm": 0.04024451449784526, |
|
"learning_rate": 0.000395620140595611, |
|
"loss": 0.3438, |
|
"step": 91 |
|
}, |
|
{ |
|
"epoch": 0.19206680584551147, |
|
"grad_norm": 0.04428176826646225, |
|
"learning_rate": 0.0003954782544861847, |
|
"loss": 0.3574, |
|
"step": 92 |
|
}, |
|
{ |
|
"epoch": 0.1941544885177453, |
|
"grad_norm": 0.045692937481720604, |
|
"learning_rate": 0.00039533413291957596, |
|
"loss": 0.3574, |
|
"step": 93 |
|
}, |
|
{ |
|
"epoch": 0.19624217118997914, |
|
"grad_norm": 0.04688683747789243, |
|
"learning_rate": 0.00039518777754393535, |
|
"loss": 0.3867, |
|
"step": 94 |
|
}, |
|
{ |
|
"epoch": 0.19832985386221294, |
|
"grad_norm": 0.05536707488449957, |
|
"learning_rate": 0.0003950391900329588, |
|
"loss": 0.3223, |
|
"step": 95 |
|
}, |
|
{ |
|
"epoch": 0.20041753653444677, |
|
"grad_norm": 0.04606231136571311, |
|
"learning_rate": 0.0003948883720858687, |
|
"loss": 0.3535, |
|
"step": 96 |
|
}, |
|
{ |
|
"epoch": 0.20250521920668058, |
|
"grad_norm": 0.04099133317121465, |
|
"learning_rate": 0.00039473532542739435, |
|
"loss": 0.3066, |
|
"step": 97 |
|
}, |
|
{ |
|
"epoch": 0.2045929018789144, |
|
"grad_norm": 0.055252229005040635, |
|
"learning_rate": 0.0003945800518077521, |
|
"loss": 0.3809, |
|
"step": 98 |
|
}, |
|
{ |
|
"epoch": 0.20668058455114824, |
|
"grad_norm": 0.04490759353014913, |
|
"learning_rate": 0.00039442255300262553, |
|
"loss": 0.373, |
|
"step": 99 |
|
}, |
|
{ |
|
"epoch": 0.20876826722338204, |
|
"grad_norm": 0.04889359614788617, |
|
"learning_rate": 0.0003942628308131452, |
|
"loss": 0.377, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.21085594989561587, |
|
"grad_norm": 0.04331101172153666, |
|
"learning_rate": 0.0003941008870658677, |
|
"loss": 0.3926, |
|
"step": 101 |
|
}, |
|
{ |
|
"epoch": 0.21294363256784968, |
|
"grad_norm": 0.039340556954521134, |
|
"learning_rate": 0.00039393672361275506, |
|
"loss": 0.3379, |
|
"step": 102 |
|
}, |
|
{ |
|
"epoch": 0.2150313152400835, |
|
"grad_norm": 0.047692728640419506, |
|
"learning_rate": 0.0003937703423311537, |
|
"loss": 0.375, |
|
"step": 103 |
|
}, |
|
{ |
|
"epoch": 0.21711899791231734, |
|
"grad_norm": 0.04193899617132502, |
|
"learning_rate": 0.00039360174512377246, |
|
"loss": 0.3203, |
|
"step": 104 |
|
}, |
|
{ |
|
"epoch": 0.21920668058455114, |
|
"grad_norm": 0.04450971768791051, |
|
"learning_rate": 0.0003934309339186614, |
|
"loss": 0.3789, |
|
"step": 105 |
|
}, |
|
{ |
|
"epoch": 0.22129436325678498, |
|
"grad_norm": 0.04610401641341007, |
|
"learning_rate": 0.00039325791066918925, |
|
"loss": 0.4004, |
|
"step": 106 |
|
}, |
|
{ |
|
"epoch": 0.22338204592901878, |
|
"grad_norm": 0.0394316515696973, |
|
"learning_rate": 0.00039308267735402156, |
|
"loss": 0.2734, |
|
"step": 107 |
|
}, |
|
{ |
|
"epoch": 0.2254697286012526, |
|
"grad_norm": 0.04552211598084851, |
|
"learning_rate": 0.00039290523597709765, |
|
"loss": 0.3555, |
|
"step": 108 |
|
}, |
|
{ |
|
"epoch": 0.22755741127348644, |
|
"grad_norm": 0.044355074180719295, |
|
"learning_rate": 0.00039272558856760797, |
|
"loss": 0.3418, |
|
"step": 109 |
|
}, |
|
{ |
|
"epoch": 0.22964509394572025, |
|
"grad_norm": 0.04664869616431499, |
|
"learning_rate": 0.0003925437371799708, |
|
"loss": 0.3301, |
|
"step": 110 |
|
}, |
|
{ |
|
"epoch": 0.23173277661795408, |
|
"grad_norm": 0.04421582409863553, |
|
"learning_rate": 0.0003923596838938086, |
|
"loss": 0.3242, |
|
"step": 111 |
|
}, |
|
{ |
|
"epoch": 0.23382045929018788, |
|
"grad_norm": 0.04914905461390972, |
|
"learning_rate": 0.00039217343081392463, |
|
"loss": 0.3027, |
|
"step": 112 |
|
}, |
|
{ |
|
"epoch": 0.2359081419624217, |
|
"grad_norm": 0.04801403896872833, |
|
"learning_rate": 0.00039198498007027836, |
|
"loss": 0.377, |
|
"step": 113 |
|
}, |
|
{ |
|
"epoch": 0.23799582463465555, |
|
"grad_norm": 0.039909379565673524, |
|
"learning_rate": 0.0003917943338179616, |
|
"loss": 0.2949, |
|
"step": 114 |
|
}, |
|
{ |
|
"epoch": 0.24008350730688935, |
|
"grad_norm": 0.04562132710784955, |
|
"learning_rate": 0.00039160149423717353, |
|
"loss": 0.3828, |
|
"step": 115 |
|
}, |
|
{ |
|
"epoch": 0.24217118997912318, |
|
"grad_norm": 0.04218187984586478, |
|
"learning_rate": 0.00039140646353319584, |
|
"loss": 0.3379, |
|
"step": 116 |
|
}, |
|
{ |
|
"epoch": 0.24425887265135698, |
|
"grad_norm": 0.13995497079449562, |
|
"learning_rate": 0.0003912092439363677, |
|
"loss": 0.3887, |
|
"step": 117 |
|
}, |
|
{ |
|
"epoch": 0.24634655532359082, |
|
"grad_norm": 0.03956892044992931, |
|
"learning_rate": 0.00039100983770205995, |
|
"loss": 0.3027, |
|
"step": 118 |
|
}, |
|
{ |
|
"epoch": 0.24843423799582465, |
|
"grad_norm": 0.043965744196170775, |
|
"learning_rate": 0.00039080824711064946, |
|
"loss": 0.3223, |
|
"step": 119 |
|
}, |
|
{ |
|
"epoch": 0.25052192066805845, |
|
"grad_norm": 0.044644460638659476, |
|
"learning_rate": 0.00039060447446749314, |
|
"loss": 0.3496, |
|
"step": 120 |
|
}, |
|
{ |
|
"epoch": 0.25260960334029225, |
|
"grad_norm": 0.05275646603548191, |
|
"learning_rate": 0.00039039852210290147, |
|
"loss": 0.3438, |
|
"step": 121 |
|
}, |
|
{ |
|
"epoch": 0.2546972860125261, |
|
"grad_norm": 0.04064451365674439, |
|
"learning_rate": 0.00039019039237211175, |
|
"loss": 0.3359, |
|
"step": 122 |
|
}, |
|
{ |
|
"epoch": 0.2567849686847599, |
|
"grad_norm": 0.048323141931080164, |
|
"learning_rate": 0.00038998008765526136, |
|
"loss": 0.377, |
|
"step": 123 |
|
}, |
|
{ |
|
"epoch": 0.2588726513569937, |
|
"grad_norm": 0.03725290930480903, |
|
"learning_rate": 0.00038976761035736064, |
|
"loss": 0.2988, |
|
"step": 124 |
|
}, |
|
{ |
|
"epoch": 0.2609603340292276, |
|
"grad_norm": 0.04052181522164268, |
|
"learning_rate": 0.00038955296290826483, |
|
"loss": 0.3203, |
|
"step": 125 |
|
}, |
|
{ |
|
"epoch": 0.2630480167014614, |
|
"grad_norm": 0.04516920645500193, |
|
"learning_rate": 0.00038933614776264695, |
|
"loss": 0.3496, |
|
"step": 126 |
|
}, |
|
{ |
|
"epoch": 0.2651356993736952, |
|
"grad_norm": 0.04887108820007405, |
|
"learning_rate": 0.00038911716739996945, |
|
"loss": 0.2734, |
|
"step": 127 |
|
}, |
|
{ |
|
"epoch": 0.267223382045929, |
|
"grad_norm": 0.04662438292740599, |
|
"learning_rate": 0.0003888960243244556, |
|
"loss": 0.2949, |
|
"step": 128 |
|
}, |
|
{ |
|
"epoch": 0.26931106471816285, |
|
"grad_norm": 0.0462101738931956, |
|
"learning_rate": 0.0003886727210650613, |
|
"loss": 0.3711, |
|
"step": 129 |
|
}, |
|
{ |
|
"epoch": 0.27139874739039666, |
|
"grad_norm": 0.045514397288015096, |
|
"learning_rate": 0.00038844726017544606, |
|
"loss": 0.3945, |
|
"step": 130 |
|
}, |
|
{ |
|
"epoch": 0.27348643006263046, |
|
"grad_norm": 0.04564176989369504, |
|
"learning_rate": 0.00038821964423394335, |
|
"loss": 0.3262, |
|
"step": 131 |
|
}, |
|
{ |
|
"epoch": 0.2755741127348643, |
|
"grad_norm": 0.0480634271209225, |
|
"learning_rate": 0.00038798987584353175, |
|
"loss": 0.4297, |
|
"step": 132 |
|
}, |
|
{ |
|
"epoch": 0.2776617954070981, |
|
"grad_norm": 0.04064367156377542, |
|
"learning_rate": 0.0003877579576318048, |
|
"loss": 0.3066, |
|
"step": 133 |
|
}, |
|
{ |
|
"epoch": 0.2797494780793319, |
|
"grad_norm": 0.06187073759126634, |
|
"learning_rate": 0.000387523892250941, |
|
"loss": 0.3828, |
|
"step": 134 |
|
}, |
|
{ |
|
"epoch": 0.2818371607515658, |
|
"grad_norm": 0.047235774434653544, |
|
"learning_rate": 0.0003872876823776737, |
|
"loss": 0.4551, |
|
"step": 135 |
|
}, |
|
{ |
|
"epoch": 0.2839248434237996, |
|
"grad_norm": 0.04613794323687757, |
|
"learning_rate": 0.00038704933071326004, |
|
"loss": 0.3555, |
|
"step": 136 |
|
}, |
|
{ |
|
"epoch": 0.2860125260960334, |
|
"grad_norm": 0.04123661418763717, |
|
"learning_rate": 0.00038680883998345046, |
|
"loss": 0.3848, |
|
"step": 137 |
|
}, |
|
{ |
|
"epoch": 0.2881002087682672, |
|
"grad_norm": 0.03990534224519547, |
|
"learning_rate": 0.0003865662129384575, |
|
"loss": 0.3223, |
|
"step": 138 |
|
}, |
|
{ |
|
"epoch": 0.29018789144050106, |
|
"grad_norm": 0.03964108882883799, |
|
"learning_rate": 0.00038632145235292407, |
|
"loss": 0.3145, |
|
"step": 139 |
|
}, |
|
{ |
|
"epoch": 0.29227557411273486, |
|
"grad_norm": 0.04771985690919652, |
|
"learning_rate": 0.00038607456102589217, |
|
"loss": 0.3555, |
|
"step": 140 |
|
}, |
|
{ |
|
"epoch": 0.29436325678496866, |
|
"grad_norm": 0.041459959268485676, |
|
"learning_rate": 0.0003858255417807701, |
|
"loss": 0.3203, |
|
"step": 141 |
|
}, |
|
{ |
|
"epoch": 0.2964509394572025, |
|
"grad_norm": 0.04488781487797441, |
|
"learning_rate": 0.0003855743974653014, |
|
"loss": 0.3516, |
|
"step": 142 |
|
}, |
|
{ |
|
"epoch": 0.2985386221294363, |
|
"grad_norm": 0.045595447329589846, |
|
"learning_rate": 0.00038532113095153086, |
|
"loss": 0.4121, |
|
"step": 143 |
|
}, |
|
{ |
|
"epoch": 0.30062630480167013, |
|
"grad_norm": 0.051789411309846184, |
|
"learning_rate": 0.0003850657451357729, |
|
"loss": 0.3945, |
|
"step": 144 |
|
}, |
|
{ |
|
"epoch": 0.302713987473904, |
|
"grad_norm": 0.04655813617947128, |
|
"learning_rate": 0.00038480824293857756, |
|
"loss": 0.377, |
|
"step": 145 |
|
}, |
|
{ |
|
"epoch": 0.3048016701461378, |
|
"grad_norm": 0.07648712725066147, |
|
"learning_rate": 0.00038454862730469773, |
|
"loss": 0.3438, |
|
"step": 146 |
|
}, |
|
{ |
|
"epoch": 0.3068893528183716, |
|
"grad_norm": 0.05310195833383878, |
|
"learning_rate": 0.0003842869012030551, |
|
"loss": 0.3594, |
|
"step": 147 |
|
}, |
|
{ |
|
"epoch": 0.3089770354906054, |
|
"grad_norm": 0.04391887584861711, |
|
"learning_rate": 0.00038402306762670636, |
|
"loss": 0.3555, |
|
"step": 148 |
|
}, |
|
{ |
|
"epoch": 0.31106471816283926, |
|
"grad_norm": 0.038985625260836355, |
|
"learning_rate": 0.000383757129592809, |
|
"loss": 0.2969, |
|
"step": 149 |
|
}, |
|
{ |
|
"epoch": 0.31315240083507306, |
|
"grad_norm": 0.050635143690619895, |
|
"learning_rate": 0.00038348909014258647, |
|
"loss": 0.3398, |
|
"step": 150 |
|
}, |
|
{ |
|
"epoch": 0.31524008350730687, |
|
"grad_norm": 0.04475955558328409, |
|
"learning_rate": 0.00038321895234129394, |
|
"loss": 0.3594, |
|
"step": 151 |
|
}, |
|
{ |
|
"epoch": 0.3173277661795407, |
|
"grad_norm": 0.046705792421880006, |
|
"learning_rate": 0.0003829467192781829, |
|
"loss": 0.3906, |
|
"step": 152 |
|
}, |
|
{ |
|
"epoch": 0.31941544885177453, |
|
"grad_norm": 0.04135295822439203, |
|
"learning_rate": 0.00038267239406646595, |
|
"loss": 0.3203, |
|
"step": 153 |
|
}, |
|
{ |
|
"epoch": 0.32150313152400833, |
|
"grad_norm": 0.03936648551856702, |
|
"learning_rate": 0.000382395979843281, |
|
"loss": 0.3086, |
|
"step": 154 |
|
}, |
|
{ |
|
"epoch": 0.3235908141962422, |
|
"grad_norm": 0.04411342571947294, |
|
"learning_rate": 0.0003821174797696558, |
|
"loss": 0.3145, |
|
"step": 155 |
|
}, |
|
{ |
|
"epoch": 0.325678496868476, |
|
"grad_norm": 0.0440989009123292, |
|
"learning_rate": 0.0003818368970304713, |
|
"loss": 0.3398, |
|
"step": 156 |
|
}, |
|
{ |
|
"epoch": 0.3277661795407098, |
|
"grad_norm": 0.04623084733338215, |
|
"learning_rate": 0.00038155423483442565, |
|
"loss": 0.373, |
|
"step": 157 |
|
}, |
|
{ |
|
"epoch": 0.3298538622129436, |
|
"grad_norm": 0.03851250195995776, |
|
"learning_rate": 0.00038126949641399725, |
|
"loss": 0.2773, |
|
"step": 158 |
|
}, |
|
{ |
|
"epoch": 0.33194154488517746, |
|
"grad_norm": 0.04376944212115029, |
|
"learning_rate": 0.0003809826850254078, |
|
"loss": 0.3613, |
|
"step": 159 |
|
}, |
|
{ |
|
"epoch": 0.33402922755741127, |
|
"grad_norm": 0.044952970087986426, |
|
"learning_rate": 0.00038069380394858526, |
|
"loss": 0.2734, |
|
"step": 160 |
|
}, |
|
{ |
|
"epoch": 0.33611691022964507, |
|
"grad_norm": 0.04338817418637471, |
|
"learning_rate": 0.0003804028564871262, |
|
"loss": 0.334, |
|
"step": 161 |
|
}, |
|
{ |
|
"epoch": 0.33820459290187893, |
|
"grad_norm": 0.04540609013390802, |
|
"learning_rate": 0.00038010984596825774, |
|
"loss": 0.3086, |
|
"step": 162 |
|
}, |
|
{ |
|
"epoch": 0.34029227557411273, |
|
"grad_norm": 0.040675218084428574, |
|
"learning_rate": 0.00037981477574280015, |
|
"loss": 0.3105, |
|
"step": 163 |
|
}, |
|
{ |
|
"epoch": 0.34237995824634654, |
|
"grad_norm": 0.04628740772476471, |
|
"learning_rate": 0.000379517649185128, |
|
"loss": 0.375, |
|
"step": 164 |
|
}, |
|
{ |
|
"epoch": 0.3444676409185804, |
|
"grad_norm": 0.045320126800110626, |
|
"learning_rate": 0.0003792184696931318, |
|
"loss": 0.3926, |
|
"step": 165 |
|
}, |
|
{ |
|
"epoch": 0.3465553235908142, |
|
"grad_norm": 0.047079880323708634, |
|
"learning_rate": 0.00037891724068817896, |
|
"loss": 0.3613, |
|
"step": 166 |
|
}, |
|
{ |
|
"epoch": 0.348643006263048, |
|
"grad_norm": 0.042006341910680284, |
|
"learning_rate": 0.0003786139656150749, |
|
"loss": 0.3848, |
|
"step": 167 |
|
}, |
|
{ |
|
"epoch": 0.35073068893528186, |
|
"grad_norm": 0.042457959635259945, |
|
"learning_rate": 0.0003783086479420235, |
|
"loss": 0.3594, |
|
"step": 168 |
|
}, |
|
{ |
|
"epoch": 0.35281837160751567, |
|
"grad_norm": 0.04729445116359038, |
|
"learning_rate": 0.0003780012911605875, |
|
"loss": 0.4023, |
|
"step": 169 |
|
}, |
|
{ |
|
"epoch": 0.35490605427974947, |
|
"grad_norm": 0.04490557924422057, |
|
"learning_rate": 0.00037769189878564843, |
|
"loss": 0.3184, |
|
"step": 170 |
|
}, |
|
{ |
|
"epoch": 0.3569937369519833, |
|
"grad_norm": 0.04659805592024877, |
|
"learning_rate": 0.0003773804743553667, |
|
"loss": 0.3613, |
|
"step": 171 |
|
}, |
|
{ |
|
"epoch": 0.35908141962421714, |
|
"grad_norm": 0.0457558395135587, |
|
"learning_rate": 0.00037706702143114085, |
|
"loss": 0.3848, |
|
"step": 172 |
|
}, |
|
{ |
|
"epoch": 0.36116910229645094, |
|
"grad_norm": 0.042225345661087835, |
|
"learning_rate": 0.0003767515435975671, |
|
"loss": 0.3652, |
|
"step": 173 |
|
}, |
|
{ |
|
"epoch": 0.36325678496868474, |
|
"grad_norm": 0.03862535581909019, |
|
"learning_rate": 0.00037643404446239805, |
|
"loss": 0.2578, |
|
"step": 174 |
|
}, |
|
{ |
|
"epoch": 0.3653444676409186, |
|
"grad_norm": 0.04511245438490728, |
|
"learning_rate": 0.0003761145276565015, |
|
"loss": 0.3789, |
|
"step": 175 |
|
}, |
|
{ |
|
"epoch": 0.3674321503131524, |
|
"grad_norm": 0.04418007608895449, |
|
"learning_rate": 0.0003757929968338193, |
|
"loss": 0.375, |
|
"step": 176 |
|
}, |
|
{ |
|
"epoch": 0.3695198329853862, |
|
"grad_norm": 0.04131570246472954, |
|
"learning_rate": 0.00037546945567132486, |
|
"loss": 0.3262, |
|
"step": 177 |
|
}, |
|
{ |
|
"epoch": 0.37160751565762007, |
|
"grad_norm": 0.04432885957928793, |
|
"learning_rate": 0.0003751439078689819, |
|
"loss": 0.3105, |
|
"step": 178 |
|
}, |
|
{ |
|
"epoch": 0.3736951983298539, |
|
"grad_norm": 0.04130281228669527, |
|
"learning_rate": 0.0003748163571497015, |
|
"loss": 0.3145, |
|
"step": 179 |
|
}, |
|
{ |
|
"epoch": 0.3757828810020877, |
|
"grad_norm": 0.041235332477716484, |
|
"learning_rate": 0.00037448680725929987, |
|
"loss": 0.3086, |
|
"step": 180 |
|
}, |
|
{ |
|
"epoch": 0.3778705636743215, |
|
"grad_norm": 0.046591187622845885, |
|
"learning_rate": 0.00037415526196645535, |
|
"loss": 0.377, |
|
"step": 181 |
|
}, |
|
{ |
|
"epoch": 0.37995824634655534, |
|
"grad_norm": 0.043131270347982746, |
|
"learning_rate": 0.00037382172506266544, |
|
"loss": 0.3223, |
|
"step": 182 |
|
}, |
|
{ |
|
"epoch": 0.38204592901878914, |
|
"grad_norm": 0.04113286864493325, |
|
"learning_rate": 0.00037348620036220325, |
|
"loss": 0.3379, |
|
"step": 183 |
|
}, |
|
{ |
|
"epoch": 0.38413361169102295, |
|
"grad_norm": 0.045126082835651934, |
|
"learning_rate": 0.00037314869170207427, |
|
"loss": 0.3906, |
|
"step": 184 |
|
}, |
|
{ |
|
"epoch": 0.3862212943632568, |
|
"grad_norm": 0.0468308767634132, |
|
"learning_rate": 0.00037280920294197186, |
|
"loss": 0.418, |
|
"step": 185 |
|
}, |
|
{ |
|
"epoch": 0.3883089770354906, |
|
"grad_norm": 0.041909756539268275, |
|
"learning_rate": 0.0003724677379642338, |
|
"loss": 0.3008, |
|
"step": 186 |
|
}, |
|
{ |
|
"epoch": 0.3903966597077244, |
|
"grad_norm": 0.039249830774930286, |
|
"learning_rate": 0.0003721243006737973, |
|
"loss": 0.293, |
|
"step": 187 |
|
}, |
|
{ |
|
"epoch": 0.3924843423799583, |
|
"grad_norm": 0.043933299859950536, |
|
"learning_rate": 0.0003717788949981549, |
|
"loss": 0.3281, |
|
"step": 188 |
|
}, |
|
{ |
|
"epoch": 0.3945720250521921, |
|
"grad_norm": 0.045457197762689416, |
|
"learning_rate": 0.0003714315248873091, |
|
"loss": 0.416, |
|
"step": 189 |
|
}, |
|
{ |
|
"epoch": 0.3966597077244259, |
|
"grad_norm": 0.038856886092809004, |
|
"learning_rate": 0.00037108219431372734, |
|
"loss": 0.3242, |
|
"step": 190 |
|
}, |
|
{ |
|
"epoch": 0.3987473903966597, |
|
"grad_norm": 0.03887481602609131, |
|
"learning_rate": 0.0003707309072722966, |
|
"loss": 0.3047, |
|
"step": 191 |
|
}, |
|
{ |
|
"epoch": 0.40083507306889354, |
|
"grad_norm": 0.04338847469385484, |
|
"learning_rate": 0.0003703776677802778, |
|
"loss": 0.3945, |
|
"step": 192 |
|
}, |
|
{ |
|
"epoch": 0.40292275574112735, |
|
"grad_norm": 0.04874209932196092, |
|
"learning_rate": 0.0003700224798772597, |
|
"loss": 0.4316, |
|
"step": 193 |
|
}, |
|
{ |
|
"epoch": 0.40501043841336115, |
|
"grad_norm": 0.0409033493154823, |
|
"learning_rate": 0.00036966534762511276, |
|
"loss": 0.3301, |
|
"step": 194 |
|
}, |
|
{ |
|
"epoch": 0.407098121085595, |
|
"grad_norm": 0.040056909386716535, |
|
"learning_rate": 0.0003693062751079427, |
|
"loss": 0.3887, |
|
"step": 195 |
|
}, |
|
{ |
|
"epoch": 0.4091858037578288, |
|
"grad_norm": 0.03764639519212107, |
|
"learning_rate": 0.0003689452664320437, |
|
"loss": 0.3125, |
|
"step": 196 |
|
}, |
|
{ |
|
"epoch": 0.4112734864300626, |
|
"grad_norm": 0.04012224378405559, |
|
"learning_rate": 0.00036858232572585173, |
|
"loss": 0.3574, |
|
"step": 197 |
|
}, |
|
{ |
|
"epoch": 0.4133611691022965, |
|
"grad_norm": 0.04149192047957985, |
|
"learning_rate": 0.00036821745713989706, |
|
"loss": 0.3242, |
|
"step": 198 |
|
}, |
|
{ |
|
"epoch": 0.4154488517745303, |
|
"grad_norm": 0.05017075921477889, |
|
"learning_rate": 0.00036785066484675683, |
|
"loss": 0.3652, |
|
"step": 199 |
|
}, |
|
{ |
|
"epoch": 0.4175365344467641, |
|
"grad_norm": 0.04227941718590804, |
|
"learning_rate": 0.00036748195304100735, |
|
"loss": 0.3242, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.4196242171189979, |
|
"grad_norm": 0.03768968002704521, |
|
"learning_rate": 0.00036711132593917635, |
|
"loss": 0.3105, |
|
"step": 201 |
|
}, |
|
{ |
|
"epoch": 0.42171189979123175, |
|
"grad_norm": 0.04194019808640771, |
|
"learning_rate": 0.00036673878777969426, |
|
"loss": 0.2891, |
|
"step": 202 |
|
}, |
|
{ |
|
"epoch": 0.42379958246346555, |
|
"grad_norm": 0.04542209099244845, |
|
"learning_rate": 0.00036636434282284633, |
|
"loss": 0.3359, |
|
"step": 203 |
|
}, |
|
{ |
|
"epoch": 0.42588726513569936, |
|
"grad_norm": 0.03795072966054866, |
|
"learning_rate": 0.0003659879953507235, |
|
"loss": 0.2178, |
|
"step": 204 |
|
}, |
|
{ |
|
"epoch": 0.4279749478079332, |
|
"grad_norm": 0.0676312462445889, |
|
"learning_rate": 0.0003656097496671737, |
|
"loss": 0.4004, |
|
"step": 205 |
|
}, |
|
{ |
|
"epoch": 0.430062630480167, |
|
"grad_norm": 0.06595187551394353, |
|
"learning_rate": 0.00036522961009775217, |
|
"loss": 0.3359, |
|
"step": 206 |
|
}, |
|
{ |
|
"epoch": 0.4321503131524008, |
|
"grad_norm": 0.048238875202028735, |
|
"learning_rate": 0.0003648475809896727, |
|
"loss": 0.375, |
|
"step": 207 |
|
}, |
|
{ |
|
"epoch": 0.4342379958246347, |
|
"grad_norm": 0.04774287102162698, |
|
"learning_rate": 0.00036446366671175735, |
|
"loss": 0.3613, |
|
"step": 208 |
|
}, |
|
{ |
|
"epoch": 0.4363256784968685, |
|
"grad_norm": 0.04447429752378728, |
|
"learning_rate": 0.0003640778716543866, |
|
"loss": 0.3594, |
|
"step": 209 |
|
}, |
|
{ |
|
"epoch": 0.4384133611691023, |
|
"grad_norm": 0.04069940254145397, |
|
"learning_rate": 0.00036369020022944946, |
|
"loss": 0.3105, |
|
"step": 210 |
|
}, |
|
{ |
|
"epoch": 0.4405010438413361, |
|
"grad_norm": 0.03646181755503713, |
|
"learning_rate": 0.0003633006568702924, |
|
"loss": 0.3047, |
|
"step": 211 |
|
}, |
|
{ |
|
"epoch": 0.44258872651356995, |
|
"grad_norm": 0.041622368664595316, |
|
"learning_rate": 0.00036290924603166945, |
|
"loss": 0.3418, |
|
"step": 212 |
|
}, |
|
{ |
|
"epoch": 0.44467640918580376, |
|
"grad_norm": 0.04687265549200718, |
|
"learning_rate": 0.00036251597218969045, |
|
"loss": 0.3926, |
|
"step": 213 |
|
}, |
|
{ |
|
"epoch": 0.44676409185803756, |
|
"grad_norm": 0.05629699568400707, |
|
"learning_rate": 0.0003621208398417705, |
|
"loss": 0.3848, |
|
"step": 214 |
|
}, |
|
{ |
|
"epoch": 0.4488517745302714, |
|
"grad_norm": 0.04583925885988264, |
|
"learning_rate": 0.0003617238535065782, |
|
"loss": 0.334, |
|
"step": 215 |
|
}, |
|
{ |
|
"epoch": 0.4509394572025052, |
|
"grad_norm": 0.04693126530215214, |
|
"learning_rate": 0.0003613250177239838, |
|
"loss": 0.3691, |
|
"step": 216 |
|
}, |
|
{ |
|
"epoch": 0.453027139874739, |
|
"grad_norm": 0.05342936402981673, |
|
"learning_rate": 0.00036092433705500797, |
|
"loss": 0.3301, |
|
"step": 217 |
|
}, |
|
{ |
|
"epoch": 0.4551148225469729, |
|
"grad_norm": 0.04089550326814837, |
|
"learning_rate": 0.000360521816081769, |
|
"loss": 0.2793, |
|
"step": 218 |
|
}, |
|
{ |
|
"epoch": 0.4572025052192067, |
|
"grad_norm": 0.04725556069406023, |
|
"learning_rate": 0.00036011745940743036, |
|
"loss": 0.3887, |
|
"step": 219 |
|
}, |
|
{ |
|
"epoch": 0.4592901878914405, |
|
"grad_norm": 0.04629637553798221, |
|
"learning_rate": 0.0003597112716561486, |
|
"loss": 0.3477, |
|
"step": 220 |
|
}, |
|
{ |
|
"epoch": 0.4613778705636743, |
|
"grad_norm": 0.042829084723361624, |
|
"learning_rate": 0.0003593032574730201, |
|
"loss": 0.3379, |
|
"step": 221 |
|
}, |
|
{ |
|
"epoch": 0.46346555323590816, |
|
"grad_norm": 0.04290003185482568, |
|
"learning_rate": 0.000358893421524028, |
|
"loss": 0.3828, |
|
"step": 222 |
|
}, |
|
{ |
|
"epoch": 0.46555323590814196, |
|
"grad_norm": 0.03695634767631775, |
|
"learning_rate": 0.0003584817684959888, |
|
"loss": 0.2539, |
|
"step": 223 |
|
}, |
|
{ |
|
"epoch": 0.46764091858037576, |
|
"grad_norm": 0.043249882204696076, |
|
"learning_rate": 0.00035806830309649887, |
|
"loss": 0.291, |
|
"step": 224 |
|
}, |
|
{ |
|
"epoch": 0.4697286012526096, |
|
"grad_norm": 0.0426499927716365, |
|
"learning_rate": 0.0003576530300538805, |
|
"loss": 0.3262, |
|
"step": 225 |
|
}, |
|
{ |
|
"epoch": 0.4718162839248434, |
|
"grad_norm": 0.038269992068262225, |
|
"learning_rate": 0.00035723595411712793, |
|
"loss": 0.2598, |
|
"step": 226 |
|
}, |
|
{ |
|
"epoch": 0.47390396659707723, |
|
"grad_norm": 0.039227521535724244, |
|
"learning_rate": 0.0003568170800558531, |
|
"loss": 0.2812, |
|
"step": 227 |
|
}, |
|
{ |
|
"epoch": 0.4759916492693111, |
|
"grad_norm": 0.0414687007362602, |
|
"learning_rate": 0.0003563964126602308, |
|
"loss": 0.3184, |
|
"step": 228 |
|
}, |
|
{ |
|
"epoch": 0.4780793319415449, |
|
"grad_norm": 0.06563073201350628, |
|
"learning_rate": 0.0003559739567409443, |
|
"loss": 0.2969, |
|
"step": 229 |
|
}, |
|
{ |
|
"epoch": 0.4801670146137787, |
|
"grad_norm": 0.04223079865704053, |
|
"learning_rate": 0.00035554971712913004, |
|
"loss": 0.334, |
|
"step": 230 |
|
}, |
|
{ |
|
"epoch": 0.4822546972860125, |
|
"grad_norm": 0.042708614625140126, |
|
"learning_rate": 0.00035512369867632236, |
|
"loss": 0.3418, |
|
"step": 231 |
|
}, |
|
{ |
|
"epoch": 0.48434237995824636, |
|
"grad_norm": 0.04063103807491919, |
|
"learning_rate": 0.00035469590625439843, |
|
"loss": 0.3477, |
|
"step": 232 |
|
}, |
|
{ |
|
"epoch": 0.48643006263048016, |
|
"grad_norm": 0.045147549230686396, |
|
"learning_rate": 0.00035426634475552187, |
|
"loss": 0.3516, |
|
"step": 233 |
|
}, |
|
{ |
|
"epoch": 0.48851774530271397, |
|
"grad_norm": 0.04106108808167586, |
|
"learning_rate": 0.00035383501909208755, |
|
"loss": 0.2812, |
|
"step": 234 |
|
}, |
|
{ |
|
"epoch": 0.4906054279749478, |
|
"grad_norm": 0.0428728366141268, |
|
"learning_rate": 0.0003534019341966647, |
|
"loss": 0.3398, |
|
"step": 235 |
|
}, |
|
{ |
|
"epoch": 0.49269311064718163, |
|
"grad_norm": 0.047772722106343216, |
|
"learning_rate": 0.00035296709502194115, |
|
"loss": 0.3848, |
|
"step": 236 |
|
}, |
|
{ |
|
"epoch": 0.49478079331941544, |
|
"grad_norm": 0.04629142912114304, |
|
"learning_rate": 0.000352530506540666, |
|
"loss": 0.2715, |
|
"step": 237 |
|
}, |
|
{ |
|
"epoch": 0.4968684759916493, |
|
"grad_norm": 0.05018937450568595, |
|
"learning_rate": 0.0003520921737455935, |
|
"loss": 0.3027, |
|
"step": 238 |
|
}, |
|
{ |
|
"epoch": 0.4989561586638831, |
|
"grad_norm": 0.04431763914166162, |
|
"learning_rate": 0.00035165210164942535, |
|
"loss": 0.3477, |
|
"step": 239 |
|
}, |
|
{ |
|
"epoch": 0.5010438413361169, |
|
"grad_norm": 0.04080914710858152, |
|
"learning_rate": 0.00035121029528475386, |
|
"loss": 0.3086, |
|
"step": 240 |
|
}, |
|
{ |
|
"epoch": 0.5031315240083507, |
|
"grad_norm": 0.04222028074641329, |
|
"learning_rate": 0.0003507667597040038, |
|
"loss": 0.3672, |
|
"step": 241 |
|
}, |
|
{ |
|
"epoch": 0.5052192066805845, |
|
"grad_norm": 0.04367534710869874, |
|
"learning_rate": 0.0003503214999793752, |
|
"loss": 0.3223, |
|
"step": 242 |
|
}, |
|
{ |
|
"epoch": 0.5073068893528184, |
|
"grad_norm": 0.04813953280858226, |
|
"learning_rate": 0.00034987452120278505, |
|
"loss": 0.3984, |
|
"step": 243 |
|
}, |
|
{ |
|
"epoch": 0.5093945720250522, |
|
"grad_norm": 0.046413128010797784, |
|
"learning_rate": 0.00034942582848580925, |
|
"loss": 0.3848, |
|
"step": 244 |
|
}, |
|
{ |
|
"epoch": 0.511482254697286, |
|
"grad_norm": 0.036986531982911595, |
|
"learning_rate": 0.0003489754269596239, |
|
"loss": 0.3066, |
|
"step": 245 |
|
}, |
|
{ |
|
"epoch": 0.5135699373695198, |
|
"grad_norm": 0.040001114155947204, |
|
"learning_rate": 0.00034852332177494685, |
|
"loss": 0.2852, |
|
"step": 246 |
|
}, |
|
{ |
|
"epoch": 0.5156576200417536, |
|
"grad_norm": 0.04605884877382518, |
|
"learning_rate": 0.0003480695181019786, |
|
"loss": 0.3535, |
|
"step": 247 |
|
}, |
|
{ |
|
"epoch": 0.5177453027139874, |
|
"grad_norm": 0.0406658437248984, |
|
"learning_rate": 0.0003476140211303433, |
|
"loss": 0.334, |
|
"step": 248 |
|
}, |
|
{ |
|
"epoch": 0.5198329853862212, |
|
"grad_norm": 0.03909067687413047, |
|
"learning_rate": 0.0003471568360690296, |
|
"loss": 0.2871, |
|
"step": 249 |
|
}, |
|
{ |
|
"epoch": 0.5219206680584552, |
|
"grad_norm": 0.04831389973643467, |
|
"learning_rate": 0.00034669796814633055, |
|
"loss": 0.3398, |
|
"step": 250 |
|
}, |
|
{ |
|
"epoch": 0.524008350730689, |
|
"grad_norm": 0.04408464727595837, |
|
"learning_rate": 0.0003462374226097844, |
|
"loss": 0.3242, |
|
"step": 251 |
|
}, |
|
{ |
|
"epoch": 0.5260960334029228, |
|
"grad_norm": 0.03798160874377182, |
|
"learning_rate": 0.00034577520472611414, |
|
"loss": 0.252, |
|
"step": 252 |
|
}, |
|
{ |
|
"epoch": 0.5281837160751566, |
|
"grad_norm": 0.04625762493599334, |
|
"learning_rate": 0.0003453113197811675, |
|
"loss": 0.334, |
|
"step": 253 |
|
}, |
|
{ |
|
"epoch": 0.5302713987473904, |
|
"grad_norm": 0.03757333278335248, |
|
"learning_rate": 0.0003448457730798564, |
|
"loss": 0.2734, |
|
"step": 254 |
|
}, |
|
{ |
|
"epoch": 0.5323590814196242, |
|
"grad_norm": 0.038808091023111425, |
|
"learning_rate": 0.0003443785699460965, |
|
"loss": 0.3652, |
|
"step": 255 |
|
}, |
|
{ |
|
"epoch": 0.534446764091858, |
|
"grad_norm": 0.03797009162534055, |
|
"learning_rate": 0.0003439097157227459, |
|
"loss": 0.3223, |
|
"step": 256 |
|
}, |
|
{ |
|
"epoch": 0.5365344467640919, |
|
"grad_norm": 0.038601658815291935, |
|
"learning_rate": 0.0003434392157715445, |
|
"loss": 0.373, |
|
"step": 257 |
|
}, |
|
{ |
|
"epoch": 0.5386221294363257, |
|
"grad_norm": 0.04680506137895324, |
|
"learning_rate": 0.00034296707547305235, |
|
"loss": 0.334, |
|
"step": 258 |
|
}, |
|
{ |
|
"epoch": 0.5407098121085595, |
|
"grad_norm": 0.041499005757660336, |
|
"learning_rate": 0.0003424933002265883, |
|
"loss": 0.3457, |
|
"step": 259 |
|
}, |
|
{ |
|
"epoch": 0.5427974947807933, |
|
"grad_norm": 0.04115196540841461, |
|
"learning_rate": 0.00034201789545016824, |
|
"loss": 0.3242, |
|
"step": 260 |
|
}, |
|
{ |
|
"epoch": 0.5448851774530271, |
|
"grad_norm": 0.04340048389863411, |
|
"learning_rate": 0.00034154086658044297, |
|
"loss": 0.2949, |
|
"step": 261 |
|
}, |
|
{ |
|
"epoch": 0.5469728601252609, |
|
"grad_norm": 0.04378711119505926, |
|
"learning_rate": 0.0003410622190726362, |
|
"loss": 0.3066, |
|
"step": 262 |
|
}, |
|
{ |
|
"epoch": 0.5490605427974948, |
|
"grad_norm": 0.04643412727247773, |
|
"learning_rate": 0.00034058195840048225, |
|
"loss": 0.3652, |
|
"step": 263 |
|
}, |
|
{ |
|
"epoch": 0.5511482254697286, |
|
"grad_norm": 0.042043153196747375, |
|
"learning_rate": 0.0003401000900561631, |
|
"loss": 0.3789, |
|
"step": 264 |
|
}, |
|
{ |
|
"epoch": 0.5532359081419624, |
|
"grad_norm": 0.040040695052475275, |
|
"learning_rate": 0.00033961661955024595, |
|
"loss": 0.3145, |
|
"step": 265 |
|
}, |
|
{ |
|
"epoch": 0.5553235908141962, |
|
"grad_norm": 0.038423365462145156, |
|
"learning_rate": 0.0003391315524116202, |
|
"loss": 0.3223, |
|
"step": 266 |
|
}, |
|
{ |
|
"epoch": 0.55741127348643, |
|
"grad_norm": 0.038539884775496434, |
|
"learning_rate": 0.00033864489418743355, |
|
"loss": 0.3086, |
|
"step": 267 |
|
}, |
|
{ |
|
"epoch": 0.5594989561586639, |
|
"grad_norm": 0.04422443806977669, |
|
"learning_rate": 0.0003381566504430298, |
|
"loss": 0.3945, |
|
"step": 268 |
|
}, |
|
{ |
|
"epoch": 0.5615866388308977, |
|
"grad_norm": 0.0377661934221125, |
|
"learning_rate": 0.00033766682676188386, |
|
"loss": 0.3398, |
|
"step": 269 |
|
}, |
|
{ |
|
"epoch": 0.5636743215031316, |
|
"grad_norm": 0.036534819703584714, |
|
"learning_rate": 0.00033717542874553896, |
|
"loss": 0.3418, |
|
"step": 270 |
|
}, |
|
{ |
|
"epoch": 0.5657620041753654, |
|
"grad_norm": 0.04440346943789282, |
|
"learning_rate": 0.0003366824620135421, |
|
"loss": 0.3984, |
|
"step": 271 |
|
}, |
|
{ |
|
"epoch": 0.5678496868475992, |
|
"grad_norm": 0.037034787962705526, |
|
"learning_rate": 0.00033618793220337965, |
|
"loss": 0.2812, |
|
"step": 272 |
|
}, |
|
{ |
|
"epoch": 0.569937369519833, |
|
"grad_norm": 0.03927620752143223, |
|
"learning_rate": 0.0003356918449704133, |
|
"loss": 0.2695, |
|
"step": 273 |
|
}, |
|
{ |
|
"epoch": 0.5720250521920668, |
|
"grad_norm": 0.04369797108825467, |
|
"learning_rate": 0.00033519420598781505, |
|
"loss": 0.3477, |
|
"step": 274 |
|
}, |
|
{ |
|
"epoch": 0.5741127348643006, |
|
"grad_norm": 0.0379874901088521, |
|
"learning_rate": 0.0003346950209465025, |
|
"loss": 0.248, |
|
"step": 275 |
|
}, |
|
{ |
|
"epoch": 0.5762004175365344, |
|
"grad_norm": 0.042755557159503056, |
|
"learning_rate": 0.0003341942955550738, |
|
"loss": 0.3242, |
|
"step": 276 |
|
}, |
|
{ |
|
"epoch": 0.5782881002087683, |
|
"grad_norm": 0.039892604797522456, |
|
"learning_rate": 0.0003336920355397421, |
|
"loss": 0.3125, |
|
"step": 277 |
|
}, |
|
{ |
|
"epoch": 0.5803757828810021, |
|
"grad_norm": 0.037020438684393274, |
|
"learning_rate": 0.0003331882466442704, |
|
"loss": 0.2832, |
|
"step": 278 |
|
}, |
|
{ |
|
"epoch": 0.5824634655532359, |
|
"grad_norm": 0.0448544789689836, |
|
"learning_rate": 0.0003326829346299057, |
|
"loss": 0.3613, |
|
"step": 279 |
|
}, |
|
{ |
|
"epoch": 0.5845511482254697, |
|
"grad_norm": 0.037923744133296695, |
|
"learning_rate": 0.00033217610527531315, |
|
"loss": 0.3262, |
|
"step": 280 |
|
}, |
|
{ |
|
"epoch": 0.5866388308977035, |
|
"grad_norm": 0.039536585438320955, |
|
"learning_rate": 0.00033166776437650985, |
|
"loss": 0.3301, |
|
"step": 281 |
|
}, |
|
{ |
|
"epoch": 0.5887265135699373, |
|
"grad_norm": 0.042309874793804436, |
|
"learning_rate": 0.0003311579177467988, |
|
"loss": 0.3281, |
|
"step": 282 |
|
}, |
|
{ |
|
"epoch": 0.5908141962421712, |
|
"grad_norm": 0.04247524788442754, |
|
"learning_rate": 0.0003306465712167023, |
|
"loss": 0.3281, |
|
"step": 283 |
|
}, |
|
{ |
|
"epoch": 0.592901878914405, |
|
"grad_norm": 0.038298773282535274, |
|
"learning_rate": 0.0003301337306338951, |
|
"loss": 0.3066, |
|
"step": 284 |
|
}, |
|
{ |
|
"epoch": 0.5949895615866388, |
|
"grad_norm": 0.04845680190635201, |
|
"learning_rate": 0.00032961940186313784, |
|
"loss": 0.3867, |
|
"step": 285 |
|
}, |
|
{ |
|
"epoch": 0.5970772442588727, |
|
"grad_norm": 0.04425999769437242, |
|
"learning_rate": 0.00032910359078620974, |
|
"loss": 0.3848, |
|
"step": 286 |
|
}, |
|
{ |
|
"epoch": 0.5991649269311065, |
|
"grad_norm": 0.04350677333004508, |
|
"learning_rate": 0.0003285863033018416, |
|
"loss": 0.3359, |
|
"step": 287 |
|
}, |
|
{ |
|
"epoch": 0.6012526096033403, |
|
"grad_norm": 0.04313248762187342, |
|
"learning_rate": 0.00032806754532564795, |
|
"loss": 0.3281, |
|
"step": 288 |
|
}, |
|
{ |
|
"epoch": 0.6033402922755741, |
|
"grad_norm": 0.04172934704529679, |
|
"learning_rate": 0.00032754732279005974, |
|
"loss": 0.3887, |
|
"step": 289 |
|
}, |
|
{ |
|
"epoch": 0.605427974947808, |
|
"grad_norm": 0.053530546182614214, |
|
"learning_rate": 0.0003270256416442564, |
|
"loss": 0.4512, |
|
"step": 290 |
|
}, |
|
{ |
|
"epoch": 0.6075156576200418, |
|
"grad_norm": 0.0585588589735363, |
|
"learning_rate": 0.0003265025078540977, |
|
"loss": 0.4023, |
|
"step": 291 |
|
}, |
|
{ |
|
"epoch": 0.6096033402922756, |
|
"grad_norm": 0.046794215765212366, |
|
"learning_rate": 0.00032597792740205573, |
|
"loss": 0.2871, |
|
"step": 292 |
|
}, |
|
{ |
|
"epoch": 0.6116910229645094, |
|
"grad_norm": 0.04076085980117636, |
|
"learning_rate": 0.0003254519062871462, |
|
"loss": 0.3027, |
|
"step": 293 |
|
}, |
|
{ |
|
"epoch": 0.6137787056367432, |
|
"grad_norm": 0.039930806408977594, |
|
"learning_rate": 0.0003249244505248603, |
|
"loss": 0.3086, |
|
"step": 294 |
|
}, |
|
{ |
|
"epoch": 0.615866388308977, |
|
"grad_norm": 0.047339086645206876, |
|
"learning_rate": 0.0003243955661470952, |
|
"loss": 0.373, |
|
"step": 295 |
|
}, |
|
{ |
|
"epoch": 0.6179540709812108, |
|
"grad_norm": 0.041035006122051215, |
|
"learning_rate": 0.00032386525920208574, |
|
"loss": 0.3281, |
|
"step": 296 |
|
}, |
|
{ |
|
"epoch": 0.6200417536534447, |
|
"grad_norm": 0.05042533676976384, |
|
"learning_rate": 0.000323333535754335, |
|
"loss": 0.293, |
|
"step": 297 |
|
}, |
|
{ |
|
"epoch": 0.6221294363256785, |
|
"grad_norm": 0.043959165085212505, |
|
"learning_rate": 0.00032280040188454494, |
|
"loss": 0.3125, |
|
"step": 298 |
|
}, |
|
{ |
|
"epoch": 0.6242171189979123, |
|
"grad_norm": 0.04458635918949398, |
|
"learning_rate": 0.0003222658636895469, |
|
"loss": 0.3105, |
|
"step": 299 |
|
}, |
|
{ |
|
"epoch": 0.6263048016701461, |
|
"grad_norm": 0.04146656432155535, |
|
"learning_rate": 0.0003217299272822316, |
|
"loss": 0.3672, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.6283924843423799, |
|
"grad_norm": 0.04285146681080291, |
|
"learning_rate": 0.00032119259879147987, |
|
"loss": 0.2773, |
|
"step": 301 |
|
}, |
|
{ |
|
"epoch": 0.6304801670146137, |
|
"grad_norm": 0.04012756626164825, |
|
"learning_rate": 0.0003206538843620919, |
|
"loss": 0.3145, |
|
"step": 302 |
|
}, |
|
{ |
|
"epoch": 0.6325678496868476, |
|
"grad_norm": 0.03395547556102837, |
|
"learning_rate": 0.0003201137901547175, |
|
"loss": 0.2344, |
|
"step": 303 |
|
}, |
|
{ |
|
"epoch": 0.6346555323590815, |
|
"grad_norm": 0.038474167468473125, |
|
"learning_rate": 0.00031957232234578506, |
|
"loss": 0.3105, |
|
"step": 304 |
|
}, |
|
{ |
|
"epoch": 0.6367432150313153, |
|
"grad_norm": 0.04345850995580467, |
|
"learning_rate": 0.00031902948712743146, |
|
"loss": 0.3711, |
|
"step": 305 |
|
}, |
|
{ |
|
"epoch": 0.6388308977035491, |
|
"grad_norm": 0.04561991595493648, |
|
"learning_rate": 0.00031848529070743113, |
|
"loss": 0.3281, |
|
"step": 306 |
|
}, |
|
{ |
|
"epoch": 0.6409185803757829, |
|
"grad_norm": 0.043639261730465534, |
|
"learning_rate": 0.00031793973930912465, |
|
"loss": 0.3535, |
|
"step": 307 |
|
}, |
|
{ |
|
"epoch": 0.6430062630480167, |
|
"grad_norm": 0.041614849807266444, |
|
"learning_rate": 0.00031739283917134835, |
|
"loss": 0.293, |
|
"step": 308 |
|
}, |
|
{ |
|
"epoch": 0.6450939457202505, |
|
"grad_norm": 0.037829924843835225, |
|
"learning_rate": 0.00031684459654836224, |
|
"loss": 0.2793, |
|
"step": 309 |
|
}, |
|
{ |
|
"epoch": 0.6471816283924844, |
|
"grad_norm": 0.043593747485495736, |
|
"learning_rate": 0.00031629501770977865, |
|
"loss": 0.2871, |
|
"step": 310 |
|
}, |
|
{ |
|
"epoch": 0.6492693110647182, |
|
"grad_norm": 0.043504619259611736, |
|
"learning_rate": 0.0003157441089404911, |
|
"loss": 0.377, |
|
"step": 311 |
|
}, |
|
{ |
|
"epoch": 0.651356993736952, |
|
"grad_norm": 0.042969570910675, |
|
"learning_rate": 0.00031519187654060153, |
|
"loss": 0.3535, |
|
"step": 312 |
|
}, |
|
{ |
|
"epoch": 0.6534446764091858, |
|
"grad_norm": 0.04444146348618118, |
|
"learning_rate": 0.0003146383268253489, |
|
"loss": 0.3633, |
|
"step": 313 |
|
}, |
|
{ |
|
"epoch": 0.6555323590814196, |
|
"grad_norm": 0.0579233541293356, |
|
"learning_rate": 0.00031408346612503684, |
|
"loss": 0.2852, |
|
"step": 314 |
|
}, |
|
{ |
|
"epoch": 0.6576200417536534, |
|
"grad_norm": 0.04116612592402649, |
|
"learning_rate": 0.00031352730078496103, |
|
"loss": 0.3184, |
|
"step": 315 |
|
}, |
|
{ |
|
"epoch": 0.6597077244258872, |
|
"grad_norm": 0.042619485954835536, |
|
"learning_rate": 0.0003129698371653368, |
|
"loss": 0.3555, |
|
"step": 316 |
|
}, |
|
{ |
|
"epoch": 0.6617954070981211, |
|
"grad_norm": 0.03978568537420636, |
|
"learning_rate": 0.00031241108164122667, |
|
"loss": 0.3184, |
|
"step": 317 |
|
}, |
|
{ |
|
"epoch": 0.6638830897703549, |
|
"grad_norm": 0.04398595731711137, |
|
"learning_rate": 0.0003118510406024667, |
|
"loss": 0.3262, |
|
"step": 318 |
|
}, |
|
{ |
|
"epoch": 0.6659707724425887, |
|
"grad_norm": 0.03557319884714253, |
|
"learning_rate": 0.0003112897204535943, |
|
"loss": 0.2539, |
|
"step": 319 |
|
}, |
|
{ |
|
"epoch": 0.6680584551148225, |
|
"grad_norm": 0.04098653755334443, |
|
"learning_rate": 0.0003107271276137745, |
|
"loss": 0.3711, |
|
"step": 320 |
|
}, |
|
{ |
|
"epoch": 0.6701461377870563, |
|
"grad_norm": 0.0377819264468114, |
|
"learning_rate": 0.0003101632685167263, |
|
"loss": 0.2891, |
|
"step": 321 |
|
}, |
|
{ |
|
"epoch": 0.6722338204592901, |
|
"grad_norm": 0.038417025071128603, |
|
"learning_rate": 0.00030959814961064973, |
|
"loss": 0.2891, |
|
"step": 322 |
|
}, |
|
{ |
|
"epoch": 0.6743215031315241, |
|
"grad_norm": 0.038602243878265344, |
|
"learning_rate": 0.0003090317773581518, |
|
"loss": 0.2734, |
|
"step": 323 |
|
}, |
|
{ |
|
"epoch": 0.6764091858037579, |
|
"grad_norm": 0.0487821791964808, |
|
"learning_rate": 0.0003084641582361723, |
|
"loss": 0.3574, |
|
"step": 324 |
|
}, |
|
{ |
|
"epoch": 0.6784968684759917, |
|
"grad_norm": 0.04544872266451946, |
|
"learning_rate": 0.00030789529873591034, |
|
"loss": 0.3008, |
|
"step": 325 |
|
}, |
|
{ |
|
"epoch": 0.6805845511482255, |
|
"grad_norm": 0.0373160417363858, |
|
"learning_rate": 0.0003073252053627496, |
|
"loss": 0.2871, |
|
"step": 326 |
|
}, |
|
{ |
|
"epoch": 0.6826722338204593, |
|
"grad_norm": 0.04358868533626725, |
|
"learning_rate": 0.0003067538846361841, |
|
"loss": 0.3066, |
|
"step": 327 |
|
}, |
|
{ |
|
"epoch": 0.6847599164926931, |
|
"grad_norm": 0.040357136071415334, |
|
"learning_rate": 0.0003061813430897439, |
|
"loss": 0.3516, |
|
"step": 328 |
|
}, |
|
{ |
|
"epoch": 0.6868475991649269, |
|
"grad_norm": 0.04862770193509928, |
|
"learning_rate": 0.0003056075872709199, |
|
"loss": 0.4062, |
|
"step": 329 |
|
}, |
|
{ |
|
"epoch": 0.6889352818371608, |
|
"grad_norm": 0.04688881879011172, |
|
"learning_rate": 0.00030503262374108927, |
|
"loss": 0.4434, |
|
"step": 330 |
|
}, |
|
{ |
|
"epoch": 0.6910229645093946, |
|
"grad_norm": 0.04276287817278344, |
|
"learning_rate": 0.00030445645907544057, |
|
"loss": 0.3105, |
|
"step": 331 |
|
}, |
|
{ |
|
"epoch": 0.6931106471816284, |
|
"grad_norm": 0.03644902974160357, |
|
"learning_rate": 0.0003038790998628981, |
|
"loss": 0.249, |
|
"step": 332 |
|
}, |
|
{ |
|
"epoch": 0.6951983298538622, |
|
"grad_norm": 0.03602308126572102, |
|
"learning_rate": 0.0003033005527060469, |
|
"loss": 0.2891, |
|
"step": 333 |
|
}, |
|
{ |
|
"epoch": 0.697286012526096, |
|
"grad_norm": 0.0397502232733115, |
|
"learning_rate": 0.0003027208242210571, |
|
"loss": 0.334, |
|
"step": 334 |
|
}, |
|
{ |
|
"epoch": 0.6993736951983298, |
|
"grad_norm": 0.042614553937913115, |
|
"learning_rate": 0.0003021399210376084, |
|
"loss": 0.3281, |
|
"step": 335 |
|
}, |
|
{ |
|
"epoch": 0.7014613778705637, |
|
"grad_norm": 0.040360864280806166, |
|
"learning_rate": 0.0003015578497988142, |
|
"loss": 0.3145, |
|
"step": 336 |
|
}, |
|
{ |
|
"epoch": 0.7035490605427975, |
|
"grad_norm": 0.03697519450333998, |
|
"learning_rate": 0.0003009746171611454, |
|
"loss": 0.2852, |
|
"step": 337 |
|
}, |
|
{ |
|
"epoch": 0.7056367432150313, |
|
"grad_norm": 0.03920168369252357, |
|
"learning_rate": 0.0003003902297943545, |
|
"loss": 0.2578, |
|
"step": 338 |
|
}, |
|
{ |
|
"epoch": 0.7077244258872651, |
|
"grad_norm": 0.04596701530986532, |
|
"learning_rate": 0.00029980469438139947, |
|
"loss": 0.3066, |
|
"step": 339 |
|
}, |
|
{ |
|
"epoch": 0.7098121085594989, |
|
"grad_norm": 0.03590053569859233, |
|
"learning_rate": 0.00029921801761836696, |
|
"loss": 0.2598, |
|
"step": 340 |
|
}, |
|
{ |
|
"epoch": 0.7118997912317327, |
|
"grad_norm": 0.039137959610818826, |
|
"learning_rate": 0.0002986302062143961, |
|
"loss": 0.3398, |
|
"step": 341 |
|
}, |
|
{ |
|
"epoch": 0.7139874739039666, |
|
"grad_norm": 0.04445491792602957, |
|
"learning_rate": 0.0002980412668916013, |
|
"loss": 0.3398, |
|
"step": 342 |
|
}, |
|
{ |
|
"epoch": 0.7160751565762005, |
|
"grad_norm": 0.040340445047287156, |
|
"learning_rate": 0.00029745120638499596, |
|
"loss": 0.3516, |
|
"step": 343 |
|
}, |
|
{ |
|
"epoch": 0.7181628392484343, |
|
"grad_norm": 0.04520397395484933, |
|
"learning_rate": 0.00029686003144241494, |
|
"loss": 0.3555, |
|
"step": 344 |
|
}, |
|
{ |
|
"epoch": 0.7202505219206681, |
|
"grad_norm": 0.045695108134425295, |
|
"learning_rate": 0.0002962677488244378, |
|
"loss": 0.2871, |
|
"step": 345 |
|
}, |
|
{ |
|
"epoch": 0.7223382045929019, |
|
"grad_norm": 0.0414553610811136, |
|
"learning_rate": 0.0002956743653043111, |
|
"loss": 0.3828, |
|
"step": 346 |
|
}, |
|
{ |
|
"epoch": 0.7244258872651357, |
|
"grad_norm": 0.043267589906942225, |
|
"learning_rate": 0.00029507988766787135, |
|
"loss": 0.3262, |
|
"step": 347 |
|
}, |
|
{ |
|
"epoch": 0.7265135699373695, |
|
"grad_norm": 0.03665181460868117, |
|
"learning_rate": 0.0002944843227134669, |
|
"loss": 0.3301, |
|
"step": 348 |
|
}, |
|
{ |
|
"epoch": 0.7286012526096033, |
|
"grad_norm": 0.04220684212991339, |
|
"learning_rate": 0.00029388767725188077, |
|
"loss": 0.3223, |
|
"step": 349 |
|
}, |
|
{ |
|
"epoch": 0.7306889352818372, |
|
"grad_norm": 0.04468007941423499, |
|
"learning_rate": 0.0002932899581062524, |
|
"loss": 0.2949, |
|
"step": 350 |
|
}, |
|
{ |
|
"epoch": 0.732776617954071, |
|
"grad_norm": 0.044433764048658723, |
|
"learning_rate": 0.00029269117211199955, |
|
"loss": 0.377, |
|
"step": 351 |
|
}, |
|
{ |
|
"epoch": 0.7348643006263048, |
|
"grad_norm": 0.046209035174827495, |
|
"learning_rate": 0.0002920913261167405, |
|
"loss": 0.3691, |
|
"step": 352 |
|
}, |
|
{ |
|
"epoch": 0.7369519832985386, |
|
"grad_norm": 0.040034133910176115, |
|
"learning_rate": 0.00029149042698021544, |
|
"loss": 0.3105, |
|
"step": 353 |
|
}, |
|
{ |
|
"epoch": 0.7390396659707724, |
|
"grad_norm": 0.04609016944022695, |
|
"learning_rate": 0.00029088848157420813, |
|
"loss": 0.3984, |
|
"step": 354 |
|
}, |
|
{ |
|
"epoch": 0.7411273486430062, |
|
"grad_norm": 0.04480775819697267, |
|
"learning_rate": 0.00029028549678246715, |
|
"loss": 0.332, |
|
"step": 355 |
|
}, |
|
{ |
|
"epoch": 0.7432150313152401, |
|
"grad_norm": 0.04387465996335384, |
|
"learning_rate": 0.0002896814795006277, |
|
"loss": 0.3984, |
|
"step": 356 |
|
}, |
|
{ |
|
"epoch": 0.7453027139874739, |
|
"grad_norm": 0.041864619992452806, |
|
"learning_rate": 0.0002890764366361318, |
|
"loss": 0.3633, |
|
"step": 357 |
|
}, |
|
{ |
|
"epoch": 0.7473903966597077, |
|
"grad_norm": 0.03717012201708029, |
|
"learning_rate": 0.0002884703751081504, |
|
"loss": 0.2324, |
|
"step": 358 |
|
}, |
|
{ |
|
"epoch": 0.7494780793319415, |
|
"grad_norm": 0.043251857860819434, |
|
"learning_rate": 0.0002878633018475035, |
|
"loss": 0.3066, |
|
"step": 359 |
|
}, |
|
{ |
|
"epoch": 0.7515657620041754, |
|
"grad_norm": 0.03591327407860244, |
|
"learning_rate": 0.00028725522379658097, |
|
"loss": 0.2852, |
|
"step": 360 |
|
}, |
|
{ |
|
"epoch": 0.7536534446764092, |
|
"grad_norm": 0.0381908896371197, |
|
"learning_rate": 0.00028664614790926355, |
|
"loss": 0.3477, |
|
"step": 361 |
|
}, |
|
{ |
|
"epoch": 0.755741127348643, |
|
"grad_norm": 0.04392871145759733, |
|
"learning_rate": 0.000286036081150843, |
|
"loss": 0.4414, |
|
"step": 362 |
|
}, |
|
{ |
|
"epoch": 0.7578288100208769, |
|
"grad_norm": 0.04392250979416697, |
|
"learning_rate": 0.0002854250304979423, |
|
"loss": 0.2832, |
|
"step": 363 |
|
}, |
|
{ |
|
"epoch": 0.7599164926931107, |
|
"grad_norm": 0.03445036859255453, |
|
"learning_rate": 0.0002848130029384363, |
|
"loss": 0.3027, |
|
"step": 364 |
|
}, |
|
{ |
|
"epoch": 0.7620041753653445, |
|
"grad_norm": 0.03588315051937214, |
|
"learning_rate": 0.00028420000547137175, |
|
"loss": 0.2793, |
|
"step": 365 |
|
}, |
|
{ |
|
"epoch": 0.7640918580375783, |
|
"grad_norm": 0.043623631697390225, |
|
"learning_rate": 0.00028358604510688666, |
|
"loss": 0.3301, |
|
"step": 366 |
|
}, |
|
{ |
|
"epoch": 0.7661795407098121, |
|
"grad_norm": 0.044899266192686, |
|
"learning_rate": 0.000282971128866131, |
|
"loss": 0.2988, |
|
"step": 367 |
|
}, |
|
{ |
|
"epoch": 0.7682672233820459, |
|
"grad_norm": 0.044441044360934975, |
|
"learning_rate": 0.0002823552637811857, |
|
"loss": 0.3262, |
|
"step": 368 |
|
}, |
|
{ |
|
"epoch": 0.7703549060542797, |
|
"grad_norm": 0.04617336800060674, |
|
"learning_rate": 0.00028173845689498286, |
|
"loss": 0.2812, |
|
"step": 369 |
|
}, |
|
{ |
|
"epoch": 0.7724425887265136, |
|
"grad_norm": 0.03481084134760086, |
|
"learning_rate": 0.0002811207152612243, |
|
"loss": 0.2637, |
|
"step": 370 |
|
}, |
|
{ |
|
"epoch": 0.7745302713987474, |
|
"grad_norm": 0.04359599728927467, |
|
"learning_rate": 0.000280502045944302, |
|
"loss": 0.3516, |
|
"step": 371 |
|
}, |
|
{ |
|
"epoch": 0.7766179540709812, |
|
"grad_norm": 0.03910171209605865, |
|
"learning_rate": 0.0002798824560192164, |
|
"loss": 0.2793, |
|
"step": 372 |
|
}, |
|
{ |
|
"epoch": 0.778705636743215, |
|
"grad_norm": 0.04498875604580853, |
|
"learning_rate": 0.0002792619525714962, |
|
"loss": 0.334, |
|
"step": 373 |
|
}, |
|
{ |
|
"epoch": 0.7807933194154488, |
|
"grad_norm": 0.046945206399684386, |
|
"learning_rate": 0.00027864054269711666, |
|
"loss": 0.3262, |
|
"step": 374 |
|
}, |
|
{ |
|
"epoch": 0.7828810020876826, |
|
"grad_norm": 0.03720210905131818, |
|
"learning_rate": 0.0002780182335024191, |
|
"loss": 0.252, |
|
"step": 375 |
|
}, |
|
{ |
|
"epoch": 0.7849686847599165, |
|
"grad_norm": 0.03971464889707817, |
|
"learning_rate": 0.00027739503210402907, |
|
"loss": 0.334, |
|
"step": 376 |
|
}, |
|
{ |
|
"epoch": 0.7870563674321504, |
|
"grad_norm": 0.047941221203412314, |
|
"learning_rate": 0.00027677094562877534, |
|
"loss": 0.3066, |
|
"step": 377 |
|
}, |
|
{ |
|
"epoch": 0.7891440501043842, |
|
"grad_norm": 0.04116312254250454, |
|
"learning_rate": 0.0002761459812136083, |
|
"loss": 0.3652, |
|
"step": 378 |
|
}, |
|
{ |
|
"epoch": 0.791231732776618, |
|
"grad_norm": 0.0408608818482939, |
|
"learning_rate": 0.00027552014600551813, |
|
"loss": 0.3359, |
|
"step": 379 |
|
}, |
|
{ |
|
"epoch": 0.7933194154488518, |
|
"grad_norm": 0.039519399095894796, |
|
"learning_rate": 0.00027489344716145344, |
|
"loss": 0.3105, |
|
"step": 380 |
|
}, |
|
{ |
|
"epoch": 0.7954070981210856, |
|
"grad_norm": 0.04468756187207934, |
|
"learning_rate": 0.0002742658918482392, |
|
"loss": 0.334, |
|
"step": 381 |
|
}, |
|
{ |
|
"epoch": 0.7974947807933194, |
|
"grad_norm": 0.03799026650779495, |
|
"learning_rate": 0.00027363748724249485, |
|
"loss": 0.3125, |
|
"step": 382 |
|
}, |
|
{ |
|
"epoch": 0.7995824634655533, |
|
"grad_norm": 0.05260052294591301, |
|
"learning_rate": 0.000273008240530552, |
|
"loss": 0.4316, |
|
"step": 383 |
|
}, |
|
{ |
|
"epoch": 0.8016701461377871, |
|
"grad_norm": 0.04069771345832873, |
|
"learning_rate": 0.00027237815890837283, |
|
"loss": 0.2969, |
|
"step": 384 |
|
}, |
|
{ |
|
"epoch": 0.8037578288100209, |
|
"grad_norm": 0.039739122060491815, |
|
"learning_rate": 0.000271747249581467, |
|
"loss": 0.3555, |
|
"step": 385 |
|
}, |
|
{ |
|
"epoch": 0.8058455114822547, |
|
"grad_norm": 0.04610276901781518, |
|
"learning_rate": 0.00027111551976480996, |
|
"loss": 0.3574, |
|
"step": 386 |
|
}, |
|
{ |
|
"epoch": 0.8079331941544885, |
|
"grad_norm": 0.04374672870028719, |
|
"learning_rate": 0.00027048297668276, |
|
"loss": 0.3047, |
|
"step": 387 |
|
}, |
|
{ |
|
"epoch": 0.8100208768267223, |
|
"grad_norm": 0.04631564065384605, |
|
"learning_rate": 0.0002698496275689758, |
|
"loss": 0.2871, |
|
"step": 388 |
|
}, |
|
{ |
|
"epoch": 0.8121085594989561, |
|
"grad_norm": 0.049176525271089545, |
|
"learning_rate": 0.00026921547966633373, |
|
"loss": 0.2871, |
|
"step": 389 |
|
}, |
|
{ |
|
"epoch": 0.81419624217119, |
|
"grad_norm": 0.040234387326007176, |
|
"learning_rate": 0.0002685805402268449, |
|
"loss": 0.3301, |
|
"step": 390 |
|
}, |
|
{ |
|
"epoch": 0.8162839248434238, |
|
"grad_norm": 0.044678234660364975, |
|
"learning_rate": 0.0002679448165115725, |
|
"loss": 0.2852, |
|
"step": 391 |
|
}, |
|
{ |
|
"epoch": 0.8183716075156576, |
|
"grad_norm": 0.043749807508504976, |
|
"learning_rate": 0.00026730831579054823, |
|
"loss": 0.3418, |
|
"step": 392 |
|
}, |
|
{ |
|
"epoch": 0.8204592901878914, |
|
"grad_norm": 0.03702597280717622, |
|
"learning_rate": 0.00026667104534268984, |
|
"loss": 0.334, |
|
"step": 393 |
|
}, |
|
{ |
|
"epoch": 0.8225469728601252, |
|
"grad_norm": 0.04051649281576064, |
|
"learning_rate": 0.00026603301245571733, |
|
"loss": 0.3652, |
|
"step": 394 |
|
}, |
|
{ |
|
"epoch": 0.824634655532359, |
|
"grad_norm": 0.04070349409408308, |
|
"learning_rate": 0.00026539422442607, |
|
"loss": 0.3164, |
|
"step": 395 |
|
}, |
|
{ |
|
"epoch": 0.826722338204593, |
|
"grad_norm": 0.03685594277413145, |
|
"learning_rate": 0.0002647546885588227, |
|
"loss": 0.2695, |
|
"step": 396 |
|
}, |
|
{ |
|
"epoch": 0.8288100208768268, |
|
"grad_norm": 0.033959522628423805, |
|
"learning_rate": 0.00026411441216760243, |
|
"loss": 0.2422, |
|
"step": 397 |
|
}, |
|
{ |
|
"epoch": 0.8308977035490606, |
|
"grad_norm": 0.03649515909769935, |
|
"learning_rate": 0.00026347340257450473, |
|
"loss": 0.252, |
|
"step": 398 |
|
}, |
|
{ |
|
"epoch": 0.8329853862212944, |
|
"grad_norm": 0.05036202800279952, |
|
"learning_rate": 0.00026283166711001, |
|
"loss": 0.3203, |
|
"step": 399 |
|
}, |
|
{ |
|
"epoch": 0.8350730688935282, |
|
"grad_norm": 0.04724871652205951, |
|
"learning_rate": 0.0002621892131128994, |
|
"loss": 0.3555, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.837160751565762, |
|
"grad_norm": 0.04523467438521327, |
|
"learning_rate": 0.0002615460479301714, |
|
"loss": 0.3809, |
|
"step": 401 |
|
}, |
|
{ |
|
"epoch": 0.8392484342379958, |
|
"grad_norm": 0.04240307801356352, |
|
"learning_rate": 0.0002609021789169571, |
|
"loss": 0.3574, |
|
"step": 402 |
|
}, |
|
{ |
|
"epoch": 0.8413361169102297, |
|
"grad_norm": 0.03998645664485665, |
|
"learning_rate": 0.0002602576134364369, |
|
"loss": 0.3008, |
|
"step": 403 |
|
}, |
|
{ |
|
"epoch": 0.8434237995824635, |
|
"grad_norm": 0.045774083398434376, |
|
"learning_rate": 0.0002596123588597555, |
|
"loss": 0.4043, |
|
"step": 404 |
|
}, |
|
{ |
|
"epoch": 0.8455114822546973, |
|
"grad_norm": 0.043894774464113946, |
|
"learning_rate": 0.0002589664225659382, |
|
"loss": 0.3301, |
|
"step": 405 |
|
}, |
|
{ |
|
"epoch": 0.8475991649269311, |
|
"grad_norm": 0.04133334314623751, |
|
"learning_rate": 0.0002583198119418063, |
|
"loss": 0.2637, |
|
"step": 406 |
|
}, |
|
{ |
|
"epoch": 0.8496868475991649, |
|
"grad_norm": 0.037286338050560974, |
|
"learning_rate": 0.00025767253438189255, |
|
"loss": 0.3164, |
|
"step": 407 |
|
}, |
|
{ |
|
"epoch": 0.8517745302713987, |
|
"grad_norm": 0.04288387043023777, |
|
"learning_rate": 0.0002570245972883568, |
|
"loss": 0.3164, |
|
"step": 408 |
|
}, |
|
{ |
|
"epoch": 0.8538622129436325, |
|
"grad_norm": 0.04117605927774004, |
|
"learning_rate": 0.000256376008070901, |
|
"loss": 0.3457, |
|
"step": 409 |
|
}, |
|
{ |
|
"epoch": 0.8559498956158664, |
|
"grad_norm": 0.040660730568059064, |
|
"learning_rate": 0.00025572677414668497, |
|
"loss": 0.3691, |
|
"step": 410 |
|
}, |
|
{ |
|
"epoch": 0.8580375782881002, |
|
"grad_norm": 0.03473424285365033, |
|
"learning_rate": 0.00025507690294024105, |
|
"loss": 0.252, |
|
"step": 411 |
|
}, |
|
{ |
|
"epoch": 0.860125260960334, |
|
"grad_norm": 0.04106073793621654, |
|
"learning_rate": 0.00025442640188338945, |
|
"loss": 0.2871, |
|
"step": 412 |
|
}, |
|
{ |
|
"epoch": 0.8622129436325678, |
|
"grad_norm": 0.03780921487044816, |
|
"learning_rate": 0.00025377527841515343, |
|
"loss": 0.2969, |
|
"step": 413 |
|
}, |
|
{ |
|
"epoch": 0.8643006263048016, |
|
"grad_norm": 0.03779696762506096, |
|
"learning_rate": 0.00025312353998167397, |
|
"loss": 0.291, |
|
"step": 414 |
|
}, |
|
{ |
|
"epoch": 0.8663883089770354, |
|
"grad_norm": 0.04781732258675552, |
|
"learning_rate": 0.00025247119403612456, |
|
"loss": 0.3301, |
|
"step": 415 |
|
}, |
|
{ |
|
"epoch": 0.8684759916492694, |
|
"grad_norm": 0.03912752520944136, |
|
"learning_rate": 0.0002518182480386261, |
|
"loss": 0.3477, |
|
"step": 416 |
|
}, |
|
{ |
|
"epoch": 0.8705636743215032, |
|
"grad_norm": 0.043540503401128916, |
|
"learning_rate": 0.0002511647094561619, |
|
"loss": 0.3398, |
|
"step": 417 |
|
}, |
|
{ |
|
"epoch": 0.872651356993737, |
|
"grad_norm": 0.04207586943763798, |
|
"learning_rate": 0.00025051058576249153, |
|
"loss": 0.3398, |
|
"step": 418 |
|
}, |
|
{ |
|
"epoch": 0.8747390396659708, |
|
"grad_norm": 0.03596881591749821, |
|
"learning_rate": 0.00024985588443806624, |
|
"loss": 0.2715, |
|
"step": 419 |
|
}, |
|
{ |
|
"epoch": 0.8768267223382046, |
|
"grad_norm": 0.042287007835322306, |
|
"learning_rate": 0.00024920061296994253, |
|
"loss": 0.3086, |
|
"step": 420 |
|
}, |
|
{ |
|
"epoch": 0.8789144050104384, |
|
"grad_norm": 0.03856865323635185, |
|
"learning_rate": 0.00024854477885169734, |
|
"loss": 0.2871, |
|
"step": 421 |
|
}, |
|
{ |
|
"epoch": 0.8810020876826722, |
|
"grad_norm": 0.03879769693027103, |
|
"learning_rate": 0.0002478883895833417, |
|
"loss": 0.2617, |
|
"step": 422 |
|
}, |
|
{ |
|
"epoch": 0.8830897703549061, |
|
"grad_norm": 0.04500551833152804, |
|
"learning_rate": 0.00024723145267123556, |
|
"loss": 0.2793, |
|
"step": 423 |
|
}, |
|
{ |
|
"epoch": 0.8851774530271399, |
|
"grad_norm": 0.03925914590146329, |
|
"learning_rate": 0.0002465739756280013, |
|
"loss": 0.3086, |
|
"step": 424 |
|
}, |
|
{ |
|
"epoch": 0.8872651356993737, |
|
"grad_norm": 0.04390031842385643, |
|
"learning_rate": 0.0002459159659724383, |
|
"loss": 0.3594, |
|
"step": 425 |
|
}, |
|
{ |
|
"epoch": 0.8893528183716075, |
|
"grad_norm": 0.04015458555102483, |
|
"learning_rate": 0.00024525743122943684, |
|
"loss": 0.2773, |
|
"step": 426 |
|
}, |
|
{ |
|
"epoch": 0.8914405010438413, |
|
"grad_norm": 0.043251502572368115, |
|
"learning_rate": 0.000244598378929892, |
|
"loss": 0.3633, |
|
"step": 427 |
|
}, |
|
{ |
|
"epoch": 0.8935281837160751, |
|
"grad_norm": 0.041718720849523466, |
|
"learning_rate": 0.00024393881661061747, |
|
"loss": 0.3418, |
|
"step": 428 |
|
}, |
|
{ |
|
"epoch": 0.8956158663883089, |
|
"grad_norm": 0.03667112112306563, |
|
"learning_rate": 0.0002432787518142596, |
|
"loss": 0.2754, |
|
"step": 429 |
|
}, |
|
{ |
|
"epoch": 0.8977035490605428, |
|
"grad_norm": 0.04034875824232465, |
|
"learning_rate": 0.0002426181920892108, |
|
"loss": 0.3242, |
|
"step": 430 |
|
}, |
|
{ |
|
"epoch": 0.8997912317327766, |
|
"grad_norm": 0.0416779068349274, |
|
"learning_rate": 0.0002419571449895236, |
|
"loss": 0.2793, |
|
"step": 431 |
|
}, |
|
{ |
|
"epoch": 0.9018789144050104, |
|
"grad_norm": 0.04492298155955886, |
|
"learning_rate": 0.00024129561807482378, |
|
"loss": 0.373, |
|
"step": 432 |
|
}, |
|
{ |
|
"epoch": 0.9039665970772442, |
|
"grad_norm": 0.0358568800030184, |
|
"learning_rate": 0.00024063361891022455, |
|
"loss": 0.293, |
|
"step": 433 |
|
}, |
|
{ |
|
"epoch": 0.906054279749478, |
|
"grad_norm": 0.041240679259523606, |
|
"learning_rate": 0.00023997115506623948, |
|
"loss": 0.3477, |
|
"step": 434 |
|
}, |
|
{ |
|
"epoch": 0.9081419624217119, |
|
"grad_norm": 0.04776747282663658, |
|
"learning_rate": 0.00023930823411869617, |
|
"loss": 0.3633, |
|
"step": 435 |
|
}, |
|
{ |
|
"epoch": 0.9102296450939458, |
|
"grad_norm": 0.03739632540656231, |
|
"learning_rate": 0.00023864486364864963, |
|
"loss": 0.3027, |
|
"step": 436 |
|
}, |
|
{ |
|
"epoch": 0.9123173277661796, |
|
"grad_norm": 0.04013506353404083, |
|
"learning_rate": 0.00023798105124229536, |
|
"loss": 0.2793, |
|
"step": 437 |
|
}, |
|
{ |
|
"epoch": 0.9144050104384134, |
|
"grad_norm": 0.044295244931196354, |
|
"learning_rate": 0.00023731680449088312, |
|
"loss": 0.377, |
|
"step": 438 |
|
}, |
|
{ |
|
"epoch": 0.9164926931106472, |
|
"grad_norm": 0.03805031241226096, |
|
"learning_rate": 0.00023665213099062938, |
|
"loss": 0.2559, |
|
"step": 439 |
|
}, |
|
{ |
|
"epoch": 0.918580375782881, |
|
"grad_norm": 0.03767845062036613, |
|
"learning_rate": 0.00023598703834263113, |
|
"loss": 0.3125, |
|
"step": 440 |
|
}, |
|
{ |
|
"epoch": 0.9206680584551148, |
|
"grad_norm": 0.04968215089246422, |
|
"learning_rate": 0.0002353215341527785, |
|
"loss": 0.3223, |
|
"step": 441 |
|
}, |
|
{ |
|
"epoch": 0.9227557411273486, |
|
"grad_norm": 0.036246340480605606, |
|
"learning_rate": 0.00023465562603166808, |
|
"loss": 0.2734, |
|
"step": 442 |
|
}, |
|
{ |
|
"epoch": 0.9248434237995825, |
|
"grad_norm": 0.034358684978586476, |
|
"learning_rate": 0.00023398932159451557, |
|
"loss": 0.3457, |
|
"step": 443 |
|
}, |
|
{ |
|
"epoch": 0.9269311064718163, |
|
"grad_norm": 0.03631746144581004, |
|
"learning_rate": 0.000233322628461069, |
|
"loss": 0.3672, |
|
"step": 444 |
|
}, |
|
{ |
|
"epoch": 0.9290187891440501, |
|
"grad_norm": 0.037623854588027666, |
|
"learning_rate": 0.0002326555542555215, |
|
"loss": 0.3047, |
|
"step": 445 |
|
}, |
|
{ |
|
"epoch": 0.9311064718162839, |
|
"grad_norm": 0.037886888034629966, |
|
"learning_rate": 0.0002319881066064239, |
|
"loss": 0.2812, |
|
"step": 446 |
|
}, |
|
{ |
|
"epoch": 0.9331941544885177, |
|
"grad_norm": 0.039683782098775454, |
|
"learning_rate": 0.00023132029314659795, |
|
"loss": 0.2969, |
|
"step": 447 |
|
}, |
|
{ |
|
"epoch": 0.9352818371607515, |
|
"grad_norm": 0.032859528564560474, |
|
"learning_rate": 0.00023065212151304843, |
|
"loss": 0.2617, |
|
"step": 448 |
|
}, |
|
{ |
|
"epoch": 0.9373695198329853, |
|
"grad_norm": 0.032092006779359715, |
|
"learning_rate": 0.00022998359934687645, |
|
"loss": 0.2656, |
|
"step": 449 |
|
}, |
|
{ |
|
"epoch": 0.9394572025052192, |
|
"grad_norm": 0.03973715339510534, |
|
"learning_rate": 0.00022931473429319143, |
|
"loss": 0.3242, |
|
"step": 450 |
|
}, |
|
{ |
|
"epoch": 0.941544885177453, |
|
"grad_norm": 0.040296933695116964, |
|
"learning_rate": 0.00022864553400102437, |
|
"loss": 0.3281, |
|
"step": 451 |
|
}, |
|
{ |
|
"epoch": 0.9436325678496869, |
|
"grad_norm": 0.03972360152628702, |
|
"learning_rate": 0.00022797600612323965, |
|
"loss": 0.3301, |
|
"step": 452 |
|
}, |
|
{ |
|
"epoch": 0.9457202505219207, |
|
"grad_norm": 0.03196535128099967, |
|
"learning_rate": 0.00022730615831644807, |
|
"loss": 0.2617, |
|
"step": 453 |
|
}, |
|
{ |
|
"epoch": 0.9478079331941545, |
|
"grad_norm": 0.041518745854850714, |
|
"learning_rate": 0.00022663599824091892, |
|
"loss": 0.3535, |
|
"step": 454 |
|
}, |
|
{ |
|
"epoch": 0.9498956158663883, |
|
"grad_norm": 0.03987544052335102, |
|
"learning_rate": 0.00022596553356049283, |
|
"loss": 0.3184, |
|
"step": 455 |
|
}, |
|
{ |
|
"epoch": 0.9519832985386222, |
|
"grad_norm": 0.03790469940548919, |
|
"learning_rate": 0.0002252947719424935, |
|
"loss": 0.2988, |
|
"step": 456 |
|
}, |
|
{ |
|
"epoch": 0.954070981210856, |
|
"grad_norm": 0.03494334490842234, |
|
"learning_rate": 0.00022462372105764065, |
|
"loss": 0.2793, |
|
"step": 457 |
|
}, |
|
{ |
|
"epoch": 0.9561586638830898, |
|
"grad_norm": 0.03463196563976093, |
|
"learning_rate": 0.0002239523885799618, |
|
"loss": 0.2637, |
|
"step": 458 |
|
}, |
|
{ |
|
"epoch": 0.9582463465553236, |
|
"grad_norm": 0.03593899157528181, |
|
"learning_rate": 0.0002232807821867048, |
|
"loss": 0.3184, |
|
"step": 459 |
|
}, |
|
{ |
|
"epoch": 0.9603340292275574, |
|
"grad_norm": 0.03890353431042402, |
|
"learning_rate": 0.00022260890955825001, |
|
"loss": 0.3008, |
|
"step": 460 |
|
}, |
|
{ |
|
"epoch": 0.9624217118997912, |
|
"grad_norm": 0.033343558090590815, |
|
"learning_rate": 0.00022193677837802235, |
|
"loss": 0.2891, |
|
"step": 461 |
|
}, |
|
{ |
|
"epoch": 0.964509394572025, |
|
"grad_norm": 0.037945187747727475, |
|
"learning_rate": 0.0002212643963324035, |
|
"loss": 0.3574, |
|
"step": 462 |
|
}, |
|
{ |
|
"epoch": 0.9665970772442589, |
|
"grad_norm": 0.03770022520852902, |
|
"learning_rate": 0.00022059177111064391, |
|
"loss": 0.3262, |
|
"step": 463 |
|
}, |
|
{ |
|
"epoch": 0.9686847599164927, |
|
"grad_norm": 0.03875032196468757, |
|
"learning_rate": 0.00021991891040477516, |
|
"loss": 0.3145, |
|
"step": 464 |
|
}, |
|
{ |
|
"epoch": 0.9707724425887265, |
|
"grad_norm": 0.03788954429707523, |
|
"learning_rate": 0.00021924582190952153, |
|
"loss": 0.3008, |
|
"step": 465 |
|
}, |
|
{ |
|
"epoch": 0.9728601252609603, |
|
"grad_norm": 0.035023715952115834, |
|
"learning_rate": 0.00021857251332221256, |
|
"loss": 0.2412, |
|
"step": 466 |
|
}, |
|
{ |
|
"epoch": 0.9749478079331941, |
|
"grad_norm": 0.037210987286251065, |
|
"learning_rate": 0.0002178989923426944, |
|
"loss": 0.3398, |
|
"step": 467 |
|
}, |
|
{ |
|
"epoch": 0.9770354906054279, |
|
"grad_norm": 0.03207771649660305, |
|
"learning_rate": 0.0002172252666732424, |
|
"loss": 0.2578, |
|
"step": 468 |
|
}, |
|
{ |
|
"epoch": 0.9791231732776617, |
|
"grad_norm": 0.031951869684034837, |
|
"learning_rate": 0.00021655134401847245, |
|
"loss": 0.2695, |
|
"step": 469 |
|
}, |
|
{ |
|
"epoch": 0.9812108559498957, |
|
"grad_norm": 0.03674391020210522, |
|
"learning_rate": 0.00021587723208525334, |
|
"loss": 0.3066, |
|
"step": 470 |
|
}, |
|
{ |
|
"epoch": 0.9832985386221295, |
|
"grad_norm": 0.035932078354231324, |
|
"learning_rate": 0.00021520293858261823, |
|
"loss": 0.3086, |
|
"step": 471 |
|
}, |
|
{ |
|
"epoch": 0.9853862212943633, |
|
"grad_norm": 0.03876657300516048, |
|
"learning_rate": 0.00021452847122167677, |
|
"loss": 0.3008, |
|
"step": 472 |
|
}, |
|
{ |
|
"epoch": 0.9874739039665971, |
|
"grad_norm": 0.032766584147543416, |
|
"learning_rate": 0.00021385383771552684, |
|
"loss": 0.3047, |
|
"step": 473 |
|
}, |
|
{ |
|
"epoch": 0.9895615866388309, |
|
"grad_norm": 0.04086665364700512, |
|
"learning_rate": 0.00021317904577916635, |
|
"loss": 0.3223, |
|
"step": 474 |
|
}, |
|
{ |
|
"epoch": 0.9916492693110647, |
|
"grad_norm": 0.043130532139654114, |
|
"learning_rate": 0.0002125041031294049, |
|
"loss": 0.3242, |
|
"step": 475 |
|
}, |
|
{ |
|
"epoch": 0.9937369519832986, |
|
"grad_norm": 0.04273564060035245, |
|
"learning_rate": 0.00021182901748477564, |
|
"loss": 0.4023, |
|
"step": 476 |
|
}, |
|
{ |
|
"epoch": 0.9958246346555324, |
|
"grad_norm": 0.03628393896960175, |
|
"learning_rate": 0.00021115379656544715, |
|
"loss": 0.2969, |
|
"step": 477 |
|
}, |
|
{ |
|
"epoch": 0.9979123173277662, |
|
"grad_norm": 0.03800046175114251, |
|
"learning_rate": 0.00021047844809313467, |
|
"loss": 0.332, |
|
"step": 478 |
|
}, |
|
{ |
|
"epoch": 1.0, |
|
"grad_norm": 0.038197115177252855, |
|
"learning_rate": 0.00020980297979101252, |
|
"loss": 0.2695, |
|
"step": 479 |
|
}, |
|
{ |
|
"epoch": 1.0020876826722338, |
|
"grad_norm": 0.032423776045183476, |
|
"learning_rate": 0.000209127399383625, |
|
"loss": 0.2021, |
|
"step": 480 |
|
}, |
|
{ |
|
"epoch": 1.0041753653444676, |
|
"grad_norm": 0.03224992988614844, |
|
"learning_rate": 0.00020845171459679866, |
|
"loss": 0.2139, |
|
"step": 481 |
|
}, |
|
{ |
|
"epoch": 1.0062630480167014, |
|
"grad_norm": 0.03676757603527534, |
|
"learning_rate": 0.00020777593315755358, |
|
"loss": 0.249, |
|
"step": 482 |
|
}, |
|
{ |
|
"epoch": 1.0083507306889352, |
|
"grad_norm": 0.03889498255916948, |
|
"learning_rate": 0.00020710006279401531, |
|
"loss": 0.3047, |
|
"step": 483 |
|
}, |
|
{ |
|
"epoch": 1.010438413361169, |
|
"grad_norm": 0.03394341377626214, |
|
"learning_rate": 0.00020642411123532605, |
|
"loss": 0.2598, |
|
"step": 484 |
|
}, |
|
{ |
|
"epoch": 1.0125260960334028, |
|
"grad_norm": 0.03195140878316827, |
|
"learning_rate": 0.0002057480862115569, |
|
"loss": 0.2266, |
|
"step": 485 |
|
}, |
|
{ |
|
"epoch": 1.0146137787056368, |
|
"grad_norm": 0.03236779419185144, |
|
"learning_rate": 0.0002050719954536187, |
|
"loss": 0.2373, |
|
"step": 486 |
|
}, |
|
{ |
|
"epoch": 1.0167014613778707, |
|
"grad_norm": 0.033762294027972366, |
|
"learning_rate": 0.0002043958466931744, |
|
"loss": 0.2324, |
|
"step": 487 |
|
}, |
|
{ |
|
"epoch": 1.0187891440501045, |
|
"grad_norm": 0.03225364955535692, |
|
"learning_rate": 0.00020371964766254988, |
|
"loss": 0.2246, |
|
"step": 488 |
|
}, |
|
{ |
|
"epoch": 1.0208768267223383, |
|
"grad_norm": 0.03549374570066917, |
|
"learning_rate": 0.00020304340609464627, |
|
"loss": 0.2559, |
|
"step": 489 |
|
}, |
|
{ |
|
"epoch": 1.022964509394572, |
|
"grad_norm": 0.039376031742883684, |
|
"learning_rate": 0.00020236712972285088, |
|
"loss": 0.291, |
|
"step": 490 |
|
}, |
|
{ |
|
"epoch": 1.0250521920668059, |
|
"grad_norm": 0.033112169954733074, |
|
"learning_rate": 0.0002016908262809491, |
|
"loss": 0.2188, |
|
"step": 491 |
|
}, |
|
{ |
|
"epoch": 1.0271398747390397, |
|
"grad_norm": 0.03556499196371877, |
|
"learning_rate": 0.00020101450350303596, |
|
"loss": 0.2363, |
|
"step": 492 |
|
}, |
|
{ |
|
"epoch": 1.0292275574112735, |
|
"grad_norm": 0.03747952002586393, |
|
"learning_rate": 0.0002003381691234275, |
|
"loss": 0.2734, |
|
"step": 493 |
|
}, |
|
{ |
|
"epoch": 1.0313152400835073, |
|
"grad_norm": 0.039212245318626486, |
|
"learning_rate": 0.00019966183087657255, |
|
"loss": 0.2812, |
|
"step": 494 |
|
}, |
|
{ |
|
"epoch": 1.033402922755741, |
|
"grad_norm": 0.038556823852449205, |
|
"learning_rate": 0.00019898549649696409, |
|
"loss": 0.252, |
|
"step": 495 |
|
}, |
|
{ |
|
"epoch": 1.0354906054279749, |
|
"grad_norm": 0.03678903694409147, |
|
"learning_rate": 0.00019830917371905095, |
|
"loss": 0.2227, |
|
"step": 496 |
|
}, |
|
{ |
|
"epoch": 1.0375782881002087, |
|
"grad_norm": 0.035536317646152164, |
|
"learning_rate": 0.00019763287027714917, |
|
"loss": 0.2451, |
|
"step": 497 |
|
}, |
|
{ |
|
"epoch": 1.0396659707724425, |
|
"grad_norm": 0.03669268286478292, |
|
"learning_rate": 0.00019695659390535377, |
|
"loss": 0.2158, |
|
"step": 498 |
|
}, |
|
{ |
|
"epoch": 1.0417536534446765, |
|
"grad_norm": 0.0365044069944666, |
|
"learning_rate": 0.0001962803523374501, |
|
"loss": 0.2334, |
|
"step": 499 |
|
}, |
|
{ |
|
"epoch": 1.0438413361169103, |
|
"grad_norm": 0.04311308542212124, |
|
"learning_rate": 0.00019560415330682567, |
|
"loss": 0.2754, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 1.0459290187891441, |
|
"grad_norm": 0.03807877662781654, |
|
"learning_rate": 0.00019492800454638128, |
|
"loss": 0.2207, |
|
"step": 501 |
|
}, |
|
{ |
|
"epoch": 1.048016701461378, |
|
"grad_norm": 0.037347719889777004, |
|
"learning_rate": 0.00019425191378844315, |
|
"loss": 0.2598, |
|
"step": 502 |
|
}, |
|
{ |
|
"epoch": 1.0501043841336117, |
|
"grad_norm": 0.041720879231187474, |
|
"learning_rate": 0.00019357588876467403, |
|
"loss": 0.2832, |
|
"step": 503 |
|
}, |
|
{ |
|
"epoch": 1.0521920668058455, |
|
"grad_norm": 0.03611878979996833, |
|
"learning_rate": 0.0001928999372059848, |
|
"loss": 0.2305, |
|
"step": 504 |
|
}, |
|
{ |
|
"epoch": 1.0542797494780793, |
|
"grad_norm": 0.03914121207636423, |
|
"learning_rate": 0.00019222406684244652, |
|
"loss": 0.2715, |
|
"step": 505 |
|
}, |
|
{ |
|
"epoch": 1.0563674321503131, |
|
"grad_norm": 0.045008260139354196, |
|
"learning_rate": 0.00019154828540320141, |
|
"loss": 0.3496, |
|
"step": 506 |
|
}, |
|
{ |
|
"epoch": 1.058455114822547, |
|
"grad_norm": 0.06925075294932237, |
|
"learning_rate": 0.00019087260061637507, |
|
"loss": 0.21, |
|
"step": 507 |
|
}, |
|
{ |
|
"epoch": 1.0605427974947808, |
|
"grad_norm": 0.042949891828889544, |
|
"learning_rate": 0.00019019702020898753, |
|
"loss": 0.3164, |
|
"step": 508 |
|
}, |
|
{ |
|
"epoch": 1.0626304801670146, |
|
"grad_norm": 0.03336924307296428, |
|
"learning_rate": 0.00018952155190686535, |
|
"loss": 0.2334, |
|
"step": 509 |
|
}, |
|
{ |
|
"epoch": 1.0647181628392484, |
|
"grad_norm": 0.037171672304770545, |
|
"learning_rate": 0.00018884620343455292, |
|
"loss": 0.2432, |
|
"step": 510 |
|
}, |
|
{ |
|
"epoch": 1.0668058455114822, |
|
"grad_norm": 0.03845099190077751, |
|
"learning_rate": 0.0001881709825152244, |
|
"loss": 0.2793, |
|
"step": 511 |
|
}, |
|
{ |
|
"epoch": 1.068893528183716, |
|
"grad_norm": 0.03411076783878538, |
|
"learning_rate": 0.00018749589687059513, |
|
"loss": 0.2041, |
|
"step": 512 |
|
}, |
|
{ |
|
"epoch": 1.07098121085595, |
|
"grad_norm": 0.03701566955647055, |
|
"learning_rate": 0.00018682095422083373, |
|
"loss": 0.2305, |
|
"step": 513 |
|
}, |
|
{ |
|
"epoch": 1.0730688935281838, |
|
"grad_norm": 0.03954320852756579, |
|
"learning_rate": 0.00018614616228447318, |
|
"loss": 0.2354, |
|
"step": 514 |
|
}, |
|
{ |
|
"epoch": 1.0751565762004176, |
|
"grad_norm": 0.04313297344221612, |
|
"learning_rate": 0.00018547152877832328, |
|
"loss": 0.2656, |
|
"step": 515 |
|
}, |
|
{ |
|
"epoch": 1.0772442588726514, |
|
"grad_norm": 0.03799814412465588, |
|
"learning_rate": 0.0001847970614173818, |
|
"loss": 0.2285, |
|
"step": 516 |
|
}, |
|
{ |
|
"epoch": 1.0793319415448852, |
|
"grad_norm": 0.04096742193043013, |
|
"learning_rate": 0.0001841227679147467, |
|
"loss": 0.25, |
|
"step": 517 |
|
}, |
|
{ |
|
"epoch": 1.081419624217119, |
|
"grad_norm": 0.03795069745138144, |
|
"learning_rate": 0.00018344865598152754, |
|
"loss": 0.2471, |
|
"step": 518 |
|
}, |
|
{ |
|
"epoch": 1.0835073068893528, |
|
"grad_norm": 0.04325821353197005, |
|
"learning_rate": 0.00018277473332675758, |
|
"loss": 0.2754, |
|
"step": 519 |
|
}, |
|
{ |
|
"epoch": 1.0855949895615866, |
|
"grad_norm": 0.04070425486872679, |
|
"learning_rate": 0.0001821010076573056, |
|
"loss": 0.2617, |
|
"step": 520 |
|
}, |
|
{ |
|
"epoch": 1.0876826722338204, |
|
"grad_norm": 0.031587246842812465, |
|
"learning_rate": 0.00018142748667778746, |
|
"loss": 0.1777, |
|
"step": 521 |
|
}, |
|
{ |
|
"epoch": 1.0897703549060542, |
|
"grad_norm": 0.036539737613616244, |
|
"learning_rate": 0.00018075417809047857, |
|
"loss": 0.2773, |
|
"step": 522 |
|
}, |
|
{ |
|
"epoch": 1.091858037578288, |
|
"grad_norm": 0.041178092025696736, |
|
"learning_rate": 0.00018008108959522497, |
|
"loss": 0.2676, |
|
"step": 523 |
|
}, |
|
{ |
|
"epoch": 1.0939457202505218, |
|
"grad_norm": 0.04046942553591543, |
|
"learning_rate": 0.00017940822888935616, |
|
"loss": 0.2217, |
|
"step": 524 |
|
}, |
|
{ |
|
"epoch": 1.0960334029227556, |
|
"grad_norm": 0.03449629028853095, |
|
"learning_rate": 0.0001787356036675966, |
|
"loss": 0.2061, |
|
"step": 525 |
|
}, |
|
{ |
|
"epoch": 1.0981210855949897, |
|
"grad_norm": 0.03986588393895996, |
|
"learning_rate": 0.0001780632216219777, |
|
"loss": 0.3125, |
|
"step": 526 |
|
}, |
|
{ |
|
"epoch": 1.1002087682672235, |
|
"grad_norm": 0.0333875432706545, |
|
"learning_rate": 0.00017739109044175003, |
|
"loss": 0.1787, |
|
"step": 527 |
|
}, |
|
{ |
|
"epoch": 1.1022964509394573, |
|
"grad_norm": 0.04046361146111627, |
|
"learning_rate": 0.00017671921781329522, |
|
"loss": 0.2715, |
|
"step": 528 |
|
}, |
|
{ |
|
"epoch": 1.104384133611691, |
|
"grad_norm": 0.04262435868048576, |
|
"learning_rate": 0.00017604761142003827, |
|
"loss": 0.2559, |
|
"step": 529 |
|
}, |
|
{ |
|
"epoch": 1.1064718162839249, |
|
"grad_norm": 0.03428115599372914, |
|
"learning_rate": 0.0001753762789423594, |
|
"loss": 0.2061, |
|
"step": 530 |
|
}, |
|
{ |
|
"epoch": 1.1085594989561587, |
|
"grad_norm": 0.03868138089293832, |
|
"learning_rate": 0.00017470522805750654, |
|
"loss": 0.2236, |
|
"step": 531 |
|
}, |
|
{ |
|
"epoch": 1.1106471816283925, |
|
"grad_norm": 0.03475622072699226, |
|
"learning_rate": 0.00017403446643950724, |
|
"loss": 0.248, |
|
"step": 532 |
|
}, |
|
{ |
|
"epoch": 1.1127348643006263, |
|
"grad_norm": 0.03979226115172334, |
|
"learning_rate": 0.00017336400175908112, |
|
"loss": 0.2988, |
|
"step": 533 |
|
}, |
|
{ |
|
"epoch": 1.11482254697286, |
|
"grad_norm": 0.036336277616070035, |
|
"learning_rate": 0.000172693841683552, |
|
"loss": 0.2236, |
|
"step": 534 |
|
}, |
|
{ |
|
"epoch": 1.116910229645094, |
|
"grad_norm": 0.039263581746387936, |
|
"learning_rate": 0.0001720239938767604, |
|
"loss": 0.332, |
|
"step": 535 |
|
}, |
|
{ |
|
"epoch": 1.1189979123173277, |
|
"grad_norm": 0.04395872414762282, |
|
"learning_rate": 0.00017135446599897565, |
|
"loss": 0.2734, |
|
"step": 536 |
|
}, |
|
{ |
|
"epoch": 1.1210855949895615, |
|
"grad_norm": 0.03867598581621191, |
|
"learning_rate": 0.00017068526570680853, |
|
"loss": 0.249, |
|
"step": 537 |
|
}, |
|
{ |
|
"epoch": 1.1231732776617953, |
|
"grad_norm": 0.035662617079617494, |
|
"learning_rate": 0.00017001640065312357, |
|
"loss": 0.2041, |
|
"step": 538 |
|
}, |
|
{ |
|
"epoch": 1.1252609603340291, |
|
"grad_norm": 0.04078186211455469, |
|
"learning_rate": 0.00016934787848695156, |
|
"loss": 0.2119, |
|
"step": 539 |
|
}, |
|
{ |
|
"epoch": 1.1273486430062631, |
|
"grad_norm": 0.03947326929596921, |
|
"learning_rate": 0.00016867970685340215, |
|
"loss": 0.2393, |
|
"step": 540 |
|
}, |
|
{ |
|
"epoch": 1.129436325678497, |
|
"grad_norm": 0.04449645249301931, |
|
"learning_rate": 0.00016801189339357616, |
|
"loss": 0.3203, |
|
"step": 541 |
|
}, |
|
{ |
|
"epoch": 1.1315240083507307, |
|
"grad_norm": 0.046340346999829446, |
|
"learning_rate": 0.00016734444574447862, |
|
"loss": 0.2598, |
|
"step": 542 |
|
}, |
|
{ |
|
"epoch": 1.1336116910229646, |
|
"grad_norm": 0.03957931201804222, |
|
"learning_rate": 0.00016667737153893105, |
|
"loss": 0.2402, |
|
"step": 543 |
|
}, |
|
{ |
|
"epoch": 1.1356993736951984, |
|
"grad_norm": 0.03703757920078097, |
|
"learning_rate": 0.00016601067840548453, |
|
"loss": 0.1924, |
|
"step": 544 |
|
}, |
|
{ |
|
"epoch": 1.1377870563674322, |
|
"grad_norm": 0.03682994100461504, |
|
"learning_rate": 0.000165344373968332, |
|
"loss": 0.2383, |
|
"step": 545 |
|
}, |
|
{ |
|
"epoch": 1.139874739039666, |
|
"grad_norm": 0.03920731523127231, |
|
"learning_rate": 0.00016467846584722153, |
|
"loss": 0.2734, |
|
"step": 546 |
|
}, |
|
{ |
|
"epoch": 1.1419624217118998, |
|
"grad_norm": 0.04276682539134295, |
|
"learning_rate": 0.0001640129616573689, |
|
"loss": 0.2852, |
|
"step": 547 |
|
}, |
|
{ |
|
"epoch": 1.1440501043841336, |
|
"grad_norm": 0.03614032190490994, |
|
"learning_rate": 0.00016334786900937067, |
|
"loss": 0.2354, |
|
"step": 548 |
|
}, |
|
{ |
|
"epoch": 1.1461377870563674, |
|
"grad_norm": 0.041006439179447914, |
|
"learning_rate": 0.00016268319550911695, |
|
"loss": 0.2432, |
|
"step": 549 |
|
}, |
|
{ |
|
"epoch": 1.1482254697286012, |
|
"grad_norm": 0.044517563506872825, |
|
"learning_rate": 0.00016201894875770466, |
|
"loss": 0.2217, |
|
"step": 550 |
|
}, |
|
{ |
|
"epoch": 1.150313152400835, |
|
"grad_norm": 0.03655321665381233, |
|
"learning_rate": 0.00016135513635135045, |
|
"loss": 0.2031, |
|
"step": 551 |
|
}, |
|
{ |
|
"epoch": 1.152400835073069, |
|
"grad_norm": 0.04079189189547784, |
|
"learning_rate": 0.00016069176588130388, |
|
"loss": 0.3027, |
|
"step": 552 |
|
}, |
|
{ |
|
"epoch": 1.1544885177453028, |
|
"grad_norm": 0.03485321031694839, |
|
"learning_rate": 0.00016002884493376054, |
|
"loss": 0.2275, |
|
"step": 553 |
|
}, |
|
{ |
|
"epoch": 1.1565762004175366, |
|
"grad_norm": 0.03052141642274577, |
|
"learning_rate": 0.00015936638108977546, |
|
"loss": 0.1992, |
|
"step": 554 |
|
}, |
|
{ |
|
"epoch": 1.1586638830897704, |
|
"grad_norm": 0.03724204209659217, |
|
"learning_rate": 0.0001587043819251762, |
|
"loss": 0.2207, |
|
"step": 555 |
|
}, |
|
{ |
|
"epoch": 1.1607515657620042, |
|
"grad_norm": 0.03701626710699115, |
|
"learning_rate": 0.00015804285501047646, |
|
"loss": 0.2041, |
|
"step": 556 |
|
}, |
|
{ |
|
"epoch": 1.162839248434238, |
|
"grad_norm": 0.04171363907307022, |
|
"learning_rate": 0.00015738180791078922, |
|
"loss": 0.2715, |
|
"step": 557 |
|
}, |
|
{ |
|
"epoch": 1.1649269311064718, |
|
"grad_norm": 0.0392243815886844, |
|
"learning_rate": 0.00015672124818574042, |
|
"loss": 0.2734, |
|
"step": 558 |
|
}, |
|
{ |
|
"epoch": 1.1670146137787056, |
|
"grad_norm": 0.034576244303318175, |
|
"learning_rate": 0.0001560611833893826, |
|
"loss": 0.207, |
|
"step": 559 |
|
}, |
|
{ |
|
"epoch": 1.1691022964509394, |
|
"grad_norm": 0.03659195217316462, |
|
"learning_rate": 0.0001554016210701081, |
|
"loss": 0.2178, |
|
"step": 560 |
|
}, |
|
{ |
|
"epoch": 1.1711899791231732, |
|
"grad_norm": 0.04227483515507797, |
|
"learning_rate": 0.00015474256877056326, |
|
"loss": 0.3086, |
|
"step": 561 |
|
}, |
|
{ |
|
"epoch": 1.173277661795407, |
|
"grad_norm": 0.0359857035728927, |
|
"learning_rate": 0.00015408403402756177, |
|
"loss": 0.2617, |
|
"step": 562 |
|
}, |
|
{ |
|
"epoch": 1.1753653444676408, |
|
"grad_norm": 0.03680442854270533, |
|
"learning_rate": 0.0001534260243719988, |
|
"loss": 0.209, |
|
"step": 563 |
|
}, |
|
{ |
|
"epoch": 1.1774530271398747, |
|
"grad_norm": 0.03670404417452906, |
|
"learning_rate": 0.0001527685473287645, |
|
"loss": 0.2197, |
|
"step": 564 |
|
}, |
|
{ |
|
"epoch": 1.1795407098121085, |
|
"grad_norm": 0.0393433887764882, |
|
"learning_rate": 0.00015211161041665832, |
|
"loss": 0.2324, |
|
"step": 565 |
|
}, |
|
{ |
|
"epoch": 1.1816283924843423, |
|
"grad_norm": 0.03642939621217633, |
|
"learning_rate": 0.00015145522114830268, |
|
"loss": 0.2393, |
|
"step": 566 |
|
}, |
|
{ |
|
"epoch": 1.1837160751565763, |
|
"grad_norm": 0.03878579783362816, |
|
"learning_rate": 0.00015079938703005752, |
|
"loss": 0.2559, |
|
"step": 567 |
|
}, |
|
{ |
|
"epoch": 1.18580375782881, |
|
"grad_norm": 0.040970554736426454, |
|
"learning_rate": 0.0001501441155619338, |
|
"loss": 0.2139, |
|
"step": 568 |
|
}, |
|
{ |
|
"epoch": 1.187891440501044, |
|
"grad_norm": 0.036820951922352504, |
|
"learning_rate": 0.0001494894142375085, |
|
"loss": 0.1719, |
|
"step": 569 |
|
}, |
|
{ |
|
"epoch": 1.1899791231732777, |
|
"grad_norm": 0.04473872009108077, |
|
"learning_rate": 0.00014883529054383817, |
|
"loss": 0.2734, |
|
"step": 570 |
|
}, |
|
{ |
|
"epoch": 1.1920668058455115, |
|
"grad_norm": 0.037460855453091094, |
|
"learning_rate": 0.00014818175196137392, |
|
"loss": 0.2334, |
|
"step": 571 |
|
}, |
|
{ |
|
"epoch": 1.1941544885177453, |
|
"grad_norm": 0.03980728407908227, |
|
"learning_rate": 0.00014752880596387551, |
|
"loss": 0.2295, |
|
"step": 572 |
|
}, |
|
{ |
|
"epoch": 1.196242171189979, |
|
"grad_norm": 0.03562870229391469, |
|
"learning_rate": 0.00014687646001832608, |
|
"loss": 0.1953, |
|
"step": 573 |
|
}, |
|
{ |
|
"epoch": 1.198329853862213, |
|
"grad_norm": 0.0412827193725494, |
|
"learning_rate": 0.00014622472158484654, |
|
"loss": 0.25, |
|
"step": 574 |
|
}, |
|
{ |
|
"epoch": 1.2004175365344467, |
|
"grad_norm": 0.03634424947703343, |
|
"learning_rate": 0.00014557359811661054, |
|
"loss": 0.207, |
|
"step": 575 |
|
}, |
|
{ |
|
"epoch": 1.2025052192066805, |
|
"grad_norm": 0.036441371511458535, |
|
"learning_rate": 0.000144923097059759, |
|
"loss": 0.249, |
|
"step": 576 |
|
}, |
|
{ |
|
"epoch": 1.2045929018789143, |
|
"grad_norm": 0.04200794927737242, |
|
"learning_rate": 0.00014427322585331505, |
|
"loss": 0.3027, |
|
"step": 577 |
|
}, |
|
{ |
|
"epoch": 1.2066805845511483, |
|
"grad_norm": 0.03620959513479551, |
|
"learning_rate": 0.00014362399192909908, |
|
"loss": 0.2324, |
|
"step": 578 |
|
}, |
|
{ |
|
"epoch": 1.2087682672233822, |
|
"grad_norm": 0.03930165573165628, |
|
"learning_rate": 0.0001429754027116433, |
|
"loss": 0.2812, |
|
"step": 579 |
|
}, |
|
{ |
|
"epoch": 1.210855949895616, |
|
"grad_norm": 0.040322537436277076, |
|
"learning_rate": 0.00014232746561810747, |
|
"loss": 0.2773, |
|
"step": 580 |
|
}, |
|
{ |
|
"epoch": 1.2129436325678498, |
|
"grad_norm": 0.03781024017014511, |
|
"learning_rate": 0.00014168018805819376, |
|
"loss": 0.2354, |
|
"step": 581 |
|
}, |
|
{ |
|
"epoch": 1.2150313152400836, |
|
"grad_norm": 0.038595664424750396, |
|
"learning_rate": 0.00014103357743406182, |
|
"loss": 0.2715, |
|
"step": 582 |
|
}, |
|
{ |
|
"epoch": 1.2171189979123174, |
|
"grad_norm": 0.034975311451833, |
|
"learning_rate": 0.00014038764114024456, |
|
"loss": 0.2041, |
|
"step": 583 |
|
}, |
|
{ |
|
"epoch": 1.2192066805845512, |
|
"grad_norm": 0.03825635258528123, |
|
"learning_rate": 0.00013974238656356313, |
|
"loss": 0.2344, |
|
"step": 584 |
|
}, |
|
{ |
|
"epoch": 1.221294363256785, |
|
"grad_norm": 0.03735073552974533, |
|
"learning_rate": 0.0001390978210830429, |
|
"loss": 0.2559, |
|
"step": 585 |
|
}, |
|
{ |
|
"epoch": 1.2233820459290188, |
|
"grad_norm": 0.040138739862162154, |
|
"learning_rate": 0.00013845395206982863, |
|
"loss": 0.2656, |
|
"step": 586 |
|
}, |
|
{ |
|
"epoch": 1.2254697286012526, |
|
"grad_norm": 0.04418658015359039, |
|
"learning_rate": 0.00013781078688710063, |
|
"loss": 0.291, |
|
"step": 587 |
|
}, |
|
{ |
|
"epoch": 1.2275574112734864, |
|
"grad_norm": 0.04239030702531795, |
|
"learning_rate": 0.00013716833288999005, |
|
"loss": 0.2637, |
|
"step": 588 |
|
}, |
|
{ |
|
"epoch": 1.2296450939457202, |
|
"grad_norm": 0.03818928317637622, |
|
"learning_rate": 0.00013652659742549532, |
|
"loss": 0.252, |
|
"step": 589 |
|
}, |
|
{ |
|
"epoch": 1.231732776617954, |
|
"grad_norm": 0.033789428188020935, |
|
"learning_rate": 0.00013588558783239762, |
|
"loss": 0.1826, |
|
"step": 590 |
|
}, |
|
{ |
|
"epoch": 1.2338204592901878, |
|
"grad_norm": 0.03479799715006895, |
|
"learning_rate": 0.00013524531144117736, |
|
"loss": 0.208, |
|
"step": 591 |
|
}, |
|
{ |
|
"epoch": 1.2359081419624216, |
|
"grad_norm": 0.0407139048904091, |
|
"learning_rate": 0.00013460577557393, |
|
"loss": 0.2832, |
|
"step": 592 |
|
}, |
|
{ |
|
"epoch": 1.2379958246346556, |
|
"grad_norm": 0.04211809721321409, |
|
"learning_rate": 0.00013396698754428266, |
|
"loss": 0.2402, |
|
"step": 593 |
|
}, |
|
{ |
|
"epoch": 1.2400835073068894, |
|
"grad_norm": 0.038313063106902565, |
|
"learning_rate": 0.00013332895465731018, |
|
"loss": 0.2539, |
|
"step": 594 |
|
}, |
|
{ |
|
"epoch": 1.2421711899791232, |
|
"grad_norm": 0.034451480337492464, |
|
"learning_rate": 0.00013269168420945178, |
|
"loss": 0.209, |
|
"step": 595 |
|
}, |
|
{ |
|
"epoch": 1.244258872651357, |
|
"grad_norm": 0.042570134603052, |
|
"learning_rate": 0.0001320551834884276, |
|
"loss": 0.2734, |
|
"step": 596 |
|
}, |
|
{ |
|
"epoch": 1.2463465553235908, |
|
"grad_norm": 0.041331289867428476, |
|
"learning_rate": 0.00013141945977315515, |
|
"loss": 0.3066, |
|
"step": 597 |
|
}, |
|
{ |
|
"epoch": 1.2484342379958246, |
|
"grad_norm": 0.03567452834168155, |
|
"learning_rate": 0.00013078452033366635, |
|
"loss": 0.2129, |
|
"step": 598 |
|
}, |
|
{ |
|
"epoch": 1.2505219206680585, |
|
"grad_norm": 0.03702357216857016, |
|
"learning_rate": 0.00013015037243102425, |
|
"loss": 0.2471, |
|
"step": 599 |
|
}, |
|
{ |
|
"epoch": 1.2526096033402923, |
|
"grad_norm": 0.03423169810999704, |
|
"learning_rate": 0.00012951702331724007, |
|
"loss": 0.2236, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 1.254697286012526, |
|
"grad_norm": 0.04865070295457581, |
|
"learning_rate": 0.00012888448023519006, |
|
"loss": 0.3223, |
|
"step": 601 |
|
}, |
|
{ |
|
"epoch": 1.2567849686847599, |
|
"grad_norm": 0.031921521655174964, |
|
"learning_rate": 0.000128252750418533, |
|
"loss": 0.1934, |
|
"step": 602 |
|
}, |
|
{ |
|
"epoch": 1.2588726513569937, |
|
"grad_norm": 0.03207714489804328, |
|
"learning_rate": 0.0001276218410916272, |
|
"loss": 0.1768, |
|
"step": 603 |
|
}, |
|
{ |
|
"epoch": 1.2609603340292277, |
|
"grad_norm": 0.03764702756962339, |
|
"learning_rate": 0.000126991759469448, |
|
"loss": 0.207, |
|
"step": 604 |
|
}, |
|
{ |
|
"epoch": 1.2630480167014615, |
|
"grad_norm": 0.03494076286574369, |
|
"learning_rate": 0.0001263625127575052, |
|
"loss": 0.2129, |
|
"step": 605 |
|
}, |
|
{ |
|
"epoch": 1.2651356993736953, |
|
"grad_norm": 0.03530986655804941, |
|
"learning_rate": 0.00012573410815176083, |
|
"loss": 0.2617, |
|
"step": 606 |
|
}, |
|
{ |
|
"epoch": 1.267223382045929, |
|
"grad_norm": 0.03741233532866471, |
|
"learning_rate": 0.00012510655283854658, |
|
"loss": 0.1914, |
|
"step": 607 |
|
}, |
|
{ |
|
"epoch": 1.269311064718163, |
|
"grad_norm": 0.03133581344814849, |
|
"learning_rate": 0.00012447985399448194, |
|
"loss": 0.1631, |
|
"step": 608 |
|
}, |
|
{ |
|
"epoch": 1.2713987473903967, |
|
"grad_norm": 0.03819672658693745, |
|
"learning_rate": 0.00012385401878639175, |
|
"loss": 0.2891, |
|
"step": 609 |
|
}, |
|
{ |
|
"epoch": 1.2734864300626305, |
|
"grad_norm": 0.04476513417542424, |
|
"learning_rate": 0.00012322905437122468, |
|
"loss": 0.3379, |
|
"step": 610 |
|
}, |
|
{ |
|
"epoch": 1.2755741127348643, |
|
"grad_norm": 0.037096901415977844, |
|
"learning_rate": 0.00012260496789597093, |
|
"loss": 0.2637, |
|
"step": 611 |
|
}, |
|
{ |
|
"epoch": 1.2776617954070981, |
|
"grad_norm": 0.033560968570005754, |
|
"learning_rate": 0.00012198176649758091, |
|
"loss": 0.1768, |
|
"step": 612 |
|
}, |
|
{ |
|
"epoch": 1.279749478079332, |
|
"grad_norm": 0.03915286043749694, |
|
"learning_rate": 0.00012135945730288333, |
|
"loss": 0.2402, |
|
"step": 613 |
|
}, |
|
{ |
|
"epoch": 1.2818371607515657, |
|
"grad_norm": 0.04243780465562693, |
|
"learning_rate": 0.00012073804742850385, |
|
"loss": 0.2266, |
|
"step": 614 |
|
}, |
|
{ |
|
"epoch": 1.2839248434237995, |
|
"grad_norm": 0.035340613835254, |
|
"learning_rate": 0.00012011754398078369, |
|
"loss": 0.2021, |
|
"step": 615 |
|
}, |
|
{ |
|
"epoch": 1.2860125260960333, |
|
"grad_norm": 0.038064762794890517, |
|
"learning_rate": 0.00011949795405569813, |
|
"loss": 0.2373, |
|
"step": 616 |
|
}, |
|
{ |
|
"epoch": 1.2881002087682671, |
|
"grad_norm": 0.03189685893047627, |
|
"learning_rate": 0.00011887928473877581, |
|
"loss": 0.2178, |
|
"step": 617 |
|
}, |
|
{ |
|
"epoch": 1.290187891440501, |
|
"grad_norm": 0.04119691632833074, |
|
"learning_rate": 0.00011826154310501723, |
|
"loss": 0.2832, |
|
"step": 618 |
|
}, |
|
{ |
|
"epoch": 1.2922755741127347, |
|
"grad_norm": 0.035970451962649, |
|
"learning_rate": 0.00011764473621881427, |
|
"loss": 0.2295, |
|
"step": 619 |
|
}, |
|
{ |
|
"epoch": 1.2943632567849686, |
|
"grad_norm": 0.03644862411137844, |
|
"learning_rate": 0.000117028871133869, |
|
"loss": 0.293, |
|
"step": 620 |
|
}, |
|
{ |
|
"epoch": 1.2964509394572026, |
|
"grad_norm": 0.04232916111512329, |
|
"learning_rate": 0.00011641395489311334, |
|
"loss": 0.2432, |
|
"step": 621 |
|
}, |
|
{ |
|
"epoch": 1.2985386221294364, |
|
"grad_norm": 0.03815713518197317, |
|
"learning_rate": 0.00011579999452862834, |
|
"loss": 0.2363, |
|
"step": 622 |
|
}, |
|
{ |
|
"epoch": 1.3006263048016702, |
|
"grad_norm": 0.03522231596610688, |
|
"learning_rate": 0.00011518699706156373, |
|
"loss": 0.2275, |
|
"step": 623 |
|
}, |
|
{ |
|
"epoch": 1.302713987473904, |
|
"grad_norm": 0.03763846715271956, |
|
"learning_rate": 0.00011457496950205784, |
|
"loss": 0.2139, |
|
"step": 624 |
|
}, |
|
{ |
|
"epoch": 1.3048016701461378, |
|
"grad_norm": 0.03937303225722648, |
|
"learning_rate": 0.00011396391884915707, |
|
"loss": 0.2246, |
|
"step": 625 |
|
}, |
|
{ |
|
"epoch": 1.3068893528183716, |
|
"grad_norm": 0.03259155185424707, |
|
"learning_rate": 0.00011335385209073645, |
|
"loss": 0.2148, |
|
"step": 626 |
|
}, |
|
{ |
|
"epoch": 1.3089770354906054, |
|
"grad_norm": 0.036437191147767146, |
|
"learning_rate": 0.00011274477620341906, |
|
"loss": 0.2168, |
|
"step": 627 |
|
}, |
|
{ |
|
"epoch": 1.3110647181628392, |
|
"grad_norm": 0.03676944565811173, |
|
"learning_rate": 0.00011213669815249659, |
|
"loss": 0.21, |
|
"step": 628 |
|
}, |
|
{ |
|
"epoch": 1.313152400835073, |
|
"grad_norm": 0.03659844810941453, |
|
"learning_rate": 0.00011152962489184955, |
|
"loss": 0.2139, |
|
"step": 629 |
|
}, |
|
{ |
|
"epoch": 1.3152400835073068, |
|
"grad_norm": 0.03855990494361526, |
|
"learning_rate": 0.0001109235633638682, |
|
"loss": 0.2373, |
|
"step": 630 |
|
}, |
|
{ |
|
"epoch": 1.3173277661795408, |
|
"grad_norm": 0.034719164079528235, |
|
"learning_rate": 0.00011031852049937237, |
|
"loss": 0.2236, |
|
"step": 631 |
|
}, |
|
{ |
|
"epoch": 1.3194154488517746, |
|
"grad_norm": 0.03650236752350518, |
|
"learning_rate": 0.00010971450321753276, |
|
"loss": 0.2197, |
|
"step": 632 |
|
}, |
|
{ |
|
"epoch": 1.3215031315240084, |
|
"grad_norm": 0.035674494077417296, |
|
"learning_rate": 0.00010911151842579195, |
|
"loss": 0.25, |
|
"step": 633 |
|
}, |
|
{ |
|
"epoch": 1.3235908141962422, |
|
"grad_norm": 0.040631215715819284, |
|
"learning_rate": 0.00010850957301978462, |
|
"loss": 0.2891, |
|
"step": 634 |
|
}, |
|
{ |
|
"epoch": 1.325678496868476, |
|
"grad_norm": 0.03124048788618308, |
|
"learning_rate": 0.00010790867388325951, |
|
"loss": 0.208, |
|
"step": 635 |
|
}, |
|
{ |
|
"epoch": 1.3277661795407099, |
|
"grad_norm": 0.04271594640915922, |
|
"learning_rate": 0.0001073088278880005, |
|
"loss": 0.3184, |
|
"step": 636 |
|
}, |
|
{ |
|
"epoch": 1.3298538622129437, |
|
"grad_norm": 0.03507030432270385, |
|
"learning_rate": 0.0001067100418937477, |
|
"loss": 0.208, |
|
"step": 637 |
|
}, |
|
{ |
|
"epoch": 1.3319415448851775, |
|
"grad_norm": 0.04020676296142397, |
|
"learning_rate": 0.00010611232274811929, |
|
"loss": 0.2988, |
|
"step": 638 |
|
}, |
|
{ |
|
"epoch": 1.3340292275574113, |
|
"grad_norm": 0.034794700042634165, |
|
"learning_rate": 0.00010551567728653311, |
|
"loss": 0.2061, |
|
"step": 639 |
|
}, |
|
{ |
|
"epoch": 1.336116910229645, |
|
"grad_norm": 0.04251059198573156, |
|
"learning_rate": 0.00010492011233212871, |
|
"loss": 0.2578, |
|
"step": 640 |
|
}, |
|
{ |
|
"epoch": 1.3382045929018789, |
|
"grad_norm": 0.031789932037755535, |
|
"learning_rate": 0.0001043256346956889, |
|
"loss": 0.2109, |
|
"step": 641 |
|
}, |
|
{ |
|
"epoch": 1.3402922755741127, |
|
"grad_norm": 0.039155457576247846, |
|
"learning_rate": 0.00010373225117556224, |
|
"loss": 0.2422, |
|
"step": 642 |
|
}, |
|
{ |
|
"epoch": 1.3423799582463465, |
|
"grad_norm": 0.035388665066062634, |
|
"learning_rate": 0.00010313996855758504, |
|
"loss": 0.2002, |
|
"step": 643 |
|
}, |
|
{ |
|
"epoch": 1.3444676409185803, |
|
"grad_norm": 0.0385841616358797, |
|
"learning_rate": 0.00010254879361500407, |
|
"loss": 0.2041, |
|
"step": 644 |
|
}, |
|
{ |
|
"epoch": 1.346555323590814, |
|
"grad_norm": 0.03971686342813934, |
|
"learning_rate": 0.00010195873310839874, |
|
"loss": 0.2256, |
|
"step": 645 |
|
}, |
|
{ |
|
"epoch": 1.348643006263048, |
|
"grad_norm": 0.04152102958645101, |
|
"learning_rate": 0.00010136979378560398, |
|
"loss": 0.2734, |
|
"step": 646 |
|
}, |
|
{ |
|
"epoch": 1.350730688935282, |
|
"grad_norm": 0.036651673265288884, |
|
"learning_rate": 0.00010078198238163299, |
|
"loss": 0.2148, |
|
"step": 647 |
|
}, |
|
{ |
|
"epoch": 1.3528183716075157, |
|
"grad_norm": 0.032824107867213306, |
|
"learning_rate": 0.00010019530561860051, |
|
"loss": 0.208, |
|
"step": 648 |
|
}, |
|
{ |
|
"epoch": 1.3549060542797495, |
|
"grad_norm": 0.03503308809553616, |
|
"learning_rate": 9.96097702056455e-05, |
|
"loss": 0.2041, |
|
"step": 649 |
|
}, |
|
{ |
|
"epoch": 1.3569937369519833, |
|
"grad_norm": 0.0404355664516907, |
|
"learning_rate": 9.902538283885465e-05, |
|
"loss": 0.2021, |
|
"step": 650 |
|
}, |
|
{ |
|
"epoch": 1.3590814196242171, |
|
"grad_norm": 0.047056225295053385, |
|
"learning_rate": 9.844215020118576e-05, |
|
"loss": 0.3125, |
|
"step": 651 |
|
}, |
|
{ |
|
"epoch": 1.361169102296451, |
|
"grad_norm": 0.0328129234031199, |
|
"learning_rate": 9.78600789623916e-05, |
|
"loss": 0.1973, |
|
"step": 652 |
|
}, |
|
{ |
|
"epoch": 1.3632567849686847, |
|
"grad_norm": 0.03175285628250321, |
|
"learning_rate": 9.727917577894297e-05, |
|
"loss": 0.1914, |
|
"step": 653 |
|
}, |
|
{ |
|
"epoch": 1.3653444676409185, |
|
"grad_norm": 0.04520466671433204, |
|
"learning_rate": 9.669944729395316e-05, |
|
"loss": 0.3125, |
|
"step": 654 |
|
}, |
|
{ |
|
"epoch": 1.3674321503131524, |
|
"grad_norm": 0.03901825127204986, |
|
"learning_rate": 9.612090013710195e-05, |
|
"loss": 0.2412, |
|
"step": 655 |
|
}, |
|
{ |
|
"epoch": 1.3695198329853862, |
|
"grad_norm": 0.037647523878022705, |
|
"learning_rate": 9.554354092455949e-05, |
|
"loss": 0.2227, |
|
"step": 656 |
|
}, |
|
{ |
|
"epoch": 1.3716075156576202, |
|
"grad_norm": 0.03796904304768066, |
|
"learning_rate": 9.496737625891076e-05, |
|
"loss": 0.2402, |
|
"step": 657 |
|
}, |
|
{ |
|
"epoch": 1.373695198329854, |
|
"grad_norm": 0.036620990435136655, |
|
"learning_rate": 9.439241272908012e-05, |
|
"loss": 0.21, |
|
"step": 658 |
|
}, |
|
{ |
|
"epoch": 1.3757828810020878, |
|
"grad_norm": 0.03753406391892656, |
|
"learning_rate": 9.381865691025613e-05, |
|
"loss": 0.2471, |
|
"step": 659 |
|
}, |
|
{ |
|
"epoch": 1.3778705636743216, |
|
"grad_norm": 0.03675743788954959, |
|
"learning_rate": 9.324611536381591e-05, |
|
"loss": 0.2256, |
|
"step": 660 |
|
}, |
|
{ |
|
"epoch": 1.3799582463465554, |
|
"grad_norm": 0.04379732725831618, |
|
"learning_rate": 9.267479463725048e-05, |
|
"loss": 0.3184, |
|
"step": 661 |
|
}, |
|
{ |
|
"epoch": 1.3820459290187892, |
|
"grad_norm": 0.04108282244231379, |
|
"learning_rate": 9.210470126408966e-05, |
|
"loss": 0.3145, |
|
"step": 662 |
|
}, |
|
{ |
|
"epoch": 1.384133611691023, |
|
"grad_norm": 0.033523837380523475, |
|
"learning_rate": 9.15358417638277e-05, |
|
"loss": 0.2021, |
|
"step": 663 |
|
}, |
|
{ |
|
"epoch": 1.3862212943632568, |
|
"grad_norm": 0.03998207042720944, |
|
"learning_rate": 9.096822264184825e-05, |
|
"loss": 0.2441, |
|
"step": 664 |
|
}, |
|
{ |
|
"epoch": 1.3883089770354906, |
|
"grad_norm": 0.037635714292856164, |
|
"learning_rate": 9.040185038935029e-05, |
|
"loss": 0.1934, |
|
"step": 665 |
|
}, |
|
{ |
|
"epoch": 1.3903966597077244, |
|
"grad_norm": 0.04063187756459359, |
|
"learning_rate": 8.983673148327369e-05, |
|
"loss": 0.2812, |
|
"step": 666 |
|
}, |
|
{ |
|
"epoch": 1.3924843423799582, |
|
"grad_norm": 0.03568609472936473, |
|
"learning_rate": 8.927287238622555e-05, |
|
"loss": 0.2578, |
|
"step": 667 |
|
}, |
|
{ |
|
"epoch": 1.394572025052192, |
|
"grad_norm": 0.044075873922673904, |
|
"learning_rate": 8.871027954640567e-05, |
|
"loss": 0.2949, |
|
"step": 668 |
|
}, |
|
{ |
|
"epoch": 1.3966597077244258, |
|
"grad_norm": 0.04018365813830774, |
|
"learning_rate": 8.814895939753331e-05, |
|
"loss": 0.3203, |
|
"step": 669 |
|
}, |
|
{ |
|
"epoch": 1.3987473903966596, |
|
"grad_norm": 0.038371138538880856, |
|
"learning_rate": 8.758891835877335e-05, |
|
"loss": 0.2451, |
|
"step": 670 |
|
}, |
|
{ |
|
"epoch": 1.4008350730688934, |
|
"grad_norm": 0.03644115907045684, |
|
"learning_rate": 8.703016283466323e-05, |
|
"loss": 0.2109, |
|
"step": 671 |
|
}, |
|
{ |
|
"epoch": 1.4029227557411272, |
|
"grad_norm": 0.0355184313530291, |
|
"learning_rate": 8.64726992150391e-05, |
|
"loss": 0.2256, |
|
"step": 672 |
|
}, |
|
{ |
|
"epoch": 1.405010438413361, |
|
"grad_norm": 0.032473676946887235, |
|
"learning_rate": 8.59165338749632e-05, |
|
"loss": 0.2266, |
|
"step": 673 |
|
}, |
|
{ |
|
"epoch": 1.407098121085595, |
|
"grad_norm": 0.03821259918901156, |
|
"learning_rate": 8.536167317465114e-05, |
|
"loss": 0.2305, |
|
"step": 674 |
|
}, |
|
{ |
|
"epoch": 1.4091858037578289, |
|
"grad_norm": 0.035842354722894494, |
|
"learning_rate": 8.480812345939855e-05, |
|
"loss": 0.2363, |
|
"step": 675 |
|
}, |
|
{ |
|
"epoch": 1.4112734864300627, |
|
"grad_norm": 0.034936854456220405, |
|
"learning_rate": 8.4255891059509e-05, |
|
"loss": 0.2344, |
|
"step": 676 |
|
}, |
|
{ |
|
"epoch": 1.4133611691022965, |
|
"grad_norm": 0.0327536953633252, |
|
"learning_rate": 8.370498229022133e-05, |
|
"loss": 0.2354, |
|
"step": 677 |
|
}, |
|
{ |
|
"epoch": 1.4154488517745303, |
|
"grad_norm": 0.03332967518468726, |
|
"learning_rate": 8.315540345163783e-05, |
|
"loss": 0.21, |
|
"step": 678 |
|
}, |
|
{ |
|
"epoch": 1.417536534446764, |
|
"grad_norm": 0.039757185578627195, |
|
"learning_rate": 8.26071608286517e-05, |
|
"loss": 0.2451, |
|
"step": 679 |
|
}, |
|
{ |
|
"epoch": 1.4196242171189979, |
|
"grad_norm": 0.033351639676043855, |
|
"learning_rate": 8.206026069087538e-05, |
|
"loss": 0.1963, |
|
"step": 680 |
|
}, |
|
{ |
|
"epoch": 1.4217118997912317, |
|
"grad_norm": 0.036626791510792185, |
|
"learning_rate": 8.151470929256893e-05, |
|
"loss": 0.2471, |
|
"step": 681 |
|
}, |
|
{ |
|
"epoch": 1.4237995824634655, |
|
"grad_norm": 0.03663750757496257, |
|
"learning_rate": 8.097051287256854e-05, |
|
"loss": 0.2559, |
|
"step": 682 |
|
}, |
|
{ |
|
"epoch": 1.4258872651356993, |
|
"grad_norm": 0.03533804105379715, |
|
"learning_rate": 8.042767765421499e-05, |
|
"loss": 0.2334, |
|
"step": 683 |
|
}, |
|
{ |
|
"epoch": 1.4279749478079333, |
|
"grad_norm": 0.03548267765082115, |
|
"learning_rate": 7.988620984528248e-05, |
|
"loss": 0.1992, |
|
"step": 684 |
|
}, |
|
{ |
|
"epoch": 1.4300626304801671, |
|
"grad_norm": 0.03365970412354974, |
|
"learning_rate": 7.934611563790803e-05, |
|
"loss": 0.2129, |
|
"step": 685 |
|
}, |
|
{ |
|
"epoch": 1.432150313152401, |
|
"grad_norm": 0.036222173404344804, |
|
"learning_rate": 7.880740120852012e-05, |
|
"loss": 0.2441, |
|
"step": 686 |
|
}, |
|
{ |
|
"epoch": 1.4342379958246347, |
|
"grad_norm": 0.03545069244831471, |
|
"learning_rate": 7.827007271776843e-05, |
|
"loss": 0.2246, |
|
"step": 687 |
|
}, |
|
{ |
|
"epoch": 1.4363256784968685, |
|
"grad_norm": 0.036141104648819485, |
|
"learning_rate": 7.773413631045314e-05, |
|
"loss": 0.2334, |
|
"step": 688 |
|
}, |
|
{ |
|
"epoch": 1.4384133611691023, |
|
"grad_norm": 0.03552223808507066, |
|
"learning_rate": 7.719959811545512e-05, |
|
"loss": 0.207, |
|
"step": 689 |
|
}, |
|
{ |
|
"epoch": 1.4405010438413361, |
|
"grad_norm": 0.040531388046511914, |
|
"learning_rate": 7.666646424566508e-05, |
|
"loss": 0.25, |
|
"step": 690 |
|
}, |
|
{ |
|
"epoch": 1.44258872651357, |
|
"grad_norm": 0.03642114412895087, |
|
"learning_rate": 7.613474079791432e-05, |
|
"loss": 0.2275, |
|
"step": 691 |
|
}, |
|
{ |
|
"epoch": 1.4446764091858038, |
|
"grad_norm": 0.03334766371093395, |
|
"learning_rate": 7.56044338529049e-05, |
|
"loss": 0.2246, |
|
"step": 692 |
|
}, |
|
{ |
|
"epoch": 1.4467640918580376, |
|
"grad_norm": 0.038435220776801955, |
|
"learning_rate": 7.50755494751398e-05, |
|
"loss": 0.2871, |
|
"step": 693 |
|
}, |
|
{ |
|
"epoch": 1.4488517745302714, |
|
"grad_norm": 0.03639976375126217, |
|
"learning_rate": 7.454809371285381e-05, |
|
"loss": 0.2236, |
|
"step": 694 |
|
}, |
|
{ |
|
"epoch": 1.4509394572025052, |
|
"grad_norm": 0.04139491524391803, |
|
"learning_rate": 7.402207259794428e-05, |
|
"loss": 0.2617, |
|
"step": 695 |
|
}, |
|
{ |
|
"epoch": 1.453027139874739, |
|
"grad_norm": 0.031147759796875986, |
|
"learning_rate": 7.34974921459023e-05, |
|
"loss": 0.1904, |
|
"step": 696 |
|
}, |
|
{ |
|
"epoch": 1.4551148225469728, |
|
"grad_norm": 0.04026433329347746, |
|
"learning_rate": 7.297435835574362e-05, |
|
"loss": 0.2393, |
|
"step": 697 |
|
}, |
|
{ |
|
"epoch": 1.4572025052192066, |
|
"grad_norm": 0.03965005806725716, |
|
"learning_rate": 7.24526772099403e-05, |
|
"loss": 0.2715, |
|
"step": 698 |
|
}, |
|
{ |
|
"epoch": 1.4592901878914404, |
|
"grad_norm": 0.040893406654203844, |
|
"learning_rate": 7.193245467435206e-05, |
|
"loss": 0.2471, |
|
"step": 699 |
|
}, |
|
{ |
|
"epoch": 1.4613778705636742, |
|
"grad_norm": 0.03636456124885245, |
|
"learning_rate": 7.141369669815841e-05, |
|
"loss": 0.2168, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 1.4634655532359082, |
|
"grad_norm": 0.03647208734905981, |
|
"learning_rate": 7.089640921379026e-05, |
|
"loss": 0.2178, |
|
"step": 701 |
|
}, |
|
{ |
|
"epoch": 1.465553235908142, |
|
"grad_norm": 0.0398294556445215, |
|
"learning_rate": 7.038059813686224e-05, |
|
"loss": 0.2695, |
|
"step": 702 |
|
}, |
|
{ |
|
"epoch": 1.4676409185803758, |
|
"grad_norm": 0.03444283718095077, |
|
"learning_rate": 6.986626936610491e-05, |
|
"loss": 0.2109, |
|
"step": 703 |
|
}, |
|
{ |
|
"epoch": 1.4697286012526096, |
|
"grad_norm": 0.036117772890174496, |
|
"learning_rate": 6.935342878329774e-05, |
|
"loss": 0.1855, |
|
"step": 704 |
|
}, |
|
{ |
|
"epoch": 1.4718162839248434, |
|
"grad_norm": 0.03485285291792577, |
|
"learning_rate": 6.884208225320121e-05, |
|
"loss": 0.1953, |
|
"step": 705 |
|
}, |
|
{ |
|
"epoch": 1.4739039665970772, |
|
"grad_norm": 0.03913119058204658, |
|
"learning_rate": 6.833223562349018e-05, |
|
"loss": 0.291, |
|
"step": 706 |
|
}, |
|
{ |
|
"epoch": 1.475991649269311, |
|
"grad_norm": 0.03619007457603297, |
|
"learning_rate": 6.782389472468687e-05, |
|
"loss": 0.2256, |
|
"step": 707 |
|
}, |
|
{ |
|
"epoch": 1.4780793319415448, |
|
"grad_norm": 0.03954136112257214, |
|
"learning_rate": 6.731706537009437e-05, |
|
"loss": 0.2148, |
|
"step": 708 |
|
}, |
|
{ |
|
"epoch": 1.4801670146137786, |
|
"grad_norm": 0.039048253382048426, |
|
"learning_rate": 6.68117533557297e-05, |
|
"loss": 0.2119, |
|
"step": 709 |
|
}, |
|
{ |
|
"epoch": 1.4822546972860124, |
|
"grad_norm": 0.03633979845664376, |
|
"learning_rate": 6.630796446025793e-05, |
|
"loss": 0.207, |
|
"step": 710 |
|
}, |
|
{ |
|
"epoch": 1.4843423799582465, |
|
"grad_norm": 0.04042749139792089, |
|
"learning_rate": 6.580570444492626e-05, |
|
"loss": 0.2617, |
|
"step": 711 |
|
}, |
|
{ |
|
"epoch": 1.4864300626304803, |
|
"grad_norm": 0.04112986991098296, |
|
"learning_rate": 6.530497905349753e-05, |
|
"loss": 0.2578, |
|
"step": 712 |
|
}, |
|
{ |
|
"epoch": 1.488517745302714, |
|
"grad_norm": 0.04014939278609065, |
|
"learning_rate": 6.480579401218502e-05, |
|
"loss": 0.2324, |
|
"step": 713 |
|
}, |
|
{ |
|
"epoch": 1.4906054279749479, |
|
"grad_norm": 0.03662270065596823, |
|
"learning_rate": 6.430815502958674e-05, |
|
"loss": 0.1689, |
|
"step": 714 |
|
}, |
|
{ |
|
"epoch": 1.4926931106471817, |
|
"grad_norm": 0.033001164537146806, |
|
"learning_rate": 6.381206779662039e-05, |
|
"loss": 0.1826, |
|
"step": 715 |
|
}, |
|
{ |
|
"epoch": 1.4947807933194155, |
|
"grad_norm": 0.04093484928652971, |
|
"learning_rate": 6.331753798645796e-05, |
|
"loss": 0.249, |
|
"step": 716 |
|
}, |
|
{ |
|
"epoch": 1.4968684759916493, |
|
"grad_norm": 0.03400770412940897, |
|
"learning_rate": 6.282457125446109e-05, |
|
"loss": 0.1709, |
|
"step": 717 |
|
}, |
|
{ |
|
"epoch": 1.498956158663883, |
|
"grad_norm": 0.03449610130369155, |
|
"learning_rate": 6.233317323811615e-05, |
|
"loss": 0.2168, |
|
"step": 718 |
|
}, |
|
{ |
|
"epoch": 1.501043841336117, |
|
"grad_norm": 0.03943007646918304, |
|
"learning_rate": 6.184334955697028e-05, |
|
"loss": 0.2617, |
|
"step": 719 |
|
}, |
|
{ |
|
"epoch": 1.5031315240083507, |
|
"grad_norm": 0.04046270086854637, |
|
"learning_rate": 6.135510581256647e-05, |
|
"loss": 0.2715, |
|
"step": 720 |
|
}, |
|
{ |
|
"epoch": 1.5052192066805845, |
|
"grad_norm": 0.036042076960770504, |
|
"learning_rate": 6.086844758837991e-05, |
|
"loss": 0.2275, |
|
"step": 721 |
|
}, |
|
{ |
|
"epoch": 1.5073068893528183, |
|
"grad_norm": 0.038478916176100005, |
|
"learning_rate": 6.0383380449754004e-05, |
|
"loss": 0.1924, |
|
"step": 722 |
|
}, |
|
{ |
|
"epoch": 1.5093945720250521, |
|
"grad_norm": 0.0578411545976244, |
|
"learning_rate": 5.98999099438369e-05, |
|
"loss": 0.2852, |
|
"step": 723 |
|
}, |
|
{ |
|
"epoch": 1.511482254697286, |
|
"grad_norm": 0.03739145944754586, |
|
"learning_rate": 5.941804159951778e-05, |
|
"loss": 0.2197, |
|
"step": 724 |
|
}, |
|
{ |
|
"epoch": 1.5135699373695197, |
|
"grad_norm": 0.036832588153257496, |
|
"learning_rate": 5.893778092736382e-05, |
|
"loss": 0.2422, |
|
"step": 725 |
|
}, |
|
{ |
|
"epoch": 1.5156576200417535, |
|
"grad_norm": 0.03671021108564163, |
|
"learning_rate": 5.845913341955711e-05, |
|
"loss": 0.1572, |
|
"step": 726 |
|
}, |
|
{ |
|
"epoch": 1.5177453027139873, |
|
"grad_norm": 0.03573073453794959, |
|
"learning_rate": 5.798210454983184e-05, |
|
"loss": 0.2168, |
|
"step": 727 |
|
}, |
|
{ |
|
"epoch": 1.5198329853862211, |
|
"grad_norm": 0.035959632617295005, |
|
"learning_rate": 5.7506699773411764e-05, |
|
"loss": 0.2451, |
|
"step": 728 |
|
}, |
|
{ |
|
"epoch": 1.5219206680584552, |
|
"grad_norm": 0.038728578675763334, |
|
"learning_rate": 5.703292452694771e-05, |
|
"loss": 0.2139, |
|
"step": 729 |
|
}, |
|
{ |
|
"epoch": 1.524008350730689, |
|
"grad_norm": 0.035308915371426546, |
|
"learning_rate": 5.6560784228455586e-05, |
|
"loss": 0.2021, |
|
"step": 730 |
|
}, |
|
{ |
|
"epoch": 1.5260960334029228, |
|
"grad_norm": 0.039840996287124834, |
|
"learning_rate": 5.609028427725418e-05, |
|
"loss": 0.1953, |
|
"step": 731 |
|
}, |
|
{ |
|
"epoch": 1.5281837160751566, |
|
"grad_norm": 0.04087634507351788, |
|
"learning_rate": 5.562143005390361e-05, |
|
"loss": 0.2617, |
|
"step": 732 |
|
}, |
|
{ |
|
"epoch": 1.5302713987473904, |
|
"grad_norm": 0.03746689779610756, |
|
"learning_rate": 5.5154226920143626e-05, |
|
"loss": 0.2197, |
|
"step": 733 |
|
}, |
|
{ |
|
"epoch": 1.5323590814196242, |
|
"grad_norm": 0.04226091666932121, |
|
"learning_rate": 5.468868021883256e-05, |
|
"loss": 0.2773, |
|
"step": 734 |
|
}, |
|
{ |
|
"epoch": 1.534446764091858, |
|
"grad_norm": 0.03255698564490662, |
|
"learning_rate": 5.422479527388591e-05, |
|
"loss": 0.1816, |
|
"step": 735 |
|
}, |
|
{ |
|
"epoch": 1.536534446764092, |
|
"grad_norm": 0.03534998758506692, |
|
"learning_rate": 5.376257739021564e-05, |
|
"loss": 0.2158, |
|
"step": 736 |
|
}, |
|
{ |
|
"epoch": 1.5386221294363258, |
|
"grad_norm": 0.04286917750609741, |
|
"learning_rate": 5.330203185366942e-05, |
|
"loss": 0.2559, |
|
"step": 737 |
|
}, |
|
{ |
|
"epoch": 1.5407098121085596, |
|
"grad_norm": 0.03987124501813775, |
|
"learning_rate": 5.284316393097042e-05, |
|
"loss": 0.2197, |
|
"step": 738 |
|
}, |
|
{ |
|
"epoch": 1.5427974947807934, |
|
"grad_norm": 0.0357279862009252, |
|
"learning_rate": 5.2385978869656705e-05, |
|
"loss": 0.209, |
|
"step": 739 |
|
}, |
|
{ |
|
"epoch": 1.5448851774530272, |
|
"grad_norm": 0.03917856556644758, |
|
"learning_rate": 5.193048189802143e-05, |
|
"loss": 0.2402, |
|
"step": 740 |
|
}, |
|
{ |
|
"epoch": 1.546972860125261, |
|
"grad_norm": 0.05340411374410496, |
|
"learning_rate": 5.147667822505318e-05, |
|
"loss": 0.2695, |
|
"step": 741 |
|
}, |
|
{ |
|
"epoch": 1.5490605427974948, |
|
"grad_norm": 0.041333285407966595, |
|
"learning_rate": 5.1024573040376115e-05, |
|
"loss": 0.2969, |
|
"step": 742 |
|
}, |
|
{ |
|
"epoch": 1.5511482254697286, |
|
"grad_norm": 0.039402130485741114, |
|
"learning_rate": 5.057417151419077e-05, |
|
"loss": 0.2402, |
|
"step": 743 |
|
}, |
|
{ |
|
"epoch": 1.5532359081419624, |
|
"grad_norm": 0.035671385682246794, |
|
"learning_rate": 5.012547879721494e-05, |
|
"loss": 0.2158, |
|
"step": 744 |
|
}, |
|
{ |
|
"epoch": 1.5553235908141962, |
|
"grad_norm": 0.038417929606667435, |
|
"learning_rate": 4.967850002062491e-05, |
|
"loss": 0.2236, |
|
"step": 745 |
|
}, |
|
{ |
|
"epoch": 1.55741127348643, |
|
"grad_norm": 0.03949905474465816, |
|
"learning_rate": 4.923324029599632e-05, |
|
"loss": 0.2715, |
|
"step": 746 |
|
}, |
|
{ |
|
"epoch": 1.5594989561586639, |
|
"grad_norm": 0.0352045032702957, |
|
"learning_rate": 4.878970471524622e-05, |
|
"loss": 0.2275, |
|
"step": 747 |
|
}, |
|
{ |
|
"epoch": 1.5615866388308977, |
|
"grad_norm": 0.036303248719977955, |
|
"learning_rate": 4.834789835057465e-05, |
|
"loss": 0.2021, |
|
"step": 748 |
|
}, |
|
{ |
|
"epoch": 1.5636743215031315, |
|
"grad_norm": 0.03443800350271572, |
|
"learning_rate": 4.790782625440655e-05, |
|
"loss": 0.1992, |
|
"step": 749 |
|
}, |
|
{ |
|
"epoch": 1.5657620041753653, |
|
"grad_norm": 0.04231633067945213, |
|
"learning_rate": 4.7469493459334066e-05, |
|
"loss": 0.2617, |
|
"step": 750 |
|
}, |
|
{ |
|
"epoch": 1.567849686847599, |
|
"grad_norm": 0.037830454141616246, |
|
"learning_rate": 4.70329049780589e-05, |
|
"loss": 0.2217, |
|
"step": 751 |
|
}, |
|
{ |
|
"epoch": 1.5699373695198329, |
|
"grad_norm": 0.037662635732679015, |
|
"learning_rate": 4.65980658033353e-05, |
|
"loss": 0.2188, |
|
"step": 752 |
|
}, |
|
{ |
|
"epoch": 1.5720250521920667, |
|
"grad_norm": 0.04067293838178101, |
|
"learning_rate": 4.616498090791248e-05, |
|
"loss": 0.293, |
|
"step": 753 |
|
}, |
|
{ |
|
"epoch": 1.5741127348643005, |
|
"grad_norm": 0.03901581955251283, |
|
"learning_rate": 4.5733655244478146e-05, |
|
"loss": 0.2266, |
|
"step": 754 |
|
}, |
|
{ |
|
"epoch": 1.5762004175365343, |
|
"grad_norm": 0.03554948290229258, |
|
"learning_rate": 4.5304093745601605e-05, |
|
"loss": 0.1943, |
|
"step": 755 |
|
}, |
|
{ |
|
"epoch": 1.5782881002087683, |
|
"grad_norm": 0.0405280460733433, |
|
"learning_rate": 4.487630132367764e-05, |
|
"loss": 0.2617, |
|
"step": 756 |
|
}, |
|
{ |
|
"epoch": 1.5803757828810021, |
|
"grad_norm": 0.03823926366289705, |
|
"learning_rate": 4.445028287087003e-05, |
|
"loss": 0.252, |
|
"step": 757 |
|
}, |
|
{ |
|
"epoch": 1.582463465553236, |
|
"grad_norm": 0.03721532941881517, |
|
"learning_rate": 4.402604325905575e-05, |
|
"loss": 0.2129, |
|
"step": 758 |
|
}, |
|
{ |
|
"epoch": 1.5845511482254697, |
|
"grad_norm": 0.029782665453401677, |
|
"learning_rate": 4.360358733976919e-05, |
|
"loss": 0.1826, |
|
"step": 759 |
|
}, |
|
{ |
|
"epoch": 1.5866388308977035, |
|
"grad_norm": 0.03989589641175334, |
|
"learning_rate": 4.3182919944146936e-05, |
|
"loss": 0.2734, |
|
"step": 760 |
|
}, |
|
{ |
|
"epoch": 1.5887265135699373, |
|
"grad_norm": 0.04383942244265336, |
|
"learning_rate": 4.2764045882872086e-05, |
|
"loss": 0.3008, |
|
"step": 761 |
|
}, |
|
{ |
|
"epoch": 1.5908141962421714, |
|
"grad_norm": 0.034818289814213685, |
|
"learning_rate": 4.2346969946119574e-05, |
|
"loss": 0.2188, |
|
"step": 762 |
|
}, |
|
{ |
|
"epoch": 1.5929018789144052, |
|
"grad_norm": 0.03770186022731002, |
|
"learning_rate": 4.1931696903501204e-05, |
|
"loss": 0.2314, |
|
"step": 763 |
|
}, |
|
{ |
|
"epoch": 1.594989561586639, |
|
"grad_norm": 0.03973578816213652, |
|
"learning_rate": 4.151823150401124e-05, |
|
"loss": 0.2402, |
|
"step": 764 |
|
}, |
|
{ |
|
"epoch": 1.5970772442588728, |
|
"grad_norm": 0.03236007667945323, |
|
"learning_rate": 4.1106578475972035e-05, |
|
"loss": 0.1924, |
|
"step": 765 |
|
}, |
|
{ |
|
"epoch": 1.5991649269311066, |
|
"grad_norm": 0.03764508247239107, |
|
"learning_rate": 4.069674252697988e-05, |
|
"loss": 0.1748, |
|
"step": 766 |
|
}, |
|
{ |
|
"epoch": 1.6012526096033404, |
|
"grad_norm": 0.03731281581422325, |
|
"learning_rate": 4.02887283438514e-05, |
|
"loss": 0.2559, |
|
"step": 767 |
|
}, |
|
{ |
|
"epoch": 1.6033402922755742, |
|
"grad_norm": 0.0372624393342072, |
|
"learning_rate": 3.988254059256971e-05, |
|
"loss": 0.2188, |
|
"step": 768 |
|
}, |
|
{ |
|
"epoch": 1.605427974947808, |
|
"grad_norm": 0.039361210964991795, |
|
"learning_rate": 3.9478183918231106e-05, |
|
"loss": 0.2578, |
|
"step": 769 |
|
}, |
|
{ |
|
"epoch": 1.6075156576200418, |
|
"grad_norm": 0.03750381742110697, |
|
"learning_rate": 3.907566294499201e-05, |
|
"loss": 0.1982, |
|
"step": 770 |
|
}, |
|
{ |
|
"epoch": 1.6096033402922756, |
|
"grad_norm": 0.03771991482783373, |
|
"learning_rate": 3.8674982276016205e-05, |
|
"loss": 0.2412, |
|
"step": 771 |
|
}, |
|
{ |
|
"epoch": 1.6116910229645094, |
|
"grad_norm": 0.04091075279365844, |
|
"learning_rate": 3.82761464934219e-05, |
|
"loss": 0.2793, |
|
"step": 772 |
|
}, |
|
{ |
|
"epoch": 1.6137787056367432, |
|
"grad_norm": 0.03859340968976196, |
|
"learning_rate": 3.787916015822954e-05, |
|
"loss": 0.2471, |
|
"step": 773 |
|
}, |
|
{ |
|
"epoch": 1.615866388308977, |
|
"grad_norm": 0.03200544736921389, |
|
"learning_rate": 3.748402781030955e-05, |
|
"loss": 0.1826, |
|
"step": 774 |
|
}, |
|
{ |
|
"epoch": 1.6179540709812108, |
|
"grad_norm": 0.0340269199931104, |
|
"learning_rate": 3.709075396833057e-05, |
|
"loss": 0.1855, |
|
"step": 775 |
|
}, |
|
{ |
|
"epoch": 1.6200417536534446, |
|
"grad_norm": 0.038072978162411425, |
|
"learning_rate": 3.66993431297076e-05, |
|
"loss": 0.2432, |
|
"step": 776 |
|
}, |
|
{ |
|
"epoch": 1.6221294363256784, |
|
"grad_norm": 0.04102889623967079, |
|
"learning_rate": 3.63097997705506e-05, |
|
"loss": 0.3145, |
|
"step": 777 |
|
}, |
|
{ |
|
"epoch": 1.6242171189979122, |
|
"grad_norm": 0.04007477882751308, |
|
"learning_rate": 3.592212834561337e-05, |
|
"loss": 0.2432, |
|
"step": 778 |
|
}, |
|
{ |
|
"epoch": 1.626304801670146, |
|
"grad_norm": 0.03819036990721501, |
|
"learning_rate": 3.553633328824266e-05, |
|
"loss": 0.1973, |
|
"step": 779 |
|
}, |
|
{ |
|
"epoch": 1.6283924843423798, |
|
"grad_norm": 0.038350798236554316, |
|
"learning_rate": 3.515241901032731e-05, |
|
"loss": 0.2676, |
|
"step": 780 |
|
}, |
|
{ |
|
"epoch": 1.6304801670146136, |
|
"grad_norm": 0.03548951603993466, |
|
"learning_rate": 3.477038990224786e-05, |
|
"loss": 0.2002, |
|
"step": 781 |
|
}, |
|
{ |
|
"epoch": 1.6325678496868476, |
|
"grad_norm": 0.034615335672358714, |
|
"learning_rate": 3.439025033282639e-05, |
|
"loss": 0.1963, |
|
"step": 782 |
|
}, |
|
{ |
|
"epoch": 1.6346555323590815, |
|
"grad_norm": 0.04345337713073096, |
|
"learning_rate": 3.401200464927654e-05, |
|
"loss": 0.2578, |
|
"step": 783 |
|
}, |
|
{ |
|
"epoch": 1.6367432150313153, |
|
"grad_norm": 0.03791537054424221, |
|
"learning_rate": 3.363565717715373e-05, |
|
"loss": 0.2227, |
|
"step": 784 |
|
}, |
|
{ |
|
"epoch": 1.638830897703549, |
|
"grad_norm": 0.04113195871715081, |
|
"learning_rate": 3.326121222030578e-05, |
|
"loss": 0.2012, |
|
"step": 785 |
|
}, |
|
{ |
|
"epoch": 1.6409185803757829, |
|
"grad_norm": 0.041884239195247946, |
|
"learning_rate": 3.288867406082372e-05, |
|
"loss": 0.2236, |
|
"step": 786 |
|
}, |
|
{ |
|
"epoch": 1.6430062630480167, |
|
"grad_norm": 0.03652594444269533, |
|
"learning_rate": 3.251804695899267e-05, |
|
"loss": 0.2139, |
|
"step": 787 |
|
}, |
|
{ |
|
"epoch": 1.6450939457202505, |
|
"grad_norm": 0.05643896674418451, |
|
"learning_rate": 3.214933515324323e-05, |
|
"loss": 0.2422, |
|
"step": 788 |
|
}, |
|
{ |
|
"epoch": 1.6471816283924845, |
|
"grad_norm": 0.04098991740754866, |
|
"learning_rate": 3.178254286010296e-05, |
|
"loss": 0.2422, |
|
"step": 789 |
|
}, |
|
{ |
|
"epoch": 1.6492693110647183, |
|
"grad_norm": 0.04098022159134268, |
|
"learning_rate": 3.1417674274148276e-05, |
|
"loss": 0.2402, |
|
"step": 790 |
|
}, |
|
{ |
|
"epoch": 1.651356993736952, |
|
"grad_norm": 0.037087942509587524, |
|
"learning_rate": 3.105473356795634e-05, |
|
"loss": 0.2295, |
|
"step": 791 |
|
}, |
|
{ |
|
"epoch": 1.653444676409186, |
|
"grad_norm": 0.03775161487801411, |
|
"learning_rate": 3.0693724892057396e-05, |
|
"loss": 0.207, |
|
"step": 792 |
|
}, |
|
{ |
|
"epoch": 1.6555323590814197, |
|
"grad_norm": 0.04001811834655189, |
|
"learning_rate": 3.0334652374887263e-05, |
|
"loss": 0.2773, |
|
"step": 793 |
|
}, |
|
{ |
|
"epoch": 1.6576200417536535, |
|
"grad_norm": 0.03971800362165568, |
|
"learning_rate": 2.997752012274031e-05, |
|
"loss": 0.2139, |
|
"step": 794 |
|
}, |
|
{ |
|
"epoch": 1.6597077244258873, |
|
"grad_norm": 0.03971183328914648, |
|
"learning_rate": 2.9622332219722193e-05, |
|
"loss": 0.2422, |
|
"step": 795 |
|
}, |
|
{ |
|
"epoch": 1.6617954070981211, |
|
"grad_norm": 0.03937647373747184, |
|
"learning_rate": 2.9269092727703375e-05, |
|
"loss": 0.252, |
|
"step": 796 |
|
}, |
|
{ |
|
"epoch": 1.663883089770355, |
|
"grad_norm": 0.03693074285981622, |
|
"learning_rate": 2.891780568627267e-05, |
|
"loss": 0.1992, |
|
"step": 797 |
|
}, |
|
{ |
|
"epoch": 1.6659707724425887, |
|
"grad_norm": 0.03819446049329271, |
|
"learning_rate": 2.8568475112690918e-05, |
|
"loss": 0.2383, |
|
"step": 798 |
|
}, |
|
{ |
|
"epoch": 1.6680584551148225, |
|
"grad_norm": 0.03385681858741079, |
|
"learning_rate": 2.8221105001845095e-05, |
|
"loss": 0.1914, |
|
"step": 799 |
|
}, |
|
{ |
|
"epoch": 1.6701461377870563, |
|
"grad_norm": 0.03481284448136041, |
|
"learning_rate": 2.7875699326202663e-05, |
|
"loss": 0.1943, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 1.6722338204592901, |
|
"grad_norm": 0.038887635380198286, |
|
"learning_rate": 2.7532262035766288e-05, |
|
"loss": 0.2236, |
|
"step": 801 |
|
}, |
|
{ |
|
"epoch": 1.674321503131524, |
|
"grad_norm": 0.039268882753983984, |
|
"learning_rate": 2.7190797058028207e-05, |
|
"loss": 0.2402, |
|
"step": 802 |
|
}, |
|
{ |
|
"epoch": 1.6764091858037578, |
|
"grad_norm": 0.03561285649555011, |
|
"learning_rate": 2.685130829792577e-05, |
|
"loss": 0.2041, |
|
"step": 803 |
|
}, |
|
{ |
|
"epoch": 1.6784968684759916, |
|
"grad_norm": 0.03548955540319579, |
|
"learning_rate": 2.6513799637796745e-05, |
|
"loss": 0.2188, |
|
"step": 804 |
|
}, |
|
{ |
|
"epoch": 1.6805845511482254, |
|
"grad_norm": 0.03987654406934741, |
|
"learning_rate": 2.61782749373346e-05, |
|
"loss": 0.2334, |
|
"step": 805 |
|
}, |
|
{ |
|
"epoch": 1.6826722338204592, |
|
"grad_norm": 0.03591235670312746, |
|
"learning_rate": 2.584473803354468e-05, |
|
"loss": 0.2217, |
|
"step": 806 |
|
}, |
|
{ |
|
"epoch": 1.684759916492693, |
|
"grad_norm": 0.03563754130765487, |
|
"learning_rate": 2.5513192740700144e-05, |
|
"loss": 0.2061, |
|
"step": 807 |
|
}, |
|
{ |
|
"epoch": 1.6868475991649268, |
|
"grad_norm": 0.03453991302292082, |
|
"learning_rate": 2.5183642850298505e-05, |
|
"loss": 0.207, |
|
"step": 808 |
|
}, |
|
{ |
|
"epoch": 1.6889352818371608, |
|
"grad_norm": 0.040646656045912735, |
|
"learning_rate": 2.4856092131018137e-05, |
|
"loss": 0.293, |
|
"step": 809 |
|
}, |
|
{ |
|
"epoch": 1.6910229645093946, |
|
"grad_norm": 0.03639422717002704, |
|
"learning_rate": 2.4530544328675186e-05, |
|
"loss": 0.2236, |
|
"step": 810 |
|
}, |
|
{ |
|
"epoch": 1.6931106471816284, |
|
"grad_norm": 0.04124184507078617, |
|
"learning_rate": 2.4207003166180765e-05, |
|
"loss": 0.2344, |
|
"step": 811 |
|
}, |
|
{ |
|
"epoch": 1.6951983298538622, |
|
"grad_norm": 0.04821861121500109, |
|
"learning_rate": 2.3885472343498515e-05, |
|
"loss": 0.2227, |
|
"step": 812 |
|
}, |
|
{ |
|
"epoch": 1.697286012526096, |
|
"grad_norm": 0.036261622126831795, |
|
"learning_rate": 2.3565955537602014e-05, |
|
"loss": 0.2266, |
|
"step": 813 |
|
}, |
|
{ |
|
"epoch": 1.6993736951983298, |
|
"grad_norm": 0.03833360109114772, |
|
"learning_rate": 2.3248456402432918e-05, |
|
"loss": 0.2598, |
|
"step": 814 |
|
}, |
|
{ |
|
"epoch": 1.7014613778705638, |
|
"grad_norm": 0.03811210564749943, |
|
"learning_rate": 2.293297856885912e-05, |
|
"loss": 0.2197, |
|
"step": 815 |
|
}, |
|
{ |
|
"epoch": 1.7035490605427976, |
|
"grad_norm": 0.03440881800931742, |
|
"learning_rate": 2.261952564463332e-05, |
|
"loss": 0.1865, |
|
"step": 816 |
|
}, |
|
{ |
|
"epoch": 1.7056367432150314, |
|
"grad_norm": 0.03784666030148996, |
|
"learning_rate": 2.2308101214351628e-05, |
|
"loss": 0.2324, |
|
"step": 817 |
|
}, |
|
{ |
|
"epoch": 1.7077244258872653, |
|
"grad_norm": 0.03706179096561306, |
|
"learning_rate": 2.1998708839412597e-05, |
|
"loss": 0.21, |
|
"step": 818 |
|
}, |
|
{ |
|
"epoch": 1.709812108559499, |
|
"grad_norm": 0.036230331032483534, |
|
"learning_rate": 2.1691352057976565e-05, |
|
"loss": 0.2471, |
|
"step": 819 |
|
}, |
|
{ |
|
"epoch": 1.7118997912317329, |
|
"grad_norm": 0.04090973690460557, |
|
"learning_rate": 2.138603438492517e-05, |
|
"loss": 0.2637, |
|
"step": 820 |
|
}, |
|
{ |
|
"epoch": 1.7139874739039667, |
|
"grad_norm": 0.051662701812649245, |
|
"learning_rate": 2.108275931182111e-05, |
|
"loss": 0.1953, |
|
"step": 821 |
|
}, |
|
{ |
|
"epoch": 1.7160751565762005, |
|
"grad_norm": 0.038240986750146325, |
|
"learning_rate": 2.0781530306868246e-05, |
|
"loss": 0.2393, |
|
"step": 822 |
|
}, |
|
{ |
|
"epoch": 1.7181628392484343, |
|
"grad_norm": 0.04106815130974131, |
|
"learning_rate": 2.048235081487202e-05, |
|
"loss": 0.2637, |
|
"step": 823 |
|
}, |
|
{ |
|
"epoch": 1.720250521920668, |
|
"grad_norm": 0.03673610429332063, |
|
"learning_rate": 2.0185224257199908e-05, |
|
"loss": 0.2207, |
|
"step": 824 |
|
}, |
|
{ |
|
"epoch": 1.7223382045929019, |
|
"grad_norm": 0.04334868227291063, |
|
"learning_rate": 1.9890154031742325e-05, |
|
"loss": 0.3223, |
|
"step": 825 |
|
}, |
|
{ |
|
"epoch": 1.7244258872651357, |
|
"grad_norm": 0.04017653068637208, |
|
"learning_rate": 1.9597143512873872e-05, |
|
"loss": 0.2656, |
|
"step": 826 |
|
}, |
|
{ |
|
"epoch": 1.7265135699373695, |
|
"grad_norm": 0.04143141277717955, |
|
"learning_rate": 1.9306196051414728e-05, |
|
"loss": 0.252, |
|
"step": 827 |
|
}, |
|
{ |
|
"epoch": 1.7286012526096033, |
|
"grad_norm": 0.030689616701832047, |
|
"learning_rate": 1.9017314974592203e-05, |
|
"loss": 0.1982, |
|
"step": 828 |
|
}, |
|
{ |
|
"epoch": 1.730688935281837, |
|
"grad_norm": 0.04149364456774439, |
|
"learning_rate": 1.873050358600279e-05, |
|
"loss": 0.2451, |
|
"step": 829 |
|
}, |
|
{ |
|
"epoch": 1.732776617954071, |
|
"grad_norm": 0.037822515436121175, |
|
"learning_rate": 1.8445765165574324e-05, |
|
"loss": 0.2246, |
|
"step": 830 |
|
}, |
|
{ |
|
"epoch": 1.7348643006263047, |
|
"grad_norm": 0.036300250288018185, |
|
"learning_rate": 1.8163102969528702e-05, |
|
"loss": 0.1709, |
|
"step": 831 |
|
}, |
|
{ |
|
"epoch": 1.7369519832985385, |
|
"grad_norm": 0.03813579853432033, |
|
"learning_rate": 1.7882520230344225e-05, |
|
"loss": 0.1982, |
|
"step": 832 |
|
}, |
|
{ |
|
"epoch": 1.7390396659707723, |
|
"grad_norm": 0.03490543301641405, |
|
"learning_rate": 1.7604020156719026e-05, |
|
"loss": 0.2256, |
|
"step": 833 |
|
}, |
|
{ |
|
"epoch": 1.7411273486430061, |
|
"grad_norm": 0.03767801651422133, |
|
"learning_rate": 1.7327605933534087e-05, |
|
"loss": 0.248, |
|
"step": 834 |
|
}, |
|
{ |
|
"epoch": 1.7432150313152401, |
|
"grad_norm": 0.03559965744869205, |
|
"learning_rate": 1.705328072181711e-05, |
|
"loss": 0.1963, |
|
"step": 835 |
|
}, |
|
{ |
|
"epoch": 1.745302713987474, |
|
"grad_norm": 0.03620103246900227, |
|
"learning_rate": 1.6781047658706094e-05, |
|
"loss": 0.2363, |
|
"step": 836 |
|
}, |
|
{ |
|
"epoch": 1.7473903966597077, |
|
"grad_norm": 0.03363208108698153, |
|
"learning_rate": 1.6510909857413593e-05, |
|
"loss": 0.208, |
|
"step": 837 |
|
}, |
|
{ |
|
"epoch": 1.7494780793319415, |
|
"grad_norm": 0.038864402335742426, |
|
"learning_rate": 1.624287040719108e-05, |
|
"loss": 0.2217, |
|
"step": 838 |
|
}, |
|
{ |
|
"epoch": 1.7515657620041754, |
|
"grad_norm": 0.03604996210348334, |
|
"learning_rate": 1.597693237329363e-05, |
|
"loss": 0.2363, |
|
"step": 839 |
|
}, |
|
{ |
|
"epoch": 1.7536534446764092, |
|
"grad_norm": 0.03870805026122142, |
|
"learning_rate": 1.5713098796944913e-05, |
|
"loss": 0.2363, |
|
"step": 840 |
|
}, |
|
{ |
|
"epoch": 1.755741127348643, |
|
"grad_norm": 0.037967280022889786, |
|
"learning_rate": 1.545137269530228e-05, |
|
"loss": 0.2031, |
|
"step": 841 |
|
}, |
|
{ |
|
"epoch": 1.757828810020877, |
|
"grad_norm": 0.032100601105533956, |
|
"learning_rate": 1.5191757061422484e-05, |
|
"loss": 0.1963, |
|
"step": 842 |
|
}, |
|
{ |
|
"epoch": 1.7599164926931108, |
|
"grad_norm": 0.03677180115256404, |
|
"learning_rate": 1.493425486422717e-05, |
|
"loss": 0.2324, |
|
"step": 843 |
|
}, |
|
{ |
|
"epoch": 1.7620041753653446, |
|
"grad_norm": 0.04179333438984321, |
|
"learning_rate": 1.467886904846918e-05, |
|
"loss": 0.2197, |
|
"step": 844 |
|
}, |
|
{ |
|
"epoch": 1.7640918580375784, |
|
"grad_norm": 0.03330069993588093, |
|
"learning_rate": 1.4425602534698645e-05, |
|
"loss": 0.1992, |
|
"step": 845 |
|
}, |
|
{ |
|
"epoch": 1.7661795407098122, |
|
"grad_norm": 0.03921593097418753, |
|
"learning_rate": 1.4174458219229868e-05, |
|
"loss": 0.2734, |
|
"step": 846 |
|
}, |
|
{ |
|
"epoch": 1.768267223382046, |
|
"grad_norm": 0.0384869660070795, |
|
"learning_rate": 1.3925438974107918e-05, |
|
"loss": 0.2285, |
|
"step": 847 |
|
}, |
|
{ |
|
"epoch": 1.7703549060542798, |
|
"grad_norm": 0.03235612278723638, |
|
"learning_rate": 1.3678547647075946e-05, |
|
"loss": 0.2021, |
|
"step": 848 |
|
}, |
|
{ |
|
"epoch": 1.7724425887265136, |
|
"grad_norm": 0.034557637083813396, |
|
"learning_rate": 1.3433787061542525e-05, |
|
"loss": 0.1982, |
|
"step": 849 |
|
}, |
|
{ |
|
"epoch": 1.7745302713987474, |
|
"grad_norm": 0.036691512115311166, |
|
"learning_rate": 1.3191160016549564e-05, |
|
"loss": 0.2695, |
|
"step": 850 |
|
}, |
|
{ |
|
"epoch": 1.7766179540709812, |
|
"grad_norm": 0.036890816812044876, |
|
"learning_rate": 1.2950669286740024e-05, |
|
"loss": 0.2119, |
|
"step": 851 |
|
}, |
|
{ |
|
"epoch": 1.778705636743215, |
|
"grad_norm": 0.03801656481599902, |
|
"learning_rate": 1.271231762232632e-05, |
|
"loss": 0.2354, |
|
"step": 852 |
|
}, |
|
{ |
|
"epoch": 1.7807933194154488, |
|
"grad_norm": 0.0431132757880389, |
|
"learning_rate": 1.2476107749058986e-05, |
|
"loss": 0.2305, |
|
"step": 853 |
|
}, |
|
{ |
|
"epoch": 1.7828810020876826, |
|
"grad_norm": 0.034619924337342727, |
|
"learning_rate": 1.2242042368195218e-05, |
|
"loss": 0.1914, |
|
"step": 854 |
|
}, |
|
{ |
|
"epoch": 1.7849686847599164, |
|
"grad_norm": 0.034941322602160714, |
|
"learning_rate": 1.2010124156468294e-05, |
|
"loss": 0.2051, |
|
"step": 855 |
|
}, |
|
{ |
|
"epoch": 1.7870563674321502, |
|
"grad_norm": 0.03577282594283845, |
|
"learning_rate": 1.1780355766056694e-05, |
|
"loss": 0.2334, |
|
"step": 856 |
|
}, |
|
{ |
|
"epoch": 1.789144050104384, |
|
"grad_norm": 0.03923352027261708, |
|
"learning_rate": 1.1552739824554026e-05, |
|
"loss": 0.2217, |
|
"step": 857 |
|
}, |
|
{ |
|
"epoch": 1.7912317327766178, |
|
"grad_norm": 0.03464671275803716, |
|
"learning_rate": 1.1327278934938723e-05, |
|
"loss": 0.1992, |
|
"step": 858 |
|
}, |
|
{ |
|
"epoch": 1.7933194154488517, |
|
"grad_norm": 0.04157357259303964, |
|
"learning_rate": 1.1103975675544443e-05, |
|
"loss": 0.2461, |
|
"step": 859 |
|
}, |
|
{ |
|
"epoch": 1.7954070981210855, |
|
"grad_norm": 0.040227594429237123, |
|
"learning_rate": 1.0882832600030624e-05, |
|
"loss": 0.2441, |
|
"step": 860 |
|
}, |
|
{ |
|
"epoch": 1.7974947807933193, |
|
"grad_norm": 0.03488250123821703, |
|
"learning_rate": 1.0663852237353067e-05, |
|
"loss": 0.1963, |
|
"step": 861 |
|
}, |
|
{ |
|
"epoch": 1.7995824634655533, |
|
"grad_norm": 0.03979475512791435, |
|
"learning_rate": 1.0447037091735223e-05, |
|
"loss": 0.2471, |
|
"step": 862 |
|
}, |
|
{ |
|
"epoch": 1.801670146137787, |
|
"grad_norm": 0.03997223499064563, |
|
"learning_rate": 1.0232389642639395e-05, |
|
"loss": 0.2471, |
|
"step": 863 |
|
}, |
|
{ |
|
"epoch": 1.803757828810021, |
|
"grad_norm": 0.04038899286733726, |
|
"learning_rate": 1.0019912344738625e-05, |
|
"loss": 0.2598, |
|
"step": 864 |
|
}, |
|
{ |
|
"epoch": 1.8058455114822547, |
|
"grad_norm": 0.03748451896837008, |
|
"learning_rate": 9.809607627888296e-06, |
|
"loss": 0.2002, |
|
"step": 865 |
|
}, |
|
{ |
|
"epoch": 1.8079331941544885, |
|
"grad_norm": 0.03427034514456546, |
|
"learning_rate": 9.601477897098576e-06, |
|
"loss": 0.1729, |
|
"step": 866 |
|
}, |
|
{ |
|
"epoch": 1.8100208768267223, |
|
"grad_norm": 0.04471055633543089, |
|
"learning_rate": 9.395525532506866e-06, |
|
"loss": 0.2256, |
|
"step": 867 |
|
}, |
|
{ |
|
"epoch": 1.812108559498956, |
|
"grad_norm": 0.03475782341698621, |
|
"learning_rate": 9.191752889350547e-06, |
|
"loss": 0.2207, |
|
"step": 868 |
|
}, |
|
{ |
|
"epoch": 1.8141962421711901, |
|
"grad_norm": 0.03828798248246293, |
|
"learning_rate": 8.990162297940097e-06, |
|
"loss": 0.2109, |
|
"step": 869 |
|
}, |
|
{ |
|
"epoch": 1.816283924843424, |
|
"grad_norm": 0.04180601763110725, |
|
"learning_rate": 8.79075606363231e-06, |
|
"loss": 0.2676, |
|
"step": 870 |
|
}, |
|
{ |
|
"epoch": 1.8183716075156577, |
|
"grad_norm": 0.03646735664376503, |
|
"learning_rate": 8.593536466804142e-06, |
|
"loss": 0.207, |
|
"step": 871 |
|
}, |
|
{ |
|
"epoch": 1.8204592901878915, |
|
"grad_norm": 0.03218572550634406, |
|
"learning_rate": 8.398505762826503e-06, |
|
"loss": 0.1797, |
|
"step": 872 |
|
}, |
|
{ |
|
"epoch": 1.8225469728601253, |
|
"grad_norm": 0.03982709976184709, |
|
"learning_rate": 8.205666182038418e-06, |
|
"loss": 0.2432, |
|
"step": 873 |
|
}, |
|
{ |
|
"epoch": 1.8246346555323592, |
|
"grad_norm": 0.037982327489911684, |
|
"learning_rate": 8.015019929721668e-06, |
|
"loss": 0.2344, |
|
"step": 874 |
|
}, |
|
{ |
|
"epoch": 1.826722338204593, |
|
"grad_norm": 0.03580893689768604, |
|
"learning_rate": 7.826569186075428e-06, |
|
"loss": 0.252, |
|
"step": 875 |
|
}, |
|
{ |
|
"epoch": 1.8288100208768268, |
|
"grad_norm": 0.035854237119299116, |
|
"learning_rate": 7.640316106191403e-06, |
|
"loss": 0.1963, |
|
"step": 876 |
|
}, |
|
{ |
|
"epoch": 1.8308977035490606, |
|
"grad_norm": 0.03548915356785303, |
|
"learning_rate": 7.4562628200292475e-06, |
|
"loss": 0.2148, |
|
"step": 877 |
|
}, |
|
{ |
|
"epoch": 1.8329853862212944, |
|
"grad_norm": 0.03888900644009877, |
|
"learning_rate": 7.274411432392026e-06, |
|
"loss": 0.2812, |
|
"step": 878 |
|
}, |
|
{ |
|
"epoch": 1.8350730688935282, |
|
"grad_norm": 0.036848786139394796, |
|
"learning_rate": 7.0947640229023675e-06, |
|
"loss": 0.1904, |
|
"step": 879 |
|
}, |
|
{ |
|
"epoch": 1.837160751565762, |
|
"grad_norm": 0.032594039620939604, |
|
"learning_rate": 6.917322645978463e-06, |
|
"loss": 0.1904, |
|
"step": 880 |
|
}, |
|
{ |
|
"epoch": 1.8392484342379958, |
|
"grad_norm": 0.038807798544498055, |
|
"learning_rate": 6.742089330810774e-06, |
|
"loss": 0.2383, |
|
"step": 881 |
|
}, |
|
{ |
|
"epoch": 1.8413361169102296, |
|
"grad_norm": 0.042279119809553053, |
|
"learning_rate": 6.5690660813386484e-06, |
|
"loss": 0.3184, |
|
"step": 882 |
|
}, |
|
{ |
|
"epoch": 1.8434237995824634, |
|
"grad_norm": 0.03575523378084652, |
|
"learning_rate": 6.398254876227561e-06, |
|
"loss": 0.1875, |
|
"step": 883 |
|
}, |
|
{ |
|
"epoch": 1.8455114822546972, |
|
"grad_norm": 0.03438492023581685, |
|
"learning_rate": 6.229657668846333e-06, |
|
"loss": 0.21, |
|
"step": 884 |
|
}, |
|
{ |
|
"epoch": 1.847599164926931, |
|
"grad_norm": 0.04355612584591555, |
|
"learning_rate": 6.0632763872449535e-06, |
|
"loss": 0.2754, |
|
"step": 885 |
|
}, |
|
{ |
|
"epoch": 1.8496868475991648, |
|
"grad_norm": 0.03533181091838774, |
|
"learning_rate": 5.899112934132345e-06, |
|
"loss": 0.1689, |
|
"step": 886 |
|
}, |
|
{ |
|
"epoch": 1.8517745302713986, |
|
"grad_norm": 0.040660080641486476, |
|
"learning_rate": 5.73716918685483e-06, |
|
"loss": 0.2471, |
|
"step": 887 |
|
}, |
|
{ |
|
"epoch": 1.8538622129436324, |
|
"grad_norm": 0.03235795768355878, |
|
"learning_rate": 5.577446997374458e-06, |
|
"loss": 0.1826, |
|
"step": 888 |
|
}, |
|
{ |
|
"epoch": 1.8559498956158664, |
|
"grad_norm": 0.03609407610302125, |
|
"learning_rate": 5.4199481922479146e-06, |
|
"loss": 0.1816, |
|
"step": 889 |
|
}, |
|
{ |
|
"epoch": 1.8580375782881002, |
|
"grad_norm": 0.03647085365814291, |
|
"learning_rate": 5.264674572605665e-06, |
|
"loss": 0.2393, |
|
"step": 890 |
|
}, |
|
{ |
|
"epoch": 1.860125260960334, |
|
"grad_norm": 0.03892481200719314, |
|
"learning_rate": 5.11162791413129e-06, |
|
"loss": 0.2021, |
|
"step": 891 |
|
}, |
|
{ |
|
"epoch": 1.8622129436325678, |
|
"grad_norm": 0.03859819801659933, |
|
"learning_rate": 4.9608099670412025e-06, |
|
"loss": 0.2402, |
|
"step": 892 |
|
}, |
|
{ |
|
"epoch": 1.8643006263048016, |
|
"grad_norm": 0.03638768037937149, |
|
"learning_rate": 4.812222456064697e-06, |
|
"loss": 0.1729, |
|
"step": 893 |
|
}, |
|
{ |
|
"epoch": 1.8663883089770354, |
|
"grad_norm": 0.03578385304837865, |
|
"learning_rate": 4.665867080424046e-06, |
|
"loss": 0.2217, |
|
"step": 894 |
|
}, |
|
{ |
|
"epoch": 1.8684759916492695, |
|
"grad_norm": 0.03899470496061886, |
|
"learning_rate": 4.521745513815345e-06, |
|
"loss": 0.2236, |
|
"step": 895 |
|
}, |
|
{ |
|
"epoch": 1.8705636743215033, |
|
"grad_norm": 0.03565199656237529, |
|
"learning_rate": 4.379859404389075e-06, |
|
"loss": 0.2158, |
|
"step": 896 |
|
}, |
|
{ |
|
"epoch": 1.872651356993737, |
|
"grad_norm": 0.03700782502510176, |
|
"learning_rate": 4.240210374731479e-06, |
|
"loss": 0.2441, |
|
"step": 897 |
|
}, |
|
{ |
|
"epoch": 1.8747390396659709, |
|
"grad_norm": 0.034012775228610695, |
|
"learning_rate": 4.10280002184591e-06, |
|
"loss": 0.1934, |
|
"step": 898 |
|
}, |
|
{ |
|
"epoch": 1.8768267223382047, |
|
"grad_norm": 0.032231872294556166, |
|
"learning_rate": 3.967629917134574e-06, |
|
"loss": 0.1689, |
|
"step": 899 |
|
}, |
|
{ |
|
"epoch": 1.8789144050104385, |
|
"grad_norm": 0.039125390388194826, |
|
"learning_rate": 3.834701606380575e-06, |
|
"loss": 0.2285, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 1.8810020876826723, |
|
"grad_norm": 0.04416962700457997, |
|
"learning_rate": 3.7040166097301877e-06, |
|
"loss": 0.2656, |
|
"step": 901 |
|
}, |
|
{ |
|
"epoch": 1.883089770354906, |
|
"grad_norm": 0.04111412261545973, |
|
"learning_rate": 3.575576421675586e-06, |
|
"loss": 0.248, |
|
"step": 902 |
|
}, |
|
{ |
|
"epoch": 1.88517745302714, |
|
"grad_norm": 0.03397805823890809, |
|
"learning_rate": 3.449382511037613e-06, |
|
"loss": 0.1904, |
|
"step": 903 |
|
}, |
|
{ |
|
"epoch": 1.8872651356993737, |
|
"grad_norm": 0.041722415283545866, |
|
"learning_rate": 3.3254363209491046e-06, |
|
"loss": 0.3301, |
|
"step": 904 |
|
}, |
|
{ |
|
"epoch": 1.8893528183716075, |
|
"grad_norm": 0.047301660349510964, |
|
"learning_rate": 3.203739268838324e-06, |
|
"loss": 0.2988, |
|
"step": 905 |
|
}, |
|
{ |
|
"epoch": 1.8914405010438413, |
|
"grad_norm": 0.03091322302919706, |
|
"learning_rate": 3.0842927464127537e-06, |
|
"loss": 0.1592, |
|
"step": 906 |
|
}, |
|
{ |
|
"epoch": 1.8935281837160751, |
|
"grad_norm": 0.038250787995771175, |
|
"learning_rate": 2.9670981196431525e-06, |
|
"loss": 0.248, |
|
"step": 907 |
|
}, |
|
{ |
|
"epoch": 1.895615866388309, |
|
"grad_norm": 0.04184624767282992, |
|
"learning_rate": 2.8521567287480344e-06, |
|
"loss": 0.2461, |
|
"step": 908 |
|
}, |
|
{ |
|
"epoch": 1.8977035490605427, |
|
"grad_norm": 0.03669678971086905, |
|
"learning_rate": 2.7394698881782367e-06, |
|
"loss": 0.1895, |
|
"step": 909 |
|
}, |
|
{ |
|
"epoch": 1.8997912317327765, |
|
"grad_norm": 0.03332356389562326, |
|
"learning_rate": 2.6290388866019533e-06, |
|
"loss": 0.1934, |
|
"step": 910 |
|
}, |
|
{ |
|
"epoch": 1.9018789144050103, |
|
"grad_norm": 0.03567454628961636, |
|
"learning_rate": 2.5208649868899704e-06, |
|
"loss": 0.2207, |
|
"step": 911 |
|
}, |
|
{ |
|
"epoch": 1.9039665970772441, |
|
"grad_norm": 0.03723478614951914, |
|
"learning_rate": 2.4149494261012097e-06, |
|
"loss": 0.2002, |
|
"step": 912 |
|
}, |
|
{ |
|
"epoch": 1.906054279749478, |
|
"grad_norm": 0.03840886510699952, |
|
"learning_rate": 2.3112934154686296e-06, |
|
"loss": 0.1982, |
|
"step": 913 |
|
}, |
|
{ |
|
"epoch": 1.9081419624217117, |
|
"grad_norm": 0.03658052263392586, |
|
"learning_rate": 2.2098981403852804e-06, |
|
"loss": 0.2295, |
|
"step": 914 |
|
}, |
|
{ |
|
"epoch": 1.9102296450939458, |
|
"grad_norm": 0.042943855139820596, |
|
"learning_rate": 2.1107647603908933e-06, |
|
"loss": 0.2422, |
|
"step": 915 |
|
}, |
|
{ |
|
"epoch": 1.9123173277661796, |
|
"grad_norm": 0.038181568479852754, |
|
"learning_rate": 2.0138944091585123e-06, |
|
"loss": 0.1768, |
|
"step": 916 |
|
}, |
|
{ |
|
"epoch": 1.9144050104384134, |
|
"grad_norm": 0.04044402128707661, |
|
"learning_rate": 1.91928819448155e-06, |
|
"loss": 0.249, |
|
"step": 917 |
|
}, |
|
{ |
|
"epoch": 1.9164926931106472, |
|
"grad_norm": 0.04021345081714543, |
|
"learning_rate": 1.8269471982611086e-06, |
|
"loss": 0.2578, |
|
"step": 918 |
|
}, |
|
{ |
|
"epoch": 1.918580375782881, |
|
"grad_norm": 0.03765027742041835, |
|
"learning_rate": 1.7368724764936118e-06, |
|
"loss": 0.2275, |
|
"step": 919 |
|
}, |
|
{ |
|
"epoch": 1.9206680584551148, |
|
"grad_norm": 0.03276750255084186, |
|
"learning_rate": 1.6490650592588363e-06, |
|
"loss": 0.1797, |
|
"step": 920 |
|
}, |
|
{ |
|
"epoch": 1.9227557411273486, |
|
"grad_norm": 0.03836166999423338, |
|
"learning_rate": 1.5635259507079224e-06, |
|
"loss": 0.2656, |
|
"step": 921 |
|
}, |
|
{ |
|
"epoch": 1.9248434237995826, |
|
"grad_norm": 0.040716455351115934, |
|
"learning_rate": 1.4802561290520934e-06, |
|
"loss": 0.2373, |
|
"step": 922 |
|
}, |
|
{ |
|
"epoch": 1.9269311064718164, |
|
"grad_norm": 0.044722992742083724, |
|
"learning_rate": 1.3992565465512863e-06, |
|
"loss": 0.2832, |
|
"step": 923 |
|
}, |
|
{ |
|
"epoch": 1.9290187891440502, |
|
"grad_norm": 0.035718501429723326, |
|
"learning_rate": 1.3205281295034512e-06, |
|
"loss": 0.2109, |
|
"step": 924 |
|
}, |
|
{ |
|
"epoch": 1.931106471816284, |
|
"grad_norm": 0.040897036162254374, |
|
"learning_rate": 1.2440717782338018e-06, |
|
"loss": 0.248, |
|
"step": 925 |
|
}, |
|
{ |
|
"epoch": 1.9331941544885178, |
|
"grad_norm": 0.037701179364082175, |
|
"learning_rate": 1.169888367084626e-06, |
|
"loss": 0.2021, |
|
"step": 926 |
|
}, |
|
{ |
|
"epoch": 1.9352818371607516, |
|
"grad_norm": 0.03800126744938023, |
|
"learning_rate": 1.0979787444051815e-06, |
|
"loss": 0.2051, |
|
"step": 927 |
|
}, |
|
{ |
|
"epoch": 1.9373695198329854, |
|
"grad_norm": 0.0376053096363303, |
|
"learning_rate": 1.0283437325421252e-06, |
|
"loss": 0.25, |
|
"step": 928 |
|
}, |
|
{ |
|
"epoch": 1.9394572025052192, |
|
"grad_norm": 0.03852373604309701, |
|
"learning_rate": 9.60984127829989e-07, |
|
"loss": 0.1895, |
|
"step": 929 |
|
}, |
|
{ |
|
"epoch": 1.941544885177453, |
|
"grad_norm": 0.05141719488909021, |
|
"learning_rate": 8.95900700582164e-07, |
|
"loss": 0.2969, |
|
"step": 930 |
|
}, |
|
{ |
|
"epoch": 1.9436325678496869, |
|
"grad_norm": 0.0422557444402279, |
|
"learning_rate": 8.330941950819738e-07, |
|
"loss": 0.2793, |
|
"step": 931 |
|
}, |
|
{ |
|
"epoch": 1.9457202505219207, |
|
"grad_norm": 0.03878673278759856, |
|
"learning_rate": 7.725653295743484e-07, |
|
"loss": 0.2412, |
|
"step": 932 |
|
}, |
|
{ |
|
"epoch": 1.9478079331941545, |
|
"grad_norm": 0.04515324231545769, |
|
"learning_rate": 7.14314796257387e-07, |
|
"loss": 0.2715, |
|
"step": 933 |
|
}, |
|
{ |
|
"epoch": 1.9498956158663883, |
|
"grad_norm": 0.04060496658163304, |
|
"learning_rate": 6.58343261274652e-07, |
|
"loss": 0.2559, |
|
"step": 934 |
|
}, |
|
{ |
|
"epoch": 1.951983298538622, |
|
"grad_norm": 0.0410883973174675, |
|
"learning_rate": 6.046513647074203e-07, |
|
"loss": 0.2676, |
|
"step": 935 |
|
}, |
|
{ |
|
"epoch": 1.9540709812108559, |
|
"grad_norm": 0.037425826708713576, |
|
"learning_rate": 5.532397205673556e-07, |
|
"loss": 0.208, |
|
"step": 936 |
|
}, |
|
{ |
|
"epoch": 1.9561586638830897, |
|
"grad_norm": 0.03711360512013879, |
|
"learning_rate": 5.04108916789603e-07, |
|
"loss": 0.2246, |
|
"step": 937 |
|
}, |
|
{ |
|
"epoch": 1.9582463465553235, |
|
"grad_norm": 0.04129502522262085, |
|
"learning_rate": 4.5725951522592734e-07, |
|
"loss": 0.2812, |
|
"step": 938 |
|
}, |
|
{ |
|
"epoch": 1.9603340292275573, |
|
"grad_norm": 0.042768062565203456, |
|
"learning_rate": 4.126920516384303e-07, |
|
"loss": 0.2734, |
|
"step": 939 |
|
}, |
|
{ |
|
"epoch": 1.962421711899791, |
|
"grad_norm": 0.03636718305173902, |
|
"learning_rate": 3.704070356932432e-07, |
|
"loss": 0.2275, |
|
"step": 940 |
|
}, |
|
{ |
|
"epoch": 1.964509394572025, |
|
"grad_norm": 0.03914519289291409, |
|
"learning_rate": 3.3040495095491006e-07, |
|
"loss": 0.2051, |
|
"step": 941 |
|
}, |
|
{ |
|
"epoch": 1.966597077244259, |
|
"grad_norm": 0.036435492987632524, |
|
"learning_rate": 2.926862548806364e-07, |
|
"loss": 0.21, |
|
"step": 942 |
|
}, |
|
{ |
|
"epoch": 1.9686847599164927, |
|
"grad_norm": 0.04520438858168889, |
|
"learning_rate": 2.5725137881529306e-07, |
|
"loss": 0.2852, |
|
"step": 943 |
|
}, |
|
{ |
|
"epoch": 1.9707724425887265, |
|
"grad_norm": 0.036719387383711895, |
|
"learning_rate": 2.2410072798624283e-07, |
|
"loss": 0.2422, |
|
"step": 944 |
|
}, |
|
{ |
|
"epoch": 1.9728601252609603, |
|
"grad_norm": 0.03890839218462144, |
|
"learning_rate": 1.9323468149892165e-07, |
|
"loss": 0.2617, |
|
"step": 945 |
|
}, |
|
{ |
|
"epoch": 1.9749478079331941, |
|
"grad_norm": 0.04401800316091134, |
|
"learning_rate": 1.6465359233233114e-07, |
|
"loss": 0.2314, |
|
"step": 946 |
|
}, |
|
{ |
|
"epoch": 1.977035490605428, |
|
"grad_norm": 0.042526472328545395, |
|
"learning_rate": 1.383577873351305e-07, |
|
"loss": 0.2734, |
|
"step": 947 |
|
}, |
|
{ |
|
"epoch": 1.9791231732776617, |
|
"grad_norm": 0.04123640104260365, |
|
"learning_rate": 1.143475672218175e-07, |
|
"loss": 0.2715, |
|
"step": 948 |
|
}, |
|
{ |
|
"epoch": 1.9812108559498958, |
|
"grad_norm": 0.041143122691388076, |
|
"learning_rate": 9.26232065693089e-08, |
|
"loss": 0.2578, |
|
"step": 949 |
|
}, |
|
{ |
|
"epoch": 1.9832985386221296, |
|
"grad_norm": 0.03646932624947557, |
|
"learning_rate": 7.318495381383184e-08, |
|
"loss": 0.1953, |
|
"step": 950 |
|
}, |
|
{ |
|
"epoch": 1.9853862212943634, |
|
"grad_norm": 0.042625916797236235, |
|
"learning_rate": 5.603303124803727e-08, |
|
"loss": 0.2344, |
|
"step": 951 |
|
}, |
|
{ |
|
"epoch": 1.9874739039665972, |
|
"grad_norm": 0.039620902455050594, |
|
"learning_rate": 4.1167635018513064e-08, |
|
"loss": 0.2598, |
|
"step": 952 |
|
}, |
|
{ |
|
"epoch": 1.989561586638831, |
|
"grad_norm": 0.0329100227494744, |
|
"learning_rate": 2.8588935123474714e-08, |
|
"loss": 0.1758, |
|
"step": 953 |
|
}, |
|
{ |
|
"epoch": 1.9916492693110648, |
|
"grad_norm": 0.03829557846462901, |
|
"learning_rate": 1.8297075410877997e-08, |
|
"loss": 0.2676, |
|
"step": 954 |
|
}, |
|
{ |
|
"epoch": 1.9937369519832986, |
|
"grad_norm": 0.037593692719248624, |
|
"learning_rate": 1.02921735767314e-08, |
|
"loss": 0.1943, |
|
"step": 955 |
|
}, |
|
{ |
|
"epoch": 1.9958246346555324, |
|
"grad_norm": 0.03806855037144764, |
|
"learning_rate": 4.574321163763884e-09, |
|
"loss": 0.2383, |
|
"step": 956 |
|
}, |
|
{ |
|
"epoch": 1.9979123173277662, |
|
"grad_norm": 0.03559582629136701, |
|
"learning_rate": 1.1435835604034495e-09, |
|
"loss": 0.1787, |
|
"step": 957 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"grad_norm": 0.03239360059033573, |
|
"learning_rate": 0.0, |
|
"loss": 0.167, |
|
"step": 958 |
|
}, |
|
{ |
|
"epoch": 2.0, |
|
"step": 958, |
|
"total_flos": 5.650597015940235e+18, |
|
"train_loss": 0.3088689816022965, |
|
"train_runtime": 6060.4121, |
|
"train_samples_per_second": 90.951, |
|
"train_steps_per_second": 0.158 |
|
} |
|
], |
|
"logging_steps": 1.0, |
|
"max_steps": 958, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 2, |
|
"save_steps": 500, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.650597015940235e+18, |
|
"train_batch_size": 72, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|