|
[ |
|
{ |
|
"current_steps": 127, |
|
"loss": 1.7356, |
|
"learning_rate": 0.0, |
|
"epoch": 0.01 |
|
}, |
|
{ |
|
"current_steps": 255, |
|
"loss": 1.7333, |
|
"learning_rate": 0.0003, |
|
"epoch": 0.01 |
|
}, |
|
{ |
|
"current_steps": 383, |
|
"loss": 1.7095, |
|
"learning_rate": 0.00029932735426008964, |
|
"epoch": 0.02 |
|
}, |
|
{ |
|
"current_steps": 511, |
|
"loss": 1.4713, |
|
"learning_rate": 0.00029865470852017935, |
|
"epoch": 0.03 |
|
}, |
|
{ |
|
"current_steps": 639, |
|
"loss": 1.3242, |
|
"learning_rate": 0.000297982062780269, |
|
"epoch": 0.03 |
|
}, |
|
{ |
|
"current_steps": 767, |
|
"loss": 1.3469, |
|
"learning_rate": 0.00029730941704035873, |
|
"epoch": 0.04 |
|
}, |
|
{ |
|
"current_steps": 895, |
|
"loss": 1.2645, |
|
"learning_rate": 0.0002966367713004484, |
|
"epoch": 0.05 |
|
}, |
|
{ |
|
"current_steps": 1023, |
|
"loss": 1.2507, |
|
"learning_rate": 0.00029596412556053806, |
|
"epoch": 0.05 |
|
}, |
|
{ |
|
"current_steps": 1151, |
|
"loss": 1.2348, |
|
"learning_rate": 0.00029529147982062777, |
|
"epoch": 0.06 |
|
}, |
|
{ |
|
"current_steps": 1279, |
|
"loss": 1.3604, |
|
"learning_rate": 0.00029461883408071743, |
|
"epoch": 0.07 |
|
}, |
|
{ |
|
"current_steps": 1407, |
|
"loss": 1.2057, |
|
"learning_rate": 0.00029394618834080715, |
|
"epoch": 0.07 |
|
}, |
|
{ |
|
"current_steps": 1535, |
|
"loss": 1.2899, |
|
"learning_rate": 0.00029327354260089687, |
|
"epoch": 0.08 |
|
}, |
|
{ |
|
"current_steps": 1663, |
|
"loss": 1.2692, |
|
"learning_rate": 0.00029260089686098653, |
|
"epoch": 0.09 |
|
}, |
|
{ |
|
"current_steps": 1791, |
|
"loss": 1.169, |
|
"learning_rate": 0.0002919282511210762, |
|
"epoch": 0.09 |
|
}, |
|
{ |
|
"current_steps": 1919, |
|
"loss": 1.1576, |
|
"learning_rate": 0.0002912556053811659, |
|
"epoch": 0.1 |
|
}, |
|
{ |
|
"current_steps": 2047, |
|
"loss": 1.1784, |
|
"learning_rate": 0.00029058295964125557, |
|
"epoch": 0.11 |
|
}, |
|
{ |
|
"current_steps": 2175, |
|
"loss": 1.1418, |
|
"learning_rate": 0.0002899103139013453, |
|
"epoch": 0.11 |
|
}, |
|
{ |
|
"current_steps": 2303, |
|
"loss": 1.2816, |
|
"learning_rate": 0.00028923766816143495, |
|
"epoch": 0.12 |
|
}, |
|
{ |
|
"current_steps": 2431, |
|
"loss": 1.2302, |
|
"learning_rate": 0.00028856502242152467, |
|
"epoch": 0.13 |
|
}, |
|
{ |
|
"current_steps": 2559, |
|
"loss": 1.1729, |
|
"learning_rate": 0.00028789237668161433, |
|
"epoch": 0.13 |
|
}, |
|
{ |
|
"current_steps": 2687, |
|
"loss": 1.1625, |
|
"learning_rate": 0.000287219730941704, |
|
"epoch": 0.14 |
|
}, |
|
{ |
|
"current_steps": 2815, |
|
"loss": 1.1987, |
|
"learning_rate": 0.0002865470852017937, |
|
"epoch": 0.15 |
|
}, |
|
{ |
|
"current_steps": 2943, |
|
"loss": 1.1673, |
|
"learning_rate": 0.00028587443946188337, |
|
"epoch": 0.15 |
|
}, |
|
{ |
|
"current_steps": 3071, |
|
"loss": 1.2074, |
|
"learning_rate": 0.0002852017937219731, |
|
"epoch": 0.16 |
|
}, |
|
{ |
|
"current_steps": 3199, |
|
"loss": 1.1848, |
|
"learning_rate": 0.00028452914798206275, |
|
"epoch": 0.17 |
|
}, |
|
{ |
|
"current_steps": 3327, |
|
"loss": 1.1209, |
|
"learning_rate": 0.0002838565022421524, |
|
"epoch": 0.17 |
|
}, |
|
{ |
|
"current_steps": 3455, |
|
"loss": 1.091, |
|
"learning_rate": 0.0002831838565022421, |
|
"epoch": 0.18 |
|
}, |
|
{ |
|
"current_steps": 3583, |
|
"loss": 1.1893, |
|
"learning_rate": 0.00028251121076233184, |
|
"epoch": 0.19 |
|
}, |
|
{ |
|
"current_steps": 3711, |
|
"loss": 1.128, |
|
"learning_rate": 0.0002818385650224215, |
|
"epoch": 0.19 |
|
}, |
|
{ |
|
"current_steps": 3839, |
|
"loss": 1.0913, |
|
"learning_rate": 0.0002811659192825112, |
|
"epoch": 0.2 |
|
}, |
|
{ |
|
"current_steps": 3967, |
|
"loss": 1.1659, |
|
"learning_rate": 0.0002804932735426009, |
|
"epoch": 0.21 |
|
}, |
|
{ |
|
"current_steps": 4095, |
|
"loss": 1.1555, |
|
"learning_rate": 0.00027982062780269054, |
|
"epoch": 0.21 |
|
}, |
|
{ |
|
"current_steps": 4223, |
|
"loss": 1.1274, |
|
"learning_rate": 0.00027914798206278026, |
|
"epoch": 0.22 |
|
}, |
|
{ |
|
"current_steps": 4351, |
|
"loss": 1.1193, |
|
"learning_rate": 0.0002784753363228699, |
|
"epoch": 0.23 |
|
}, |
|
{ |
|
"current_steps": 4479, |
|
"loss": 1.1114, |
|
"learning_rate": 0.00027780269058295964, |
|
"epoch": 0.23 |
|
}, |
|
{ |
|
"current_steps": 4607, |
|
"loss": 1.1123, |
|
"learning_rate": 0.0002771300448430493, |
|
"epoch": 0.24 |
|
}, |
|
{ |
|
"current_steps": 4735, |
|
"loss": 1.135, |
|
"learning_rate": 0.00027645739910313896, |
|
"epoch": 0.25 |
|
}, |
|
{ |
|
"current_steps": 4863, |
|
"loss": 1.1067, |
|
"learning_rate": 0.0002757847533632287, |
|
"epoch": 0.25 |
|
}, |
|
{ |
|
"current_steps": 4991, |
|
"loss": 1.0959, |
|
"learning_rate": 0.00027511210762331834, |
|
"epoch": 0.26 |
|
}, |
|
{ |
|
"current_steps": 5119, |
|
"loss": 1.0699, |
|
"learning_rate": 0.00027443946188340806, |
|
"epoch": 0.27 |
|
}, |
|
{ |
|
"current_steps": 5247, |
|
"loss": 1.1366, |
|
"learning_rate": 0.0002737668161434977, |
|
"epoch": 0.27 |
|
}, |
|
{ |
|
"current_steps": 5375, |
|
"loss": 1.1146, |
|
"learning_rate": 0.0002730941704035874, |
|
"epoch": 0.28 |
|
}, |
|
{ |
|
"current_steps": 5503, |
|
"loss": 1.1146, |
|
"learning_rate": 0.0002730941704035874, |
|
"epoch": 0.28 |
|
} |
|
] |