|
{ |
|
"best_metric": null, |
|
"best_model_checkpoint": null, |
|
"epoch": 1.0, |
|
"eval_steps": 1660, |
|
"global_step": 33185, |
|
"is_hyper_param_search": false, |
|
"is_local_process_zero": true, |
|
"is_world_process_zero": true, |
|
"log_history": [ |
|
{ |
|
"epoch": 3.0134096730450506e-05, |
|
"grad_norm": 8.0, |
|
"learning_rate": 2e-06, |
|
"loss": 1.334, |
|
"step": 1 |
|
}, |
|
{ |
|
"epoch": 0.0030134096730450506, |
|
"grad_norm": 0.34765625, |
|
"learning_rate": 0.0002, |
|
"loss": 0.9372, |
|
"step": 100 |
|
}, |
|
{ |
|
"epoch": 0.006026819346090101, |
|
"grad_norm": 1.2734375, |
|
"learning_rate": 0.0004, |
|
"loss": 0.6477, |
|
"step": 200 |
|
}, |
|
{ |
|
"epoch": 0.009040229019135152, |
|
"grad_norm": 0.59375, |
|
"learning_rate": 0.0006, |
|
"loss": 0.5967, |
|
"step": 300 |
|
}, |
|
{ |
|
"epoch": 0.012053638692180202, |
|
"grad_norm": 1.40625, |
|
"learning_rate": 0.0008, |
|
"loss": 0.5828, |
|
"step": 400 |
|
}, |
|
{ |
|
"epoch": 0.015067048365225252, |
|
"grad_norm": 0.453125, |
|
"learning_rate": 0.001, |
|
"loss": 0.5774, |
|
"step": 500 |
|
}, |
|
{ |
|
"epoch": 0.018080458038270304, |
|
"grad_norm": 0.890625, |
|
"learning_rate": 0.0012, |
|
"loss": 0.557, |
|
"step": 600 |
|
}, |
|
{ |
|
"epoch": 0.021093867711315353, |
|
"grad_norm": 0.396484375, |
|
"learning_rate": 0.0014, |
|
"loss": 0.5573, |
|
"step": 700 |
|
}, |
|
{ |
|
"epoch": 0.024107277384360404, |
|
"grad_norm": 0.416015625, |
|
"learning_rate": 0.0016, |
|
"loss": 0.5577, |
|
"step": 800 |
|
}, |
|
{ |
|
"epoch": 0.027120687057405453, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.0018000000000000002, |
|
"loss": 0.563, |
|
"step": 900 |
|
}, |
|
{ |
|
"epoch": 0.030134096730450505, |
|
"grad_norm": 0.58984375, |
|
"learning_rate": 0.002, |
|
"loss": 0.568, |
|
"step": 1000 |
|
}, |
|
{ |
|
"epoch": 0.03314750640349556, |
|
"grad_norm": 0.33203125, |
|
"learning_rate": 0.0019999571252319053, |
|
"loss": 0.5969, |
|
"step": 1100 |
|
}, |
|
{ |
|
"epoch": 0.03616091607654061, |
|
"grad_norm": 0.5234375, |
|
"learning_rate": 0.0019998285050126107, |
|
"loss": 0.6365, |
|
"step": 1200 |
|
}, |
|
{ |
|
"epoch": 0.03917432574958565, |
|
"grad_norm": 0.302734375, |
|
"learning_rate": 0.0019996141515967, |
|
"loss": 0.6351, |
|
"step": 1300 |
|
}, |
|
{ |
|
"epoch": 0.042187735422630705, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.001999314085407178, |
|
"loss": 0.5537, |
|
"step": 1400 |
|
}, |
|
{ |
|
"epoch": 0.04520114509567576, |
|
"grad_norm": 0.271484375, |
|
"learning_rate": 0.0019989283350335314, |
|
"loss": 0.5484, |
|
"step": 1500 |
|
}, |
|
{ |
|
"epoch": 0.04821455476872081, |
|
"grad_norm": 0.375, |
|
"learning_rate": 0.0019984569372289993, |
|
"loss": 0.5583, |
|
"step": 1600 |
|
}, |
|
{ |
|
"epoch": 0.050022600572547836, |
|
"eval_peoplespeech-clean-transcription_loss": 4.166935443878174, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 15.2594, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.194, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.066, |
|
"step": 1660 |
|
}, |
|
{ |
|
"epoch": 0.05122796444176586, |
|
"grad_norm": 0.2578125, |
|
"learning_rate": 0.0019978999369070737, |
|
"loss": 0.5722, |
|
"step": 1700 |
|
}, |
|
{ |
|
"epoch": 0.054241374114810906, |
|
"grad_norm": 0.3203125, |
|
"learning_rate": 0.001997257387137221, |
|
"loss": 0.5699, |
|
"step": 1800 |
|
}, |
|
{ |
|
"epoch": 0.05725478378785596, |
|
"grad_norm": 0.27734375, |
|
"learning_rate": 0.0019965293491398237, |
|
"loss": 0.595, |
|
"step": 1900 |
|
}, |
|
{ |
|
"epoch": 0.06026819346090101, |
|
"grad_norm": 0.455078125, |
|
"learning_rate": 0.001995715892280349, |
|
"loss": 0.561, |
|
"step": 2000 |
|
}, |
|
{ |
|
"epoch": 0.06328160313394605, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 0.00199481709406274, |
|
"loss": 0.5553, |
|
"step": 2100 |
|
}, |
|
{ |
|
"epoch": 0.06629501280699111, |
|
"grad_norm": 0.27734375, |
|
"learning_rate": 0.0019938330401220307, |
|
"loss": 0.5668, |
|
"step": 2200 |
|
}, |
|
{ |
|
"epoch": 0.06930842248003616, |
|
"grad_norm": 0.318359375, |
|
"learning_rate": 0.0019927638242161864, |
|
"loss": 0.574, |
|
"step": 2300 |
|
}, |
|
{ |
|
"epoch": 0.07232183215308122, |
|
"grad_norm": 0.1767578125, |
|
"learning_rate": 0.001991609548217171, |
|
"loss": 0.6076, |
|
"step": 2400 |
|
}, |
|
{ |
|
"epoch": 0.07533524182612626, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 0.001990370322101242, |
|
"loss": 0.6369, |
|
"step": 2500 |
|
}, |
|
{ |
|
"epoch": 0.0783486514991713, |
|
"grad_norm": 0.314453125, |
|
"learning_rate": 0.001989046263938472, |
|
"loss": 0.6106, |
|
"step": 2600 |
|
}, |
|
{ |
|
"epoch": 0.08136206117221637, |
|
"grad_norm": 0.240234375, |
|
"learning_rate": 0.0019876374998814973, |
|
"loss": 0.5993, |
|
"step": 2700 |
|
}, |
|
{ |
|
"epoch": 0.08437547084526141, |
|
"grad_norm": 0.51953125, |
|
"learning_rate": 0.0019861441641535007, |
|
"loss": 0.5933, |
|
"step": 2800 |
|
}, |
|
{ |
|
"epoch": 0.08738888051830647, |
|
"grad_norm": 0.609375, |
|
"learning_rate": 0.001984566399035423, |
|
"loss": 0.5937, |
|
"step": 2900 |
|
}, |
|
{ |
|
"epoch": 0.09040229019135151, |
|
"grad_norm": 0.21484375, |
|
"learning_rate": 0.001982904354852404, |
|
"loss": 0.5881, |
|
"step": 3000 |
|
}, |
|
{ |
|
"epoch": 0.09341569986439656, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 0.0019811581899594646, |
|
"loss": 0.5817, |
|
"step": 3100 |
|
}, |
|
{ |
|
"epoch": 0.09642910953744162, |
|
"grad_norm": 0.2138671875, |
|
"learning_rate": 0.0019793280707264154, |
|
"loss": 0.588, |
|
"step": 3200 |
|
}, |
|
{ |
|
"epoch": 0.09944251921048666, |
|
"grad_norm": 0.23828125, |
|
"learning_rate": 0.0019774141715220065, |
|
"loss": 0.5813, |
|
"step": 3300 |
|
}, |
|
{ |
|
"epoch": 0.10004520114509567, |
|
"eval_peoplespeech-clean-transcription_loss": 4.31866455078125, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.293, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.478, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.07, |
|
"step": 3320 |
|
}, |
|
{ |
|
"epoch": 0.10245592888353172, |
|
"grad_norm": 0.2080078125, |
|
"learning_rate": 0.0019754166746973156, |
|
"loss": 0.5784, |
|
"step": 3400 |
|
}, |
|
{ |
|
"epoch": 0.10546933855657677, |
|
"grad_norm": 0.2119140625, |
|
"learning_rate": 0.0019733357705683705, |
|
"loss": 0.582, |
|
"step": 3500 |
|
}, |
|
{ |
|
"epoch": 0.10848274822962181, |
|
"grad_norm": 0.248046875, |
|
"learning_rate": 0.001971171657398021, |
|
"loss": 0.5877, |
|
"step": 3600 |
|
}, |
|
{ |
|
"epoch": 0.11149615790266687, |
|
"grad_norm": 0.255859375, |
|
"learning_rate": 0.001968924541377045, |
|
"loss": 0.5788, |
|
"step": 3700 |
|
}, |
|
{ |
|
"epoch": 0.11450956757571192, |
|
"grad_norm": 0.271484375, |
|
"learning_rate": 0.001966594636604506, |
|
"loss": 0.5739, |
|
"step": 3800 |
|
}, |
|
{ |
|
"epoch": 0.11752297724875697, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 0.001964182165067352, |
|
"loss": 0.5748, |
|
"step": 3900 |
|
}, |
|
{ |
|
"epoch": 0.12053638692180202, |
|
"grad_norm": 7.28125, |
|
"learning_rate": 0.001961687356619266, |
|
"loss": 0.5746, |
|
"step": 4000 |
|
}, |
|
{ |
|
"epoch": 0.12354979659484706, |
|
"grad_norm": 0.25, |
|
"learning_rate": 0.001959110448958769, |
|
"loss": 0.5877, |
|
"step": 4100 |
|
}, |
|
{ |
|
"epoch": 0.1265632062678921, |
|
"grad_norm": 0.2421875, |
|
"learning_rate": 0.001956451687606567, |
|
"loss": 0.5652, |
|
"step": 4200 |
|
}, |
|
{ |
|
"epoch": 0.12957661594093717, |
|
"grad_norm": 0.26171875, |
|
"learning_rate": 0.0019537113258821636, |
|
"loss": 0.5842, |
|
"step": 4300 |
|
}, |
|
{ |
|
"epoch": 0.13259002561398223, |
|
"grad_norm": 0.2158203125, |
|
"learning_rate": 0.001950889624879722, |
|
"loss": 0.5687, |
|
"step": 4400 |
|
}, |
|
{ |
|
"epoch": 0.13560343528702729, |
|
"grad_norm": 0.32421875, |
|
"learning_rate": 0.0019479868534431892, |
|
"loss": 0.5715, |
|
"step": 4500 |
|
}, |
|
{ |
|
"epoch": 0.13861684496007232, |
|
"grad_norm": 0.251953125, |
|
"learning_rate": 0.001945003288140681, |
|
"loss": 0.5811, |
|
"step": 4600 |
|
}, |
|
{ |
|
"epoch": 0.14163025463311738, |
|
"grad_norm": 0.44921875, |
|
"learning_rate": 0.0019419392132381317, |
|
"loss": 0.5936, |
|
"step": 4700 |
|
}, |
|
{ |
|
"epoch": 0.14464366430616243, |
|
"grad_norm": 0.24609375, |
|
"learning_rate": 0.0019387949206722099, |
|
"loss": 0.5861, |
|
"step": 4800 |
|
}, |
|
{ |
|
"epoch": 0.14765707397920746, |
|
"grad_norm": 0.255859375, |
|
"learning_rate": 0.0019355707100225034, |
|
"loss": 0.5867, |
|
"step": 4900 |
|
}, |
|
{ |
|
"epoch": 0.1500678017176435, |
|
"eval_peoplespeech-clean-transcription_loss": 4.08488130569458, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.7738, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.332, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.068, |
|
"step": 4980 |
|
}, |
|
{ |
|
"epoch": 0.15067048365225252, |
|
"grad_norm": 0.22265625, |
|
"learning_rate": 0.0019322668884829768, |
|
"loss": 0.5827, |
|
"step": 5000 |
|
}, |
|
{ |
|
"epoch": 0.15368389332529758, |
|
"grad_norm": 0.267578125, |
|
"learning_rate": 0.0019288837708327019, |
|
"loss": 0.5829, |
|
"step": 5100 |
|
}, |
|
{ |
|
"epoch": 0.1566973029983426, |
|
"grad_norm": 0.2236328125, |
|
"learning_rate": 0.0019254216794058665, |
|
"loss": 0.574, |
|
"step": 5200 |
|
}, |
|
{ |
|
"epoch": 0.15971071267138767, |
|
"grad_norm": 0.232421875, |
|
"learning_rate": 0.0019218809440610645, |
|
"loss": 0.5907, |
|
"step": 5300 |
|
}, |
|
{ |
|
"epoch": 0.16272412234443273, |
|
"grad_norm": 0.234375, |
|
"learning_rate": 0.0019182619021498664, |
|
"loss": 0.5736, |
|
"step": 5400 |
|
}, |
|
{ |
|
"epoch": 0.1657375320174778, |
|
"grad_norm": 0.2451171875, |
|
"learning_rate": 0.001914564898484678, |
|
"loss": 0.586, |
|
"step": 5500 |
|
}, |
|
{ |
|
"epoch": 0.16875094169052282, |
|
"grad_norm": 0.2080078125, |
|
"learning_rate": 0.0019107902853058875, |
|
"loss": 0.583, |
|
"step": 5600 |
|
}, |
|
{ |
|
"epoch": 0.17176435136356788, |
|
"grad_norm": 0.2314453125, |
|
"learning_rate": 0.0019069384222483061, |
|
"loss": 0.589, |
|
"step": 5700 |
|
}, |
|
{ |
|
"epoch": 0.17477776103661294, |
|
"grad_norm": 0.390625, |
|
"learning_rate": 0.0019030096763069007, |
|
"loss": 0.569, |
|
"step": 5800 |
|
}, |
|
{ |
|
"epoch": 0.17779117070965797, |
|
"grad_norm": 0.24609375, |
|
"learning_rate": 0.0018990044218018295, |
|
"loss": 0.5914, |
|
"step": 5900 |
|
}, |
|
{ |
|
"epoch": 0.18080458038270303, |
|
"grad_norm": 0.205078125, |
|
"learning_rate": 0.0018949230403427768, |
|
"loss": 0.5936, |
|
"step": 6000 |
|
}, |
|
{ |
|
"epoch": 0.1838179900557481, |
|
"grad_norm": 0.2353515625, |
|
"learning_rate": 0.0018907659207925951, |
|
"loss": 0.5959, |
|
"step": 6100 |
|
}, |
|
{ |
|
"epoch": 0.18683139972879312, |
|
"grad_norm": 0.220703125, |
|
"learning_rate": 0.0018865334592302553, |
|
"loss": 0.5734, |
|
"step": 6200 |
|
}, |
|
{ |
|
"epoch": 0.18984480940183818, |
|
"grad_norm": 0.2431640625, |
|
"learning_rate": 0.0018822260589131075, |
|
"loss": 0.5815, |
|
"step": 6300 |
|
}, |
|
{ |
|
"epoch": 0.19285821907488324, |
|
"grad_norm": 0.2158203125, |
|
"learning_rate": 0.0018778441302384629, |
|
"loss": 0.58, |
|
"step": 6400 |
|
}, |
|
{ |
|
"epoch": 0.1958716287479283, |
|
"grad_norm": 0.2734375, |
|
"learning_rate": 0.0018733880907044892, |
|
"loss": 0.5807, |
|
"step": 6500 |
|
}, |
|
{ |
|
"epoch": 0.19888503842097333, |
|
"grad_norm": 0.2216796875, |
|
"learning_rate": 0.0018688583648704348, |
|
"loss": 0.5741, |
|
"step": 6600 |
|
}, |
|
{ |
|
"epoch": 0.20009040229019134, |
|
"eval_peoplespeech-clean-transcription_loss": 4.054948806762695, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.5864, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.388, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.069, |
|
"step": 6640 |
|
}, |
|
{ |
|
"epoch": 0.20189844809401838, |
|
"grad_norm": 0.263671875, |
|
"learning_rate": 0.0018642553843161765, |
|
"loss": 0.5808, |
|
"step": 6700 |
|
}, |
|
{ |
|
"epoch": 0.20491185776706344, |
|
"grad_norm": 0.1962890625, |
|
"learning_rate": 0.0018595795876011011, |
|
"loss": 0.572, |
|
"step": 6800 |
|
}, |
|
{ |
|
"epoch": 0.20792526744010847, |
|
"grad_norm": 0.224609375, |
|
"learning_rate": 0.001854831420222319, |
|
"loss": 0.5738, |
|
"step": 6900 |
|
}, |
|
{ |
|
"epoch": 0.21093867711315353, |
|
"grad_norm": 0.18359375, |
|
"learning_rate": 0.001850011334572219, |
|
"loss": 0.5631, |
|
"step": 7000 |
|
}, |
|
{ |
|
"epoch": 0.2139520867861986, |
|
"grad_norm": 0.2001953125, |
|
"learning_rate": 0.0018451197898953675, |
|
"loss": 0.5656, |
|
"step": 7100 |
|
}, |
|
{ |
|
"epoch": 0.21696549645924362, |
|
"grad_norm": 0.1708984375, |
|
"learning_rate": 0.0018401572522447499, |
|
"loss": 0.5501, |
|
"step": 7200 |
|
}, |
|
{ |
|
"epoch": 0.21997890613228868, |
|
"grad_norm": 0.1845703125, |
|
"learning_rate": 0.0018351241944373684, |
|
"loss": 0.5487, |
|
"step": 7300 |
|
}, |
|
{ |
|
"epoch": 0.22299231580533374, |
|
"grad_norm": 0.14453125, |
|
"learning_rate": 0.0018300210960091926, |
|
"loss": 0.535, |
|
"step": 7400 |
|
}, |
|
{ |
|
"epoch": 0.2260057254783788, |
|
"grad_norm": 0.111328125, |
|
"learning_rate": 0.0018248484431694705, |
|
"loss": 0.5265, |
|
"step": 7500 |
|
}, |
|
{ |
|
"epoch": 0.22901913515142383, |
|
"grad_norm": 0.05126953125, |
|
"learning_rate": 0.0018196067287544043, |
|
"loss": 0.4819, |
|
"step": 7600 |
|
}, |
|
{ |
|
"epoch": 0.2320325448244689, |
|
"grad_norm": 0.04296875, |
|
"learning_rate": 0.0018142964521801936, |
|
"loss": 0.4168, |
|
"step": 7700 |
|
}, |
|
{ |
|
"epoch": 0.23504595449751395, |
|
"grad_norm": 0.033447265625, |
|
"learning_rate": 0.001808918119395454, |
|
"loss": 0.3548, |
|
"step": 7800 |
|
}, |
|
{ |
|
"epoch": 0.23805936417055898, |
|
"grad_norm": 0.03271484375, |
|
"learning_rate": 0.0018034722428330089, |
|
"loss": 0.3206, |
|
"step": 7900 |
|
}, |
|
{ |
|
"epoch": 0.24107277384360404, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.0017979593413610688, |
|
"loss": 0.3043, |
|
"step": 8000 |
|
}, |
|
{ |
|
"epoch": 0.2440861835166491, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0017923799402337944, |
|
"loss": 0.2899, |
|
"step": 8100 |
|
}, |
|
{ |
|
"epoch": 0.24709959318969413, |
|
"grad_norm": 0.028076171875, |
|
"learning_rate": 0.0017867345710412504, |
|
"loss": 0.2772, |
|
"step": 8200 |
|
}, |
|
{ |
|
"epoch": 0.2501130028627392, |
|
"grad_norm": 0.030029296875, |
|
"learning_rate": 0.00178102377165876, |
|
"loss": 0.2692, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.2501130028627392, |
|
"eval_peoplespeech-clean-transcription_loss": 1.788245677947998, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.7068, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.669, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.073, |
|
"step": 8300 |
|
}, |
|
{ |
|
"epoch": 0.2531264125357842, |
|
"grad_norm": 0.0262451171875, |
|
"learning_rate": 0.0017752480861956536, |
|
"loss": 0.2649, |
|
"step": 8400 |
|
}, |
|
{ |
|
"epoch": 0.2561398222088293, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0017694080649434314, |
|
"loss": 0.2574, |
|
"step": 8500 |
|
}, |
|
{ |
|
"epoch": 0.25915323188187434, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0017635042643233307, |
|
"loss": 0.2522, |
|
"step": 8600 |
|
}, |
|
{ |
|
"epoch": 0.26216664155491937, |
|
"grad_norm": 0.02880859375, |
|
"learning_rate": 0.0017575372468333127, |
|
"loss": 0.2487, |
|
"step": 8700 |
|
}, |
|
{ |
|
"epoch": 0.26518005122796445, |
|
"grad_norm": 0.0230712890625, |
|
"learning_rate": 0.001751507580994468, |
|
"loss": 0.2407, |
|
"step": 8800 |
|
}, |
|
{ |
|
"epoch": 0.2681934609010095, |
|
"grad_norm": 0.031005859375, |
|
"learning_rate": 0.0017454158412968522, |
|
"loss": 0.238, |
|
"step": 8900 |
|
}, |
|
{ |
|
"epoch": 0.27120687057405457, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0017392626081447465, |
|
"loss": 0.2347, |
|
"step": 9000 |
|
}, |
|
{ |
|
"epoch": 0.2742202802470996, |
|
"grad_norm": 0.0306396484375, |
|
"learning_rate": 0.0017330484678013609, |
|
"loss": 0.2343, |
|
"step": 9100 |
|
}, |
|
{ |
|
"epoch": 0.27723368992014463, |
|
"grad_norm": 0.0262451171875, |
|
"learning_rate": 0.0017267740123329753, |
|
"loss": 0.2324, |
|
"step": 9200 |
|
}, |
|
{ |
|
"epoch": 0.2802470995931897, |
|
"grad_norm": 0.02880859375, |
|
"learning_rate": 0.0017204398395525308, |
|
"loss": 0.2294, |
|
"step": 9300 |
|
}, |
|
{ |
|
"epoch": 0.28326050926623475, |
|
"grad_norm": 0.025390625, |
|
"learning_rate": 0.0017140465529626692, |
|
"loss": 0.2278, |
|
"step": 9400 |
|
}, |
|
{ |
|
"epoch": 0.2862739189392798, |
|
"grad_norm": 0.027587890625, |
|
"learning_rate": 0.0017075947616982349, |
|
"loss": 0.2247, |
|
"step": 9500 |
|
}, |
|
{ |
|
"epoch": 0.28928732861232487, |
|
"grad_norm": 0.024658203125, |
|
"learning_rate": 0.001701085080468237, |
|
"loss": 0.2204, |
|
"step": 9600 |
|
}, |
|
{ |
|
"epoch": 0.2923007382853699, |
|
"grad_norm": 0.030029296875, |
|
"learning_rate": 0.0016945181294972828, |
|
"loss": 0.2201, |
|
"step": 9700 |
|
}, |
|
{ |
|
"epoch": 0.29531414795841493, |
|
"grad_norm": 0.025634765625, |
|
"learning_rate": 0.0016878945344664831, |
|
"loss": 0.2196, |
|
"step": 9800 |
|
}, |
|
{ |
|
"epoch": 0.29832755763146, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.0016812149264538402, |
|
"loss": 0.2163, |
|
"step": 9900 |
|
}, |
|
{ |
|
"epoch": 0.300135603435287, |
|
"eval_peoplespeech-clean-transcription_loss": 1.6014918088912964, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.7355, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.659, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.073, |
|
"step": 9960 |
|
}, |
|
{ |
|
"epoch": 0.30134096730450505, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.0016744799418741193, |
|
"loss": 0.2143, |
|
"step": 10000 |
|
}, |
|
{ |
|
"epoch": 0.3043543769775501, |
|
"grad_norm": 0.027099609375, |
|
"learning_rate": 0.001667690222418214, |
|
"loss": 0.214, |
|
"step": 10100 |
|
}, |
|
{ |
|
"epoch": 0.30736778665059517, |
|
"grad_norm": 0.0260009765625, |
|
"learning_rate": 0.0016608464149920064, |
|
"loss": 0.2111, |
|
"step": 10200 |
|
}, |
|
{ |
|
"epoch": 0.3103811963236402, |
|
"grad_norm": 0.02734375, |
|
"learning_rate": 0.0016539491716547332, |
|
"loss": 0.2124, |
|
"step": 10300 |
|
}, |
|
{ |
|
"epoch": 0.3133946059966852, |
|
"grad_norm": 0.025146484375, |
|
"learning_rate": 0.0016469991495568573, |
|
"loss": 0.2071, |
|
"step": 10400 |
|
}, |
|
{ |
|
"epoch": 0.3164080156697303, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0016399970108774587, |
|
"loss": 0.2106, |
|
"step": 10500 |
|
}, |
|
{ |
|
"epoch": 0.31942142534277534, |
|
"grad_norm": 0.025634765625, |
|
"learning_rate": 0.001632943422761141, |
|
"loss": 0.2075, |
|
"step": 10600 |
|
}, |
|
{ |
|
"epoch": 0.3224348350158204, |
|
"grad_norm": 0.0269775390625, |
|
"learning_rate": 0.0016258390572544716, |
|
"loss": 0.2065, |
|
"step": 10700 |
|
}, |
|
{ |
|
"epoch": 0.32544824468886546, |
|
"grad_norm": 0.024169921875, |
|
"learning_rate": 0.001618684591241946, |
|
"loss": 0.2065, |
|
"step": 10800 |
|
}, |
|
{ |
|
"epoch": 0.3284616543619105, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0016114807063815008, |
|
"loss": 0.2055, |
|
"step": 10900 |
|
}, |
|
{ |
|
"epoch": 0.3314750640349556, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0016042280890395642, |
|
"loss": 0.2043, |
|
"step": 11000 |
|
}, |
|
{ |
|
"epoch": 0.3344884737080006, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.0015969274302256621, |
|
"loss": 0.2006, |
|
"step": 11100 |
|
}, |
|
{ |
|
"epoch": 0.33750188338104564, |
|
"grad_norm": 0.0245361328125, |
|
"learning_rate": 0.00158957942552658, |
|
"loss": 0.2021, |
|
"step": 11200 |
|
}, |
|
{ |
|
"epoch": 0.34051529305409073, |
|
"grad_norm": 0.02783203125, |
|
"learning_rate": 0.00158218477504009, |
|
"loss": 0.2042, |
|
"step": 11300 |
|
}, |
|
{ |
|
"epoch": 0.34352870272713576, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.0015747441833082476, |
|
"loss": 0.2043, |
|
"step": 11400 |
|
}, |
|
{ |
|
"epoch": 0.3465421124001808, |
|
"grad_norm": 0.0263671875, |
|
"learning_rate": 0.0015672583592502632, |
|
"loss": 0.1991, |
|
"step": 11500 |
|
}, |
|
{ |
|
"epoch": 0.3495555220732259, |
|
"grad_norm": 0.0281982421875, |
|
"learning_rate": 0.0015597280160949602, |
|
"loss": 0.1994, |
|
"step": 11600 |
|
}, |
|
{ |
|
"epoch": 0.3501582040078349, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5406583547592163, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.5635, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.395, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.069, |
|
"step": 11620 |
|
}, |
|
{ |
|
"epoch": 0.3525689317462709, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.0015521538713128204, |
|
"loss": 0.2, |
|
"step": 11700 |
|
}, |
|
{ |
|
"epoch": 0.35558234141931594, |
|
"grad_norm": 0.0244140625, |
|
"learning_rate": 0.001544536646547623, |
|
"loss": 0.1978, |
|
"step": 11800 |
|
}, |
|
{ |
|
"epoch": 0.358595751092361, |
|
"grad_norm": 0.0255126953125, |
|
"learning_rate": 0.0015368770675476915, |
|
"loss": 0.1974, |
|
"step": 11900 |
|
}, |
|
{ |
|
"epoch": 0.36160916076540606, |
|
"grad_norm": 0.025390625, |
|
"learning_rate": 0.001529175864096744, |
|
"loss": 0.1963, |
|
"step": 12000 |
|
}, |
|
{ |
|
"epoch": 0.3646225704384511, |
|
"grad_norm": 0.027587890625, |
|
"learning_rate": 0.0015214337699443632, |
|
"loss": 0.1958, |
|
"step": 12100 |
|
}, |
|
{ |
|
"epoch": 0.3676359801114962, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0015136515227360855, |
|
"loss": 0.1974, |
|
"step": 12200 |
|
}, |
|
{ |
|
"epoch": 0.3706493897845412, |
|
"grad_norm": 0.029052734375, |
|
"learning_rate": 0.0015058298639431193, |
|
"loss": 0.1974, |
|
"step": 12300 |
|
}, |
|
{ |
|
"epoch": 0.37366279945758624, |
|
"grad_norm": 0.0245361328125, |
|
"learning_rate": 0.0014979695387917036, |
|
"loss": 0.1924, |
|
"step": 12400 |
|
}, |
|
{ |
|
"epoch": 0.3766762091306313, |
|
"grad_norm": 0.024169921875, |
|
"learning_rate": 0.0014900712961920999, |
|
"loss": 0.1925, |
|
"step": 12500 |
|
}, |
|
{ |
|
"epoch": 0.37968961880367635, |
|
"grad_norm": 0.021484375, |
|
"learning_rate": 0.0014821358886672414, |
|
"loss": 0.1935, |
|
"step": 12600 |
|
}, |
|
{ |
|
"epoch": 0.3827030284767214, |
|
"grad_norm": 0.0252685546875, |
|
"learning_rate": 0.0014741640722810332, |
|
"loss": 0.1925, |
|
"step": 12700 |
|
}, |
|
{ |
|
"epoch": 0.38571643814976647, |
|
"grad_norm": 0.024658203125, |
|
"learning_rate": 0.0014661566065663168, |
|
"loss": 0.1936, |
|
"step": 12800 |
|
}, |
|
{ |
|
"epoch": 0.3887298478228115, |
|
"grad_norm": 0.0263671875, |
|
"learning_rate": 0.0014581142544525052, |
|
"loss": 0.1928, |
|
"step": 12900 |
|
}, |
|
{ |
|
"epoch": 0.3917432574958566, |
|
"grad_norm": 0.025390625, |
|
"learning_rate": 0.0014500377821928911, |
|
"loss": 0.1927, |
|
"step": 13000 |
|
}, |
|
{ |
|
"epoch": 0.3947566671689016, |
|
"grad_norm": 0.0245361328125, |
|
"learning_rate": 0.0014419279592916417, |
|
"loss": 0.1931, |
|
"step": 13100 |
|
}, |
|
{ |
|
"epoch": 0.39777007684194665, |
|
"grad_norm": 0.025390625, |
|
"learning_rate": 0.001433785558430481, |
|
"loss": 0.1903, |
|
"step": 13200 |
|
}, |
|
{ |
|
"epoch": 0.4001808045803827, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5337910652160645, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.9866, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.27, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.067, |
|
"step": 13280 |
|
}, |
|
{ |
|
"epoch": 0.40078348651499174, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.0014256113553950739, |
|
"loss": 0.1917, |
|
"step": 13300 |
|
}, |
|
{ |
|
"epoch": 0.40379689618803677, |
|
"grad_norm": 0.023681640625, |
|
"learning_rate": 0.0014174061290011075, |
|
"loss": 0.1893, |
|
"step": 13400 |
|
}, |
|
{ |
|
"epoch": 0.4068103058610818, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.0014091706610200902, |
|
"loss": 0.1909, |
|
"step": 13500 |
|
}, |
|
{ |
|
"epoch": 0.4098237155341269, |
|
"grad_norm": 0.02294921875, |
|
"learning_rate": 0.0014009057361048665, |
|
"loss": 0.19, |
|
"step": 13600 |
|
}, |
|
{ |
|
"epoch": 0.4128371252071719, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.001392612141714856, |
|
"loss": 0.1913, |
|
"step": 13700 |
|
}, |
|
{ |
|
"epoch": 0.41585053488021695, |
|
"grad_norm": 0.02294921875, |
|
"learning_rate": 0.0013842906680410286, |
|
"loss": 0.1898, |
|
"step": 13800 |
|
}, |
|
{ |
|
"epoch": 0.41886394455326204, |
|
"grad_norm": 0.026611328125, |
|
"learning_rate": 0.0013759421079306145, |
|
"loss": 0.1892, |
|
"step": 13900 |
|
}, |
|
{ |
|
"epoch": 0.42187735422630707, |
|
"grad_norm": 0.0252685546875, |
|
"learning_rate": 0.001367567256811567, |
|
"loss": 0.1893, |
|
"step": 14000 |
|
}, |
|
{ |
|
"epoch": 0.4248907638993521, |
|
"grad_norm": 0.02783203125, |
|
"learning_rate": 0.0013591669126167736, |
|
"loss": 0.1898, |
|
"step": 14100 |
|
}, |
|
{ |
|
"epoch": 0.4279041735723972, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.001350741875708033, |
|
"loss": 0.1874, |
|
"step": 14200 |
|
}, |
|
{ |
|
"epoch": 0.4309175832454422, |
|
"grad_norm": 0.029541015625, |
|
"learning_rate": 0.0013422929487997973, |
|
"loss": 0.188, |
|
"step": 14300 |
|
}, |
|
{ |
|
"epoch": 0.43393099291848725, |
|
"grad_norm": 0.0252685546875, |
|
"learning_rate": 0.0013338209368826933, |
|
"loss": 0.1879, |
|
"step": 14400 |
|
}, |
|
{ |
|
"epoch": 0.43694440259153233, |
|
"grad_norm": 0.0260009765625, |
|
"learning_rate": 0.0013253266471468235, |
|
"loss": 0.1865, |
|
"step": 14500 |
|
}, |
|
{ |
|
"epoch": 0.43995781226457736, |
|
"grad_norm": 0.0213623046875, |
|
"learning_rate": 0.0013168108889048602, |
|
"loss": 0.1859, |
|
"step": 14600 |
|
}, |
|
{ |
|
"epoch": 0.4429712219376224, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.0013082744735149366, |
|
"loss": 0.1872, |
|
"step": 14700 |
|
}, |
|
{ |
|
"epoch": 0.4459846316106675, |
|
"grad_norm": 0.0223388671875, |
|
"learning_rate": 0.0012997182143033416, |
|
"loss": 0.1867, |
|
"step": 14800 |
|
}, |
|
{ |
|
"epoch": 0.4489980412837125, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.00129114292648703, |
|
"loss": 0.1867, |
|
"step": 14900 |
|
}, |
|
{ |
|
"epoch": 0.45020340515293056, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5441259145736694, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.9015, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.295, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.067, |
|
"step": 14940 |
|
}, |
|
{ |
|
"epoch": 0.4520114509567576, |
|
"grad_norm": 0.0211181640625, |
|
"learning_rate": 0.001282549427095949, |
|
"loss": 0.1866, |
|
"step": 15000 |
|
}, |
|
{ |
|
"epoch": 0.45502486062980263, |
|
"grad_norm": 0.029541015625, |
|
"learning_rate": 0.0012739385348951955, |
|
"loss": 0.1852, |
|
"step": 15100 |
|
}, |
|
{ |
|
"epoch": 0.45803827030284766, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0012653110703070055, |
|
"loss": 0.1849, |
|
"step": 15200 |
|
}, |
|
{ |
|
"epoch": 0.46105167997589275, |
|
"grad_norm": 0.030029296875, |
|
"learning_rate": 0.001256667855332587, |
|
"loss": 0.1846, |
|
"step": 15300 |
|
}, |
|
{ |
|
"epoch": 0.4640650896489378, |
|
"grad_norm": 0.0225830078125, |
|
"learning_rate": 0.0012480097134738009, |
|
"loss": 0.185, |
|
"step": 15400 |
|
}, |
|
{ |
|
"epoch": 0.4670784993219828, |
|
"grad_norm": 0.02490234375, |
|
"learning_rate": 0.0012393374696547015, |
|
"loss": 0.1861, |
|
"step": 15500 |
|
}, |
|
{ |
|
"epoch": 0.4700919089950279, |
|
"grad_norm": 0.0223388671875, |
|
"learning_rate": 0.0012306519501429395, |
|
"loss": 0.1877, |
|
"step": 15600 |
|
}, |
|
{ |
|
"epoch": 0.4731053186680729, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0012219539824710357, |
|
"loss": 0.1859, |
|
"step": 15700 |
|
}, |
|
{ |
|
"epoch": 0.47611872834111796, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.0012132443953575397, |
|
"loss": 0.1847, |
|
"step": 15800 |
|
}, |
|
{ |
|
"epoch": 0.47913213801416304, |
|
"grad_norm": 0.0267333984375, |
|
"learning_rate": 0.0012045240186280676, |
|
"loss": 0.1853, |
|
"step": 15900 |
|
}, |
|
{ |
|
"epoch": 0.4821455476872081, |
|
"grad_norm": 0.0211181640625, |
|
"learning_rate": 0.0011957936831362426, |
|
"loss": 0.185, |
|
"step": 16000 |
|
}, |
|
{ |
|
"epoch": 0.4851589573602531, |
|
"grad_norm": 0.029541015625, |
|
"learning_rate": 0.0011870542206845298, |
|
"loss": 0.1849, |
|
"step": 16100 |
|
}, |
|
{ |
|
"epoch": 0.4881723670332982, |
|
"grad_norm": 0.0213623046875, |
|
"learning_rate": 0.001178306463944987, |
|
"loss": 0.1835, |
|
"step": 16200 |
|
}, |
|
{ |
|
"epoch": 0.4911857767063432, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0011695512463799286, |
|
"loss": 0.1837, |
|
"step": 16300 |
|
}, |
|
{ |
|
"epoch": 0.49419918637938826, |
|
"grad_norm": 0.0234375, |
|
"learning_rate": 0.0011607894021625166, |
|
"loss": 0.1847, |
|
"step": 16400 |
|
}, |
|
{ |
|
"epoch": 0.49721259605243334, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0011520217660972811, |
|
"loss": 0.1853, |
|
"step": 16500 |
|
}, |
|
{ |
|
"epoch": 0.5002260057254784, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.0011432491735405852, |
|
"loss": 0.1827, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.5002260057254784, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5211623907089233, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.7246, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.346, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.068, |
|
"step": 16600 |
|
}, |
|
{ |
|
"epoch": 0.5032394153985235, |
|
"grad_norm": 0.0291748046875, |
|
"learning_rate": 0.0011344724603210318, |
|
"loss": 0.1818, |
|
"step": 16700 |
|
}, |
|
{ |
|
"epoch": 0.5062528250715684, |
|
"grad_norm": 0.025146484375, |
|
"learning_rate": 0.0011256924626598297, |
|
"loss": 0.1831, |
|
"step": 16800 |
|
}, |
|
{ |
|
"epoch": 0.5092662347446135, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0011169100170911204, |
|
"loss": 0.184, |
|
"step": 16900 |
|
}, |
|
{ |
|
"epoch": 0.5122796444176586, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0011081259603822747, |
|
"loss": 0.1833, |
|
"step": 17000 |
|
}, |
|
{ |
|
"epoch": 0.5152930540907036, |
|
"grad_norm": 0.0294189453125, |
|
"learning_rate": 0.0010993411294541694, |
|
"loss": 0.1841, |
|
"step": 17100 |
|
}, |
|
{ |
|
"epoch": 0.5183064637637487, |
|
"grad_norm": 0.0223388671875, |
|
"learning_rate": 0.001090556361301446, |
|
"loss": 0.1849, |
|
"step": 17200 |
|
}, |
|
{ |
|
"epoch": 0.5213198734367938, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.0010817724929127646, |
|
"loss": 0.1831, |
|
"step": 17300 |
|
}, |
|
{ |
|
"epoch": 0.5243332831098387, |
|
"grad_norm": 0.021484375, |
|
"learning_rate": 0.00107299036119106, |
|
"loss": 0.1822, |
|
"step": 17400 |
|
}, |
|
{ |
|
"epoch": 0.5273466927828838, |
|
"grad_norm": 0.02880859375, |
|
"learning_rate": 0.0010642108028738003, |
|
"loss": 0.1819, |
|
"step": 17500 |
|
}, |
|
{ |
|
"epoch": 0.5303601024559289, |
|
"grad_norm": 0.021728515625, |
|
"learning_rate": 0.0010554346544532672, |
|
"loss": 0.1839, |
|
"step": 17600 |
|
}, |
|
{ |
|
"epoch": 0.5333735121289739, |
|
"grad_norm": 0.031494140625, |
|
"learning_rate": 0.0010466627520968577, |
|
"loss": 0.1858, |
|
"step": 17700 |
|
}, |
|
{ |
|
"epoch": 0.536386921802019, |
|
"grad_norm": 0.02197265625, |
|
"learning_rate": 0.001037895931567414, |
|
"loss": 0.1837, |
|
"step": 17800 |
|
}, |
|
{ |
|
"epoch": 0.539400331475064, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0010291350281435962, |
|
"loss": 0.1819, |
|
"step": 17900 |
|
}, |
|
{ |
|
"epoch": 0.5424137411481091, |
|
"grad_norm": 0.024169921875, |
|
"learning_rate": 0.0010203808765402993, |
|
"loss": 0.1835, |
|
"step": 18000 |
|
}, |
|
{ |
|
"epoch": 0.5454271508211541, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0010116343108291233, |
|
"loss": 0.1828, |
|
"step": 18100 |
|
}, |
|
{ |
|
"epoch": 0.5484405604941992, |
|
"grad_norm": 0.0234375, |
|
"learning_rate": 0.0010028961643589044, |
|
"loss": 0.1802, |
|
"step": 18200 |
|
}, |
|
{ |
|
"epoch": 0.5502486062980262, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5098843574523926, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.7248, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.346, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.068, |
|
"step": 18260 |
|
}, |
|
{ |
|
"epoch": 0.5514539701672443, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.0009941672696763173, |
|
"loss": 0.1835, |
|
"step": 18300 |
|
}, |
|
{ |
|
"epoch": 0.5544673798402893, |
|
"grad_norm": 0.0255126953125, |
|
"learning_rate": 0.0009854484584465506, |
|
"loss": 0.1815, |
|
"step": 18400 |
|
}, |
|
{ |
|
"epoch": 0.5574807895133344, |
|
"grad_norm": 0.03076171875, |
|
"learning_rate": 0.0009767405613740716, |
|
"loss": 0.1817, |
|
"step": 18500 |
|
}, |
|
{ |
|
"epoch": 0.5604941991863794, |
|
"grad_norm": 0.0230712890625, |
|
"learning_rate": 0.0009680444081234734, |
|
"loss": 0.1822, |
|
"step": 18600 |
|
}, |
|
{ |
|
"epoch": 0.5635076088594244, |
|
"grad_norm": 0.0299072265625, |
|
"learning_rate": 0.0009593608272404317, |
|
"loss": 0.183, |
|
"step": 18700 |
|
}, |
|
{ |
|
"epoch": 0.5665210185324695, |
|
"grad_norm": 0.0213623046875, |
|
"learning_rate": 0.0009506906460727618, |
|
"loss": 0.1813, |
|
"step": 18800 |
|
}, |
|
{ |
|
"epoch": 0.5695344282055146, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.0009420346906915895, |
|
"loss": 0.1826, |
|
"step": 18900 |
|
}, |
|
{ |
|
"epoch": 0.5725478378785596, |
|
"grad_norm": 0.021728515625, |
|
"learning_rate": 0.0009333937858126477, |
|
"loss": 0.1799, |
|
"step": 19000 |
|
}, |
|
{ |
|
"epoch": 0.5755612475516046, |
|
"grad_norm": 0.0303955078125, |
|
"learning_rate": 0.0009247687547176979, |
|
"loss": 0.1819, |
|
"step": 19100 |
|
}, |
|
{ |
|
"epoch": 0.5785746572246497, |
|
"grad_norm": 0.021728515625, |
|
"learning_rate": 0.0009161604191760915, |
|
"loss": 0.1804, |
|
"step": 19200 |
|
}, |
|
{ |
|
"epoch": 0.5815880668976947, |
|
"grad_norm": 0.028564453125, |
|
"learning_rate": 0.000907569599366473, |
|
"loss": 0.181, |
|
"step": 19300 |
|
}, |
|
{ |
|
"epoch": 0.5846014765707398, |
|
"grad_norm": 0.0223388671875, |
|
"learning_rate": 0.000898997113798635, |
|
"loss": 0.1798, |
|
"step": 19400 |
|
}, |
|
{ |
|
"epoch": 0.5876148862437849, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.0008904437792355364, |
|
"loss": 0.179, |
|
"step": 19500 |
|
}, |
|
{ |
|
"epoch": 0.5906282959168299, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.0008819104106154776, |
|
"loss": 0.1808, |
|
"step": 19600 |
|
}, |
|
{ |
|
"epoch": 0.593641705589875, |
|
"grad_norm": 0.028076171875, |
|
"learning_rate": 0.0008733978209744609, |
|
"loss": 0.1802, |
|
"step": 19700 |
|
}, |
|
{ |
|
"epoch": 0.59665511526292, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0008649068213687225, |
|
"loss": 0.1813, |
|
"step": 19800 |
|
}, |
|
{ |
|
"epoch": 0.599668524935965, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.0008564382207974612, |
|
"loss": 0.1807, |
|
"step": 19900 |
|
}, |
|
{ |
|
"epoch": 0.600271206870574, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5082225799560547, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.0089, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.569, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.071, |
|
"step": 19920 |
|
}, |
|
{ |
|
"epoch": 0.6026819346090101, |
|
"grad_norm": 0.0234375, |
|
"learning_rate": 0.0008479928261257557, |
|
"loss": 0.1807, |
|
"step": 20000 |
|
}, |
|
{ |
|
"epoch": 0.6056953442820552, |
|
"grad_norm": 0.0284423828125, |
|
"learning_rate": 0.0008395714420076905, |
|
"loss": 0.1813, |
|
"step": 20100 |
|
}, |
|
{ |
|
"epoch": 0.6087087539551002, |
|
"grad_norm": 0.0223388671875, |
|
"learning_rate": 0.0008311748708096898, |
|
"loss": 0.1794, |
|
"step": 20200 |
|
}, |
|
{ |
|
"epoch": 0.6117221636281452, |
|
"grad_norm": 0.033935546875, |
|
"learning_rate": 0.0008228039125340721, |
|
"loss": 0.1809, |
|
"step": 20300 |
|
}, |
|
{ |
|
"epoch": 0.6147355733011903, |
|
"grad_norm": 0.0234375, |
|
"learning_rate": 0.0008144593647428254, |
|
"loss": 0.1796, |
|
"step": 20400 |
|
}, |
|
{ |
|
"epoch": 0.6177489829742353, |
|
"grad_norm": 0.0281982421875, |
|
"learning_rate": 0.0008061420224816187, |
|
"loss": 0.1807, |
|
"step": 20500 |
|
}, |
|
{ |
|
"epoch": 0.6207623926472804, |
|
"grad_norm": 0.0242919921875, |
|
"learning_rate": 0.0007978526782040547, |
|
"loss": 0.1806, |
|
"step": 20600 |
|
}, |
|
{ |
|
"epoch": 0.6237758023203255, |
|
"grad_norm": 0.0308837890625, |
|
"learning_rate": 0.0007895921216961628, |
|
"loss": 0.1802, |
|
"step": 20700 |
|
}, |
|
{ |
|
"epoch": 0.6267892119933705, |
|
"grad_norm": 0.0244140625, |
|
"learning_rate": 0.0007813611400011535, |
|
"loss": 0.1806, |
|
"step": 20800 |
|
}, |
|
{ |
|
"epoch": 0.6298026216664155, |
|
"grad_norm": 0.0322265625, |
|
"learning_rate": 0.0007731605173444294, |
|
"loss": 0.1799, |
|
"step": 20900 |
|
}, |
|
{ |
|
"epoch": 0.6328160313394606, |
|
"grad_norm": 0.0228271484375, |
|
"learning_rate": 0.0007649910350588683, |
|
"loss": 0.1797, |
|
"step": 21000 |
|
}, |
|
{ |
|
"epoch": 0.6358294410125056, |
|
"grad_norm": 0.025146484375, |
|
"learning_rate": 0.000756853471510377, |
|
"loss": 0.1779, |
|
"step": 21100 |
|
}, |
|
{ |
|
"epoch": 0.6388428506855507, |
|
"grad_norm": 0.0240478515625, |
|
"learning_rate": 0.0007487486020237337, |
|
"loss": 0.1786, |
|
"step": 21200 |
|
}, |
|
{ |
|
"epoch": 0.6418562603585958, |
|
"grad_norm": 0.0302734375, |
|
"learning_rate": 0.0007406771988087153, |
|
"loss": 0.1795, |
|
"step": 21300 |
|
}, |
|
{ |
|
"epoch": 0.6448696700316408, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.0007326400308865245, |
|
"loss": 0.1827, |
|
"step": 21400 |
|
}, |
|
{ |
|
"epoch": 0.6478830797046858, |
|
"grad_norm": 0.02978515625, |
|
"learning_rate": 0.0007246378640165184, |
|
"loss": 0.1797, |
|
"step": 21500 |
|
}, |
|
{ |
|
"epoch": 0.6502938074431219, |
|
"eval_peoplespeech-clean-transcription_loss": 1.4861868619918823, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.4228, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.437, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.069, |
|
"step": 21580 |
|
}, |
|
{ |
|
"epoch": 0.6508964893777309, |
|
"grad_norm": 0.0218505859375, |
|
"learning_rate": 0.0007166714606232492, |
|
"loss": 0.1812, |
|
"step": 21600 |
|
}, |
|
{ |
|
"epoch": 0.6539098990507759, |
|
"grad_norm": 0.0291748046875, |
|
"learning_rate": 0.0007087415797238248, |
|
"loss": 0.1826, |
|
"step": 21700 |
|
}, |
|
{ |
|
"epoch": 0.656923308723821, |
|
"grad_norm": 0.0260009765625, |
|
"learning_rate": 0.0007008489768555886, |
|
"loss": 0.18, |
|
"step": 21800 |
|
}, |
|
{ |
|
"epoch": 0.6599367183968661, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.0006929944040041347, |
|
"loss": 0.1786, |
|
"step": 21900 |
|
}, |
|
{ |
|
"epoch": 0.6629501280699112, |
|
"grad_norm": 0.02294921875, |
|
"learning_rate": 0.0006851786095316618, |
|
"loss": 0.1805, |
|
"step": 22000 |
|
}, |
|
{ |
|
"epoch": 0.6659635377429561, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.000677402338105672, |
|
"loss": 0.1786, |
|
"step": 22100 |
|
}, |
|
{ |
|
"epoch": 0.6689769474160012, |
|
"grad_norm": 0.024658203125, |
|
"learning_rate": 0.0006696663306280182, |
|
"loss": 0.181, |
|
"step": 22200 |
|
}, |
|
{ |
|
"epoch": 0.6719903570890463, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0006619713241643147, |
|
"loss": 0.1797, |
|
"step": 22300 |
|
}, |
|
{ |
|
"epoch": 0.6750037667620913, |
|
"grad_norm": 0.0263671875, |
|
"learning_rate": 0.0006543180518737122, |
|
"loss": 0.1793, |
|
"step": 22400 |
|
}, |
|
{ |
|
"epoch": 0.6780171764351364, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.0006467072429390431, |
|
"loss": 0.178, |
|
"step": 22500 |
|
}, |
|
{ |
|
"epoch": 0.6810305861081815, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.0006391396224973473, |
|
"loss": 0.1793, |
|
"step": 22600 |
|
}, |
|
{ |
|
"epoch": 0.6840439957812264, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0006316159115707838, |
|
"loss": 0.1793, |
|
"step": 22700 |
|
}, |
|
{ |
|
"epoch": 0.6870574054542715, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0006241368269979337, |
|
"loss": 0.177, |
|
"step": 22800 |
|
}, |
|
{ |
|
"epoch": 0.6900708151273166, |
|
"grad_norm": 0.028076171875, |
|
"learning_rate": 0.0006167030813654996, |
|
"loss": 0.1781, |
|
"step": 22900 |
|
}, |
|
{ |
|
"epoch": 0.6930842248003616, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.0006093153829404155, |
|
"loss": 0.1782, |
|
"step": 23000 |
|
}, |
|
{ |
|
"epoch": 0.6960976344734067, |
|
"grad_norm": 0.0296630859375, |
|
"learning_rate": 0.0006019744356023627, |
|
"loss": 0.179, |
|
"step": 23100 |
|
}, |
|
{ |
|
"epoch": 0.6991110441464518, |
|
"grad_norm": 0.0250244140625, |
|
"learning_rate": 0.0005946809387767075, |
|
"loss": 0.1788, |
|
"step": 23200 |
|
}, |
|
{ |
|
"epoch": 0.7003164080156697, |
|
"eval_peoplespeech-clean-transcription_loss": 1.503227710723877, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 14.6551, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.367, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.068, |
|
"step": 23240 |
|
}, |
|
{ |
|
"epoch": 0.7021244538194967, |
|
"grad_norm": 0.03125, |
|
"learning_rate": 0.000587435587367861, |
|
"loss": 0.1789, |
|
"step": 23300 |
|
}, |
|
{ |
|
"epoch": 0.7051378634925418, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0005802390716930713, |
|
"loss": 0.1771, |
|
"step": 23400 |
|
}, |
|
{ |
|
"epoch": 0.7081512731655869, |
|
"grad_norm": 0.0311279296875, |
|
"learning_rate": 0.0005730920774166495, |
|
"loss": 0.1793, |
|
"step": 23500 |
|
}, |
|
{ |
|
"epoch": 0.7111646828386319, |
|
"grad_norm": 0.02587890625, |
|
"learning_rate": 0.0005659952854846461, |
|
"loss": 0.1773, |
|
"step": 23600 |
|
}, |
|
{ |
|
"epoch": 0.714178092511677, |
|
"grad_norm": 0.0299072265625, |
|
"learning_rate": 0.0005589493720599683, |
|
"loss": 0.1785, |
|
"step": 23700 |
|
}, |
|
{ |
|
"epoch": 0.717191502184722, |
|
"grad_norm": 0.02587890625, |
|
"learning_rate": 0.0005519550084579583, |
|
"loss": 0.1776, |
|
"step": 23800 |
|
}, |
|
{ |
|
"epoch": 0.720204911857767, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.0005450128610824328, |
|
"loss": 0.179, |
|
"step": 23900 |
|
}, |
|
{ |
|
"epoch": 0.7232183215308121, |
|
"grad_norm": 0.0224609375, |
|
"learning_rate": 0.0005381235913621889, |
|
"loss": 0.1779, |
|
"step": 24000 |
|
}, |
|
{ |
|
"epoch": 0.7262317312038572, |
|
"grad_norm": 0.02783203125, |
|
"learning_rate": 0.0005312878556879856, |
|
"loss": 0.1776, |
|
"step": 24100 |
|
}, |
|
{ |
|
"epoch": 0.7292451408769022, |
|
"grad_norm": 0.0267333984375, |
|
"learning_rate": 0.0005245063053500047, |
|
"loss": 0.1796, |
|
"step": 24200 |
|
}, |
|
{ |
|
"epoch": 0.7322585505499473, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.0005177795864757979, |
|
"loss": 0.179, |
|
"step": 24300 |
|
}, |
|
{ |
|
"epoch": 0.7352719602229923, |
|
"grad_norm": 0.024169921875, |
|
"learning_rate": 0.0005111083399687246, |
|
"loss": 0.179, |
|
"step": 24400 |
|
}, |
|
{ |
|
"epoch": 0.7382853698960373, |
|
"grad_norm": 0.0264892578125, |
|
"learning_rate": 0.0005044932014468884, |
|
"loss": 0.178, |
|
"step": 24500 |
|
}, |
|
{ |
|
"epoch": 0.7412987795690824, |
|
"grad_norm": 0.0238037109375, |
|
"learning_rate": 0.0004979348011825788, |
|
"loss": 0.1797, |
|
"step": 24600 |
|
}, |
|
{ |
|
"epoch": 0.7443121892421275, |
|
"grad_norm": 0.0264892578125, |
|
"learning_rate": 0.000491433764042219, |
|
"loss": 0.1793, |
|
"step": 24700 |
|
}, |
|
{ |
|
"epoch": 0.7473255989151725, |
|
"grad_norm": 0.025390625, |
|
"learning_rate": 0.0004849907094268304, |
|
"loss": 0.1788, |
|
"step": 24800 |
|
}, |
|
{ |
|
"epoch": 0.7503390085882176, |
|
"grad_norm": 0.02490234375, |
|
"learning_rate": 0.0004786062512130186, |
|
"loss": 0.1784, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.7503390085882176, |
|
"eval_peoplespeech-clean-transcription_loss": 1.5016210079193115, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.8104, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.634, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.072, |
|
"step": 24900 |
|
}, |
|
{ |
|
"epoch": 0.7533524182612626, |
|
"grad_norm": 0.0234375, |
|
"learning_rate": 0.00047228099769448437, |
|
"loss": 0.1769, |
|
"step": 25000 |
|
}, |
|
{ |
|
"epoch": 0.7563658279343076, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.00046601555152406694, |
|
"loss": 0.1781, |
|
"step": 25100 |
|
}, |
|
{ |
|
"epoch": 0.7593792376073527, |
|
"grad_norm": 0.0238037109375, |
|
"learning_rate": 0.0004598105096563256, |
|
"loss": 0.1773, |
|
"step": 25200 |
|
}, |
|
{ |
|
"epoch": 0.7623926472803978, |
|
"grad_norm": 0.0262451171875, |
|
"learning_rate": 0.00045366646329066243, |
|
"loss": 0.1782, |
|
"step": 25300 |
|
}, |
|
{ |
|
"epoch": 0.7654060569534428, |
|
"grad_norm": 0.0242919921875, |
|
"learning_rate": 0.0004475839978149959, |
|
"loss": 0.1768, |
|
"step": 25400 |
|
}, |
|
{ |
|
"epoch": 0.7684194666264879, |
|
"grad_norm": 0.0269775390625, |
|
"learning_rate": 0.00044156369274998554, |
|
"loss": 0.1776, |
|
"step": 25500 |
|
}, |
|
{ |
|
"epoch": 0.7714328762995329, |
|
"grad_norm": 0.0262451171875, |
|
"learning_rate": 0.00043560612169381583, |
|
"loss": 0.1763, |
|
"step": 25600 |
|
}, |
|
{ |
|
"epoch": 0.7744462859725779, |
|
"grad_norm": 0.0260009765625, |
|
"learning_rate": 0.00042971185226754895, |
|
"loss": 0.1775, |
|
"step": 25700 |
|
}, |
|
{ |
|
"epoch": 0.777459695645623, |
|
"grad_norm": 0.0252685546875, |
|
"learning_rate": 0.00042388144606103926, |
|
"loss": 0.1791, |
|
"step": 25800 |
|
}, |
|
{ |
|
"epoch": 0.7804731053186681, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.00041811545857942936, |
|
"loss": 0.1802, |
|
"step": 25900 |
|
}, |
|
{ |
|
"epoch": 0.7834865149917132, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.00041241443919022124, |
|
"loss": 0.1768, |
|
"step": 26000 |
|
}, |
|
{ |
|
"epoch": 0.7864999246647582, |
|
"grad_norm": 0.02978515625, |
|
"learning_rate": 0.0004067789310709359, |
|
"loss": 0.181, |
|
"step": 26100 |
|
}, |
|
{ |
|
"epoch": 0.7895133343378032, |
|
"grad_norm": 0.02490234375, |
|
"learning_rate": 0.0004012094711573591, |
|
"loss": 0.1794, |
|
"step": 26200 |
|
}, |
|
{ |
|
"epoch": 0.7925267440108483, |
|
"grad_norm": 0.027587890625, |
|
"learning_rate": 0.0003957065900923845, |
|
"loss": 0.1784, |
|
"step": 26300 |
|
}, |
|
{ |
|
"epoch": 0.7955401536838933, |
|
"grad_norm": 0.0245361328125, |
|
"learning_rate": 0.00039027081217545554, |
|
"loss": 0.1777, |
|
"step": 26400 |
|
}, |
|
{ |
|
"epoch": 0.7985535633569384, |
|
"grad_norm": 0.02880859375, |
|
"learning_rate": 0.0003849026553126118, |
|
"loss": 0.1762, |
|
"step": 26500 |
|
}, |
|
{ |
|
"epoch": 0.8003616091607654, |
|
"eval_peoplespeech-clean-transcription_loss": 1.4861080646514893, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.7251, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.663, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.073, |
|
"step": 26560 |
|
}, |
|
{ |
|
"epoch": 0.8015669730299835, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.0003796026309671429, |
|
"loss": 0.1792, |
|
"step": 26600 |
|
}, |
|
{ |
|
"epoch": 0.8045803827030285, |
|
"grad_norm": 0.0267333984375, |
|
"learning_rate": 0.0003743712441108592, |
|
"loss": 0.1795, |
|
"step": 26700 |
|
}, |
|
{ |
|
"epoch": 0.8075937923760735, |
|
"grad_norm": 0.0250244140625, |
|
"learning_rate": 0.00036920899317597976, |
|
"loss": 0.1764, |
|
"step": 26800 |
|
}, |
|
{ |
|
"epoch": 0.8106072020491186, |
|
"grad_norm": 0.0269775390625, |
|
"learning_rate": 0.00036411637000764133, |
|
"loss": 0.18, |
|
"step": 26900 |
|
}, |
|
{ |
|
"epoch": 0.8136206117221636, |
|
"grad_norm": 0.0240478515625, |
|
"learning_rate": 0.00035909385981703777, |
|
"loss": 0.1783, |
|
"step": 27000 |
|
}, |
|
{ |
|
"epoch": 0.8166340213952087, |
|
"grad_norm": 0.02587890625, |
|
"learning_rate": 0.0003541419411351909, |
|
"loss": 0.1801, |
|
"step": 27100 |
|
}, |
|
{ |
|
"epoch": 0.8196474310682538, |
|
"grad_norm": 0.02734375, |
|
"learning_rate": 0.0003492610857673564, |
|
"loss": 0.1781, |
|
"step": 27200 |
|
}, |
|
{ |
|
"epoch": 0.8226608407412987, |
|
"grad_norm": 0.0263671875, |
|
"learning_rate": 0.000344451758748072, |
|
"loss": 0.1772, |
|
"step": 27300 |
|
}, |
|
{ |
|
"epoch": 0.8256742504143438, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.00033971441829685036, |
|
"loss": 0.1762, |
|
"step": 27400 |
|
}, |
|
{ |
|
"epoch": 0.8286876600873889, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.0003350495157745207, |
|
"loss": 0.1785, |
|
"step": 27500 |
|
}, |
|
{ |
|
"epoch": 0.8317010697604339, |
|
"grad_norm": 0.0245361328125, |
|
"learning_rate": 0.00033045749564022497, |
|
"loss": 0.1778, |
|
"step": 27600 |
|
}, |
|
{ |
|
"epoch": 0.834714479433479, |
|
"grad_norm": 0.0289306640625, |
|
"learning_rate": 0.00032593879540907076, |
|
"loss": 0.1773, |
|
"step": 27700 |
|
}, |
|
{ |
|
"epoch": 0.8377278891065241, |
|
"grad_norm": 0.022705078125, |
|
"learning_rate": 0.0003214938456104454, |
|
"loss": 0.1785, |
|
"step": 27800 |
|
}, |
|
{ |
|
"epoch": 0.840741298779569, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.0003171230697469978, |
|
"loss": 0.179, |
|
"step": 27900 |
|
}, |
|
{ |
|
"epoch": 0.8437547084526141, |
|
"grad_norm": 0.026611328125, |
|
"learning_rate": 0.00031282688425428686, |
|
"loss": 0.1778, |
|
"step": 28000 |
|
}, |
|
{ |
|
"epoch": 0.8467681181256592, |
|
"grad_norm": 0.0260009765625, |
|
"learning_rate": 0.0003086056984611053, |
|
"loss": 0.1782, |
|
"step": 28100 |
|
}, |
|
{ |
|
"epoch": 0.8497815277987042, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.00030445991455047927, |
|
"loss": 0.1764, |
|
"step": 28200 |
|
}, |
|
{ |
|
"epoch": 0.8503842097333132, |
|
"eval_peoplespeech-clean-transcription_loss": 1.485144853591919, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.3295, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.801, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.075, |
|
"step": 28220 |
|
}, |
|
{ |
|
"epoch": 0.8527949374717493, |
|
"grad_norm": 0.0279541015625, |
|
"learning_rate": 0.000300389927521351, |
|
"loss": 0.178, |
|
"step": 28300 |
|
}, |
|
{ |
|
"epoch": 0.8558083471447944, |
|
"grad_norm": 0.0255126953125, |
|
"learning_rate": 0.0002963961251509423, |
|
"loss": 0.1794, |
|
"step": 28400 |
|
}, |
|
{ |
|
"epoch": 0.8588217568178393, |
|
"grad_norm": 0.0289306640625, |
|
"learning_rate": 0.0002924788879578099, |
|
"loss": 0.1774, |
|
"step": 28500 |
|
}, |
|
{ |
|
"epoch": 0.8618351664908844, |
|
"grad_norm": 0.02734375, |
|
"learning_rate": 0.00028863858916559, |
|
"loss": 0.178, |
|
"step": 28600 |
|
}, |
|
{ |
|
"epoch": 0.8648485761639295, |
|
"grad_norm": 0.0247802734375, |
|
"learning_rate": 0.0002848755946674383, |
|
"loss": 0.1756, |
|
"step": 28700 |
|
}, |
|
{ |
|
"epoch": 0.8678619858369745, |
|
"grad_norm": 0.0264892578125, |
|
"learning_rate": 0.00028119026299116905, |
|
"loss": 0.1783, |
|
"step": 28800 |
|
}, |
|
{ |
|
"epoch": 0.8708753955100196, |
|
"grad_norm": 0.029052734375, |
|
"learning_rate": 0.0002775829452650956, |
|
"loss": 0.1789, |
|
"step": 28900 |
|
}, |
|
{ |
|
"epoch": 0.8738888051830647, |
|
"grad_norm": 0.0262451171875, |
|
"learning_rate": 0.00027405398518457575, |
|
"loss": 0.1763, |
|
"step": 29000 |
|
}, |
|
{ |
|
"epoch": 0.8769022148561096, |
|
"grad_norm": 0.0281982421875, |
|
"learning_rate": 0.0002706037189792652, |
|
"loss": 0.1771, |
|
"step": 29100 |
|
}, |
|
{ |
|
"epoch": 0.8799156245291547, |
|
"grad_norm": 0.0263671875, |
|
"learning_rate": 0.00026723247538108254, |
|
"loss": 0.1795, |
|
"step": 29200 |
|
}, |
|
{ |
|
"epoch": 0.8829290342021998, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.00026394057559288856, |
|
"loss": 0.1779, |
|
"step": 29300 |
|
}, |
|
{ |
|
"epoch": 0.8859424438752448, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.00026072833325788375, |
|
"loss": 0.1779, |
|
"step": 29400 |
|
}, |
|
{ |
|
"epoch": 0.8889558535482899, |
|
"grad_norm": 0.02587890625, |
|
"learning_rate": 0.0002575960544297239, |
|
"loss": 0.1783, |
|
"step": 29500 |
|
}, |
|
{ |
|
"epoch": 0.891969263221335, |
|
"grad_norm": 0.0235595703125, |
|
"learning_rate": 0.0002545440375433609, |
|
"loss": 0.1766, |
|
"step": 29600 |
|
}, |
|
{ |
|
"epoch": 0.8949826728943799, |
|
"grad_norm": 0.030517578125, |
|
"learning_rate": 0.0002515725733866084, |
|
"loss": 0.1782, |
|
"step": 29700 |
|
}, |
|
{ |
|
"epoch": 0.897996082567425, |
|
"grad_norm": 0.024658203125, |
|
"learning_rate": 0.000248681945072437, |
|
"loss": 0.1769, |
|
"step": 29800 |
|
}, |
|
{ |
|
"epoch": 0.9004068103058611, |
|
"eval_peoplespeech-clean-transcription_loss": 1.481724500656128, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.9492, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.588, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.072, |
|
"step": 29880 |
|
}, |
|
{ |
|
"epoch": 0.9010094922404701, |
|
"grad_norm": 0.0272216796875, |
|
"learning_rate": 0.000245872428011999, |
|
"loss": 0.1763, |
|
"step": 29900 |
|
}, |
|
{ |
|
"epoch": 0.9040229019135152, |
|
"grad_norm": 0.0255126953125, |
|
"learning_rate": 0.00024314428988838856, |
|
"loss": 0.1764, |
|
"step": 30000 |
|
}, |
|
{ |
|
"epoch": 0.9070363115865602, |
|
"grad_norm": 0.0274658203125, |
|
"learning_rate": 0.000240497790631138, |
|
"loss": 0.1754, |
|
"step": 30100 |
|
}, |
|
{ |
|
"epoch": 0.9100497212596053, |
|
"grad_norm": 0.02587890625, |
|
"learning_rate": 0.00023793318239145138, |
|
"loss": 0.1781, |
|
"step": 30200 |
|
}, |
|
{ |
|
"epoch": 0.9130631309326503, |
|
"grad_norm": 0.0299072265625, |
|
"learning_rate": 0.00023545070951818084, |
|
"loss": 0.1776, |
|
"step": 30300 |
|
}, |
|
{ |
|
"epoch": 0.9160765406056953, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.00023305060853454597, |
|
"loss": 0.1789, |
|
"step": 30400 |
|
}, |
|
{ |
|
"epoch": 0.9190899502787404, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.00023073310811559807, |
|
"loss": 0.1793, |
|
"step": 30500 |
|
}, |
|
{ |
|
"epoch": 0.9221033599517855, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.00022849842906643277, |
|
"loss": 0.1793, |
|
"step": 30600 |
|
}, |
|
{ |
|
"epoch": 0.9251167696248305, |
|
"grad_norm": 0.024169921875, |
|
"learning_rate": 0.00022634678430115206, |
|
"loss": 0.177, |
|
"step": 30700 |
|
}, |
|
{ |
|
"epoch": 0.9281301792978756, |
|
"grad_norm": 0.0247802734375, |
|
"learning_rate": 0.0002242783788225793, |
|
"loss": 0.1788, |
|
"step": 30800 |
|
}, |
|
{ |
|
"epoch": 0.9311435889709206, |
|
"grad_norm": 0.0257568359375, |
|
"learning_rate": 0.00022229340970272572, |
|
"loss": 0.1777, |
|
"step": 30900 |
|
}, |
|
{ |
|
"epoch": 0.9341569986439656, |
|
"grad_norm": 0.02392578125, |
|
"learning_rate": 0.00022039206606401526, |
|
"loss": 0.1776, |
|
"step": 31000 |
|
}, |
|
{ |
|
"epoch": 0.9371704083170107, |
|
"grad_norm": 0.02880859375, |
|
"learning_rate": 0.0002185745290612646, |
|
"loss": 0.1774, |
|
"step": 31100 |
|
}, |
|
{ |
|
"epoch": 0.9401838179900558, |
|
"grad_norm": 0.0286865234375, |
|
"learning_rate": 0.00021684097186442405, |
|
"loss": 0.1786, |
|
"step": 31200 |
|
}, |
|
{ |
|
"epoch": 0.9431972276631008, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.0002151915596420774, |
|
"loss": 0.1777, |
|
"step": 31300 |
|
}, |
|
{ |
|
"epoch": 0.9462106373361459, |
|
"grad_norm": 0.0252685546875, |
|
"learning_rate": 0.0002136264495457057, |
|
"loss": 0.1781, |
|
"step": 31400 |
|
}, |
|
{ |
|
"epoch": 0.9492240470091909, |
|
"grad_norm": 0.028076171875, |
|
"learning_rate": 0.00021214579069471447, |
|
"loss": 0.1772, |
|
"step": 31500 |
|
}, |
|
{ |
|
"epoch": 0.9504294108784089, |
|
"eval_peoplespeech-clean-transcription_loss": 1.4812726974487305, |
|
"eval_peoplespeech-clean-transcription_model_preparation_time": 0.0062, |
|
"eval_peoplespeech-clean-transcription_runtime": 13.7554, |
|
"eval_peoplespeech-clean-transcription_samples_per_second": 4.653, |
|
"eval_peoplespeech-clean-transcription_steps_per_second": 0.073, |
|
"step": 31540 |
|
}, |
|
{ |
|
"epoch": 0.9522374566822359, |
|
"grad_norm": 0.0230712890625, |
|
"learning_rate": 0.0002107497241622257, |
|
"loss": 0.1767, |
|
"step": 31600 |
|
}, |
|
{ |
|
"epoch": 0.955250866355281, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.00020943838296163657, |
|
"loss": 0.1788, |
|
"step": 31700 |
|
}, |
|
{ |
|
"epoch": 0.9582642760283261, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.00020821189203394706, |
|
"loss": 0.1783, |
|
"step": 31800 |
|
}, |
|
{ |
|
"epoch": 0.9612776857013711, |
|
"grad_norm": 0.0238037109375, |
|
"learning_rate": 0.00020707036823585488, |
|
"loss": 0.1753, |
|
"step": 31900 |
|
}, |
|
{ |
|
"epoch": 0.9642910953744162, |
|
"grad_norm": 0.027587890625, |
|
"learning_rate": 0.00020601392032862275, |
|
"loss": 0.1752, |
|
"step": 32000 |
|
}, |
|
{ |
|
"epoch": 0.9673045050474612, |
|
"grad_norm": 0.0255126953125, |
|
"learning_rate": 0.00020504264896771505, |
|
"loss": 0.177, |
|
"step": 32100 |
|
}, |
|
{ |
|
"epoch": 0.9703179147205062, |
|
"grad_norm": 0.025634765625, |
|
"learning_rate": 0.00020415664669320817, |
|
"loss": 0.1786, |
|
"step": 32200 |
|
}, |
|
{ |
|
"epoch": 0.9733313243935513, |
|
"grad_norm": 0.0244140625, |
|
"learning_rate": 0.00020335599792097327, |
|
"loss": 0.1764, |
|
"step": 32300 |
|
}, |
|
{ |
|
"epoch": 0.9763447340665964, |
|
"grad_norm": 0.0240478515625, |
|
"learning_rate": 0.00020264077893463362, |
|
"loss": 0.1781, |
|
"step": 32400 |
|
}, |
|
{ |
|
"epoch": 0.9793581437396414, |
|
"grad_norm": 0.0277099609375, |
|
"learning_rate": 0.00020201105787829627, |
|
"loss": 0.1761, |
|
"step": 32500 |
|
}, |
|
{ |
|
"epoch": 0.9823715534126864, |
|
"grad_norm": 0.027099609375, |
|
"learning_rate": 0.00020146689475005947, |
|
"loss": 0.1786, |
|
"step": 32600 |
|
}, |
|
{ |
|
"epoch": 0.9853849630857315, |
|
"grad_norm": 0.0242919921875, |
|
"learning_rate": 0.00020100834139629646, |
|
"loss": 0.1776, |
|
"step": 32700 |
|
}, |
|
{ |
|
"epoch": 0.9883983727587765, |
|
"grad_norm": 0.02685546875, |
|
"learning_rate": 0.00020063544150671555, |
|
"loss": 0.1765, |
|
"step": 32800 |
|
}, |
|
{ |
|
"epoch": 0.9914117824318216, |
|
"grad_norm": 0.026123046875, |
|
"learning_rate": 0.00020034823061019724, |
|
"loss": 0.1768, |
|
"step": 32900 |
|
}, |
|
{ |
|
"epoch": 0.9944251921048667, |
|
"grad_norm": 0.026611328125, |
|
"learning_rate": 0.00020014673607140958, |
|
"loss": 0.1771, |
|
"step": 33000 |
|
}, |
|
{ |
|
"epoch": 0.9974386017779117, |
|
"grad_norm": 0.02490234375, |
|
"learning_rate": 0.00020003097708820057, |
|
"loss": 0.1781, |
|
"step": 33100 |
|
} |
|
], |
|
"logging_steps": 100, |
|
"max_steps": 33185, |
|
"num_input_tokens_seen": 0, |
|
"num_train_epochs": 1, |
|
"save_steps": 8297, |
|
"stateful_callbacks": { |
|
"TrainerControl": { |
|
"args": { |
|
"should_epoch_stop": false, |
|
"should_evaluate": false, |
|
"should_log": false, |
|
"should_save": true, |
|
"should_training_stop": true |
|
}, |
|
"attributes": {} |
|
} |
|
}, |
|
"total_flos": 5.732431512509769e+19, |
|
"train_batch_size": 672, |
|
"trial_name": null, |
|
"trial_params": null |
|
} |
|
|