Research Area:  Machine Learning
Training large, deep neural network acoustic models for speech recognition on large datasets takes a long time on a single GPU, motivating research on parallel training algorithms. We present an approach for training a bidirectional LSTM acoustic model on the 2000-hour Switchboard corpus. The model we train achieves state-of-the-art word error rate, 7.5% on the Hub5-2000 Switchboard test set and 13.1% on the Callhome test set, and scales to an unprecedented 96 learners while employing only 12 global reductions per epoch of training. As our implementation incurs far fewer reductions than prior work, it does not require aggressively optimized communication primitives to reach state-of-the-art performance in a short amount of time. With 48 NVIDIA V100 GPUs training takes 5 hours; with 96 GPUs, training takes around 3 hours.
Keywords:  
Author(s) Name:  Guojing Cong; Brian Kingsbury; Chih-Chieh Yang; Tianyi Liu
Journal name:  
Conferrence name:  ICASSP 2020 - 2020 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP)
Publisher name:  IEEE
DOI:  10.1109/ICASSP40776.2020.9053993
Volume Information:  Volume 2020
Paper Link:   https://ieeexplore.ieee.org/abstract/document/9053993