The hybrid deep neural Network (DNN) and hidden Markov model (HMM) have recently achieved great performance in speech recognition. However, the computing hardware was not adequate to learn deep neural networks with more hidden layers from big data set...
The hybrid deep neural Network (DNN) and hidden Markov model (HMM) have recently achieved great performance in speech recognition. However, the computing hardware was not adequate to learn deep neural networks with more hidden layers from big data sets. Further, despite the powerful performance of a DNN-based acoustic model, the time-consuming learning process has been a problem. This paper proposes a novel DNN-based acoustic modeling framework for speech recognition. The new model adopts parallel training in multiple DNNs. Several hierarchically structured DNNs are trained separately in parallel, using multiple computing units. Weights are averaged after each epoch. The suggested structure separates DNN into 10 and shows approximately 7.5 times faster in training time than baseline hybrid deep neural network. This improvement in average training time is mainly attributed to the use of multiple GPUs and the fact that training is based on only a subset of data in parallel. The WSJ data set was used for proposed parallel DNN performance verification.