Abstract

In this paper, we show how new training principles and optimization techniques for neural networks can be used for different network structures. In particular, we revisit the Recurrent Neural Network (RNN), which explicitly models the Markovian dynamics of a set of observations through a non-linear function with a much larger hidden state space than traditional sequence models such as an HMM. We apply pretraining principles used for Deep Neural Networks (DNNs) and second-order optimization techniques to train an RNN. Moreover, we explore its application in the Aurora2 speech recognition task under mismatched noise conditions using a Tandem approach. We observe top performance on clean speech, and under high noise conditions, compared to multi-layer perceptrons (MLPs) and DNNs, with the added benefit of being a "deeper" model than an MLP but more compact than a DNN.

Keywords

Computer scienceRecurrent neural networkPerceptronArtificial neural networkHidden Markov modelArtificial intelligenceTime delay neural networkSet (abstract data type)Noise (video)Speech recognitionTask (project management)Deep neural networksMultilayer perceptronMachine learningPattern recognition (psychology)

Affiliated Institutions

Related Publications

Publication Info

Year
2012
Type
article
Pages
4085-4088
Citations
139
Access
Closed

External Links

Social Impact

Social media, news, blog, policy document mentions

Citation Metrics

139
OpenAlex

Cite This

Oriol Vinyals, Suman Ravuri, Daniel Povey (2012). Revisiting Recurrent Neural Networks for robust ASR. , 4085-4088. https://doi.org/10.1109/icassp.2012.6288816

Identifiers

DOI
10.1109/icassp.2012.6288816