Multi-rate neural networks for efficient acoustic modeling

Subscribers:
343,000
Published on ● Video Link: https://www.youtube.com/watch?v=9uA0577ffNE



Duration: 1:27:42
87 views
0


In sequence recognition, the problem of long-span dependency in input sequences is typically tackled using recurrent neural network architectures, and robustness to sequential distortions is achieved using training data representative of a variety of these distortions. However, both these solutions substantially increase the training time. Thus low computation complexity during training is critical for acoustic modeling. This talk proposes the use of multi-rate neural network architectures to satisfy the design requirement of computational efficiency. In these architectures the network is partitioned into groups of units, operating at various sampling rates. As the network evaluates certain groups only once every few time steps, the computational cost is reduced. This talk will focus on the multi-rate feed-forward convolutional architecture. It will present results on several large vocabulary continuous speech recognition (LVCSR) tasks with training data ranging from 3 to 1800 hours to show the effectiveness of this architecture in efficiently learning wider temporal dependencies in both small and large data scenarios. Further it will discuss the use of this architecture for robust acoustic modeling in far-field environments. This model was shown to provide state-of-art results in the ASpIRE far-field recognition challenge. This talk will also discuss some preliminary results of multi-rate recurrent neural network based acoustic models.




Other Videos By Microsoft Research


2016-06-22Oral Session: Solving Random Quadratic Systems of Equations-Nearly as Easy as Solving Linear Systems
2016-06-22Towards Understandable Neural Networks for High Level AI Tasks - Part 7
2016-06-22The Once and Future Internet
2016-06-22The First Order World of Galton-Watson Trees
2016-06-22Verasco, a formally verified C static analyzer
2016-06-22Symposium: Deep Learning - Max Jaderberg
2016-06-22Satisfiability of Ordering CSPs Above Average Is Fixed-Parameter Tractable
2016-06-22Symposium: Deep Learning - Harri Valpola
2016-06-22Machine Learning as Creative Tool for Designing Real-Time Expressive Interactions
2016-06-22Symposium: Deep Learning - Sergey Ioffe
2016-06-22Multi-rate neural networks for efficient acoustic modeling
2016-06-22Robust Spectral Inference for Joint Stochastic Matrix Factorization and Topic Modeling
2016-06-22Computational Limits in Statistical Inference: Hidden Cliques and Sum of Squares
2016-06-22Extreme Classification: A New Paradigm for Ranking & Recommendation
2016-06-22A Lasserre-Based (1+epsilon)-Approximation for Makespan Scheduling with Precedence Constraints
2016-06-22Oral Session: Learning Theory and Algorithms for Forecasting Non-stationary Time Series
2016-06-22Recent Developments in Combinatorial Optimization
2016-06-22Invited Talks: Computational Principles for Deep Neuronal Architectures
2016-06-22Coalescence in Branching Trees and Branching Random Walks
2016-06-22Oral Session: Randomized Block Krylov Methods
2016-06-22Oral Session: Fast Convergence of Regularized Learning in Games



Tags:
microsoft research
acoustic modeling