mSDA: A fast and easy-to-use way to improve bag-of-words features

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=jrLVGdVNx7A



Duration: 1:08:03
986 views
10


Machine learning algorithms rely heavily on the representation of the data they are presented with. In particular, text documents (and often images) are traditionally expressed as bag-of-words feature vectors (e.g. as tf-idf). Recently Glorot et al. showed that stacked denoising autoencoders (SDA), a deep learning algorithm, can learn representations that are far superior over variants of bag-of-words. Unfortunately, training SDAs often requires a prohibitive amount of computation time and is non-trivial for non-experts. In this work, we show that with a few modifications of the SDA model, we can relax the optimization over the hidden weights into convex optimization problems with closed form solutions. Further, we show that the expected value of the hidden weights after infinitely many training iterations can also be computed in closed form. The resulting transformation (which we call marginalized-SDA) can be computed in no more than 20 lines of straight-forward Matlab code and requires no prior expertise in machine learning. The representations learned with mSDA behave similar to those obtained with SDA, but the training time is reduced by several orders of magnitudes. For example, mSDA matches the world-record on the Amazon transfer learning benchmark, however the training time shrinks from several days to a few minutes.




Other Videos By Microsoft Research


2016-07-28Approximating the Expansion Profile and Almost Optimal Local Graph Clustering
2016-07-28Stochastic Dual Coordinate Ascent and its Proximal Extension for Regularized Loss Minimization
2016-07-28A Practical Approach to Reduce the Power Consumption of LCD Displays
2016-07-28CryptDB: Processing Queries on an Encrypted Database
2016-07-28Performing Time, Space and Light
2016-07-28Probabilistic Methods for Efficient Search & Statistical Learning in Extremely HighDimensional Data
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 4
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 2
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 1
2016-07-28Bug Finding Techniques for Programs with Infinitely Many States
2016-07-28mSDA: A fast and easy-to-use way to improve bag-of-words features
2016-07-28Verifying the integrity of peripherals' firmware
2016-07-28Privacy, Audit and Accountability
2016-07-28One Mouse per Child
2016-07-28The Benefits Of Being Out Of Focus: Making the Most of Lens PSF
2016-07-28Algorithms for bipartite matching problems with connections to sparsification and streaming
2016-07-28MIMD on GPU
2016-07-28The Case for Continuous Time
2016-07-28Starfish: A MADDER and Self-tuning System for Big Data Analytics
2016-07-28Spatial Coding for Large-scale Partial-duplicate Image Search
2016-07-28Testing Atomicity of Composed Concurrent Operations & Automatic Fine-Grain Locking



Tags:
microsoft research