Large Linear Classification when Data Cannot Fit In Memory

Subscribers:
351,000
Published on ● Video Link: https://www.youtube.com/watch?v=rAKbbOWpUJ4



Duration: 1:11:42
226 views
5


Linear classification is a useful tool for dealing with large-scale data in applications such as document classification and natural language processing. Recent developments of linear classification have shown that the training process can be efficiently conducted. However, when the data size exceeds the memory capacity, most training methods suffer from very slow convergence due to the severe disk swapping. In the first part of the talk, we describe a block minimization framework for data larger than memory. Under the framework, a solver splits data into blocks and stores them into separate files. Then, at each time, the solver trains a data block loaded from disk. In the second part of the talk, we introduce a selective block minimization (SBM) algorithm, a block minimization method that makes use of selective sampling. At each step, SBM updates the model using data consisting of two parts: (1) new data loaded from disk and (2) a set of informative samples already in memory from previous steps. We prove that, by updating the linear model in the dual form, the proposed method fully utilizes the data in memory and converges to a globally optimal solution on the entire data. Experiments show that the SBM algorithm dramatically reduces the number of blocks loaded from disk and consequently obtains an accurate and stable model quickly on both binary and multi-class classification. Joint work with Hsiang-Fu Yu, Cho-Jui Hsieh, Chih-Jen Lin and Dan Roth.




Other Videos By Microsoft Research


2016-07-28Approximating the Expansion Profile and Almost Optimal Local Graph Clustering
2016-07-28Stochastic Dual Coordinate Ascent and its Proximal Extension for Regularized Loss Minimization
2016-07-28A Practical Approach to Reduce the Power Consumption of LCD Displays
2016-07-28CryptDB: Processing Queries on an Encrypted Database
2016-07-28Performing Time, Space and Light
2016-07-28Probabilistic Methods for Efficient Search & Statistical Learning in Extremely HighDimensional Data
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 4
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 2
2016-07-28Quantum Computation for Quantum Chemistry: Status, Challenges, and Prospects - Session 1
2016-07-28Bug Finding Techniques for Programs with Infinitely Many States
2016-07-28Large Linear Classification when Data Cannot Fit In Memory
2016-07-28Verifying the integrity of peripherals' firmware
2016-07-28Privacy, Audit and Accountability
2016-07-28One Mouse per Child
2016-07-28The Benefits Of Being Out Of Focus: Making the Most of Lens PSF
2016-07-28Algorithms for bipartite matching problems with connections to sparsification and streaming
2016-07-28MIMD on GPU
2016-07-28The Case for Continuous Time
2016-07-28Starfish: A MADDER and Self-tuning System for Big Data Analytics
2016-07-28Spatial Coding for Large-scale Partial-duplicate Image Search
2016-07-28Testing Atomicity of Composed Concurrent Operations & Automatic Fine-Grain Locking



Tags:
microsoft research