Communication-Avoiding Algorithms and Fast Matrix Multiplication

Subscribers:
342,000
Published on ● Video Link: https://www.youtube.com/watch?v=nwdkeAxqZDc



Duration: 1:01:05
872 views
20


As our computing capabilities grow, the size and complexity of numerical simulations and data analysis that today's computational scientists conduct continue to increase. The gap between the peak capabilities of the hardware and the achieved performance of many of these computations is caused in large part by the high cost of communication, or movement of data, between processors and throughout the memory hierarchy of a single processor. As expected, much of this communication cannot be avoided when using parallelism to solve our problems; however, many standard algorithms in linear algebra move more data than necessary. We have proved lower bounds on the communication required by a wide class of algorithms and can determine whether standard approaches attain these bounds. In several cases where gaps exist between algorithms and lower bounds, we have developed new algorithms that communicate asymptotically less than previous ones and outperform them on a variety of platforms. These include algorithms for solving linear systems, least squares problems, eigenvalue problems, and parallelization of Strassen's matrix multiplication algorithm. In particular, not only does Strassen's algorithm require less computation that the classical matrix multiplication algorithm, it also requires less communication. In fact, fast matrix multiplication algorithms with smaller exponent than Strassen's in their computational complexity require even less communication. I'll talk about recent development in implementations of fast algorithms that exploit this fact and where fast matrix multiplication, often thought of as a theoretical pursuit, can be practically useful.




Other Videos By Microsoft Research


2016-06-21A Fast Distributed Algorithm for Ξ±-Fair Packing Problems
2016-06-21Theory Day Session 3
2016-06-21Provable Submodular Minimization via Wolfe’s Algorithm
2016-06-21A Faster Cutting Plane Method and its Implications for Combinatorial and Convex Optimization
2016-06-21Stochastic Methods for Complex Performance Measures: A Tale of Two Families
2016-06-21Parallel Bayesian Network Structure Learning for Genome-Scale Gene Networks
2016-06-21Empirical Inference for Intelligent Systems
2016-06-21Ordered Stick-breaking Prior for Sequential MCMC Inference of Bayesian Non-Parametric Models
2016-06-21Effective-Resistance-Reducing Flows, Spectrally Thin Trees and Asymmetric TSP
2016-06-21Non-Convex Robust PCA
2016-06-21Communication-Avoiding Algorithms and Fast Matrix Multiplication
2016-06-21The Forza Motorsport 5 Original Soundtrack, An Insider's View
2016-06-21Reasoning about GADT Pattern Matching in Haskell
2016-06-21Non-Convex Robust PCA - Part 2
2016-06-21Modeling, Quantifying, and Limiting Adversary Knowledge
2016-06-21Sampling Techniques for Constraint Satisfaction and Beyond
2016-06-21Cell Classification of FT-IR Spectroscopic Data for Histopathology using Neural Networks
2016-06-21Sequential Equilibrium Distributions in Multi-Stage Games with Infinite Sets of Types and Actions
2016-06-21Randomized Interior Point Methods for Sampling and Optimization
2016-06-21Introduction to large-scale optimization - Part1
2016-06-21Welcome and Microsoft Research Asia Update



Tags:
microsoft research
algorithms