Fast Neural Kernel Embeddings for General Activations

Subscribers:
348,000
Published on ● Video Link: https://www.youtube.com/watch?v=SfV166gNLUY



Duration: 33:44
924 views
15


A Google TechTalk, presented by Insu Han, 2023-02-02
Algorithms Seminar Series. ABSTRACT: Infinite width limit has shed light on generalization and optimization aspects of deep learning by establishing connections between neural networks and kernel methods. Despite their importance, the utility of these kernel methods was limited in large-scale learning settings due to their (super-)quadratic runtime and memory complexities. Moreover, most prior works on neural kernels have focused on the ReLU activation, mainly due to its popularity but also due to the difficulty of computing such kernels for general activations. In this work, we overcome such difficulties by providing methods to work with general activations. First, we compile and expand the list of activation functions admitting exact dual activation expressions to compute neural kernels. When the exact computation is unknown, we present methods to effectively approximate them. We propose a fast sketching method that approximates any multi-layered Neural Network Gaussian Process (NNGP) kernel and Neural Tangent Kernel (NTK) matrices for a wide range of activation functions, going beyond the commonly analyzed ReLU activation. This is done by showing how to approximate the neural kernels using the truncated Hermite expansion of any desired activation functions. While most prior works require data points on the unit sphere, our methods do not suffer from such limitations and are applicable to any dataset of points in ℝd. Furthermore, we provide a subspace embedding for NNGP and NTK matrices with near input-sparsity runtime and near-optimal target dimension which applies to any \emph{homogeneous} dual activation functions with rapidly convergent Taylor expansion. Empirically, with respect to exact convolutional NTK (CNTK) computation, our method achieves 106× speedup for approximate CNTK of a 5-layer Myrtle network on CIFAR-10 dataset.

Bio: Insu Han is a postdoctoral research fellow in the Department of Electrical Engineering at Yale University, hosted by Prof. Amin Karbasi. He obtained his Ph.D. degree at Korea Advanced Institute of Science and Technology (KAIST) in 2021 under the supervision of Prof. Jinwoo Shin. He was a recipient of Microsoft Research Asia Fellowship in 2019. His research interests focus on approximate algorithm design and analysis for large-scale machine learning problems and their applications.




Other Videos By Google TechTalks


2023-03-09Evan Shapiro | CEO Mina Foundation | web3 talks | Feb 16th 2023 | MC: Marlon Ruiz
2023-03-07Zürich Go Meetup: Zero-effort Type-safe Parsing of JSON and XML
2023-03-07Zürich Go Meetup: Let’s Build a Game with Go
2023-03-07Zürich Go Meetup: Run Go programs on your Raspberry Pi with gokrazy!
2023-03-03Online Covering: Secretaries, Prophets and Universal Maps
2023-03-03Auto-bidding in Online Advertising: Campaign Management and Fairness
2023-03-03Tree Learning: Optimal Algorithms and Sample Complexity
2023-03-03A Fast Algorithm for Adaptive Private Mean Estimation
2023-02-13Piers Ridyard | CEO RDX Works | Radix Protocol | web3 talks | Dec 7th 2022 | MC: Blake DeBenon
2023-02-10Sergey Gorbunov | Co-Founder Axelar | web3 talks | Jan 26th 2023 | MC: Marlon Ruiz
2023-02-10Fast Neural Kernel Embeddings for General Activations
2023-02-10Pathwise Conditioning and Non-Euclidean Gaussian Processes
2023-02-10Privacy-Preserving Machine Learning with Fully Homomorphic Encryption
2023-01-19Charles Hoskinson | CEO of Input Output Global | web3 talks | Jan 5th 2023 | MC: Marlon Ruiz
2023-01-18Control, Confidentiality, and the Right to be Forgotten
2023-01-18The Saddle Point Accountant for Differential Privacy
2023-01-18Analog vs. Digital Epsilons: Implementation Considerations Considerations for Differential Privacy
2023-01-18Secure Self-supervised Learning
2023-01-18Example Memorization in Learning: Batch and Streaming
2023-01-18Marginal-based Methods for Differentially Private Synthetic Data
2023-01-18Private Convex Optimization via Exponential Mechanism