Self-Tuning Networks: Amortizing the Hypergradient Computation for Hyperparameter Optimization

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=eoLeANtBGKs



Duration: 1:00:50
1,698 views
55


Optimization of many deep learning hyperparameters can be formulated as a bilevel optimization problem. While most black-box and gradient-based approaches require many independent training runs, we aim to adapt hyperparameters online as the network trains. The main challenge is to approximate the response Jacobian, which captures how the minimum of the inner objective changes as the hyperparameters are perturbed. To do this, we introduce the self-tuning network (STN), which fits a hypernetwork to approximate the best response function in the vicinity of the current hyperparameters. Differentiating through the hypernetwork lets us efficiently approximate the gradient of the validation loss with respect to the hyperparameters. We train the hypernetwork and hyperparameters jointly. Empirically, we can find hyperparameter settings competitive with Bayesian Optimization in a single run of training, and in some cases find hyperparameter schedules that outperform any fixed hyperparameter value.

Roger Grosse is an Assistant Professor of Computer Science at the University of Toronto, and a founding member of the Vector Institute for Artificial Intelligence. He received his Ph.D. in computer science from MIT, and then spent two years as a postdoc at the University of Toronto. He holds a Canada Research Chair in Probabilistic Inference and Deep Learning, an Ontario MRIS Early Researcher Award, and a CIFAR Canadian AI Chair.

Learn more about the 2020-2021 Directions in ML: AutoML and Automating Algorithms virtual speaker series: https://aka.ms/diml




Other Videos By Microsoft Research


2021-04-29Virtual Lake Nona Impact Forum “Health Innovation in the New Reality”
2021-04-28Sound Capture and Speech Enhancement for Communication and Distant Speech Recognition
2021-04-27Virtual Lake Nona Impact Forum “Health Innovation in the New Reality”
2021-04-26FastNeRF: High-Fidelity Neural Rendering at 200FPS [Condensed]
2021-04-21Research for Industries (RFI) Lecture Series: Warren Powell
2021-04-21Research for Industries (RFI) Lecture Series: Andreas Haeberlen
2021-04-13Discovering hidden connections in art with deep, interpretable visual analogies
2021-04-13ZeRO & Fastest BERT: Increasing the scale and speed of deep learning training in DeepSpeed
2021-04-13Interactive sound simulation: Rendering immersive soundscapes in games and virtual reality
2021-04-13A prototype implementation of 4G packet gateway Microsoft Catapult FPGA platform
2021-04-12Self-Tuning Networks: Amortizing the Hypergradient Computation for Hyperparameter Optimization
2021-04-06Ultra-dense data storage and extreme parallelism with electronic-molecular systems
2021-04-06Harmonizing the declarative and imperative in database systems
2021-04-06Domain-specific language model pretraining for biomedical natural language processing
2021-03-30Platform Biography: A framework for analyzing the structures and dynamics of social media
2021-03-30Building multimodal, integrative AI systems with Platform for Situated Intelligence
2021-03-29From player to creator: Designing video games on gaming handhelds with Microsoft TileCode webinar
2021-03-29Camera-based non-contact health sensing
2021-03-29Foundations of causal inference and its impacts on machine learning webinar
2021-03-29Avatars: Finding a sense of self and others in the virtual world
2021-03-25In pursuit of responsible AI: Bringing principles to practice