[WeightWatcher] Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory

Published on ● Video Link: https://www.youtube.com/watch?v=DymfJGOOK_4



Duration: 1:08:16
707 views
19


For slides and more information on the paper, visit https://aisc.ai.science/events/2019-11-06

Discussion lead & author: Charles Martin

Abstract:

Random Matrix Theory (RMT) is applied to analyze weight matrices of Deep Neural Networks (DNNs), including both production quality, pre-trained models such as AlexNet and Inception, and smaller models trained from scratch, such as LeNet5 and a miniature-AlexNet. Empirical and theoretical results clearly indicate that the DNN training process itself implicitly implements a form of Self-Regularization. The empirical spectral density (ESD) of DNN layer matrices displays signatures of traditionally-regularized statistical models, even in the absence of exogenously specifying traditional forms of explicit regularization. Building on relatively recent results in RMT, most notably its extension to Universality classes of Heavy-Tailed matrices, we develop a theory to identify 5+1 Phases of Training, corresponding to increasing amounts of Implicit Self-Regularization. These phases can be observed during the training process as well as in the final learned DNNs. For smaller and/or older DNNs, this Implicit Self-Regularization is like traditional Tikhonov regularization, in that there is a "size scale" separating signal from noise. For state-of-the-art DNNs, however, we identify a novel form of Heavy-Tailed Self-Regularization, similar to the self-organization seen in the statistical physics of disordered systems. This results from correlations arising at all size scales, which arises implicitly due to the training process itself. This implicit Self-Regularization can depend strongly on the many knobs of the training process. By exploiting the generalization gap phenomena, we demonstrate that we can cause a small model to exhibit all 5+1 phases of training simply by changing the batch size. This demonstrates that---all else being equal---DNN optimization with larger batch sizes leads to less-well implicitly-regularized models, and it provides an explanation for the generalization gap phenomena.




Other Videos By LLMs Explained - Aggregate Intellect - AI.SCIENCE


2020-01-08Overview of Modern Anomaly and Novelty Detection | AISC
2020-01-06Annotating Object Instances With a Polygon RNN | AISC
2019-12-11Predicting translational progress in biomedical research | AISC
2019-12-09AlphaStar explained: Grandmaster level in StarCraft II with multi-agent RL
2019-12-04How Can We Be So Dense? The Benefits of Using Highly Sparse Representations | AISC
2019-12-02[RoBERT & ToBERT] Hierarchical Transformers for Long Document Classification | AISC
2019-11-25[OpenAI] Solving Rubik's Cube with a Robot Hand | AISC
2019-11-18Top-K Off-Policy Correction for a REINFORCE Recommender System | AISC
2019-11-13Overview of Unsupervised & Semi-supervised learning | AISC
2019-11-11Building products for Continous Delivery in Machine Learning | AISC
2019-11-06[WeightWatcher] Self-Regularization in Deep Neural Networks: Evidence from Random Matrix Theory
2019-11-04Defending Against Fake Neural News | AISC
2019-10-28[RecSys Challenge 2019 2nd Place] Robust Contextual Models for In-Session Personalization | AISC
2019-10-23Deep learning enables rapid identification of potent DDR1 kinase inhibitors | AISC
2019-10-22Restricted Boltzmann Machines for Collaborative Filtering | AISC
2019-10-15Location Intelligence Products: Goals and Challenges | AISC
2019-10-15RecSys, Reverse Engineering User's Needs and Desires | AISC
2019-10-08Breaking the Softmax Bottleneck via Learnable Monotonic Pointwise Non-linearities | AISC
2019-10-07DeepStack: Expert-Level Artificial Intelligence in Heads-Up No-Limit Poker | AISC
2019-09-30AISC Abstract Night September Edition | AISC
2019-09-26EigenDamage: Structured Pruning in the Kronecker-Factored Eigenbasis



Tags:
weightwatcher
deep learning
random matrix theory