Understanding Oversmoothing in Graph Neural Networks (GNNs): Insights from Two Theoretical Studies

Subscribers:
344,000
Published on ● Video Link: https://www.youtube.com/watch?v=MLiEoJOhXJA



Duration: 59:46
985 views
17


A Google TechTalk, presented by Xinyi Wu, 2024-01-18
A Google Algorithm Seminar. ABSTRACT: Oversmoothing in Graph Neural Networks (GNNs) refers to the phenomenon where increasing network depth leads to homogeneous node representations. Over the last few years, it has remained as one of the central challenges of building more powerful Graph Neural Networks (GNNs). In this talk, I will discuss two recent papers on this phenomenon and provide some new insights.

The first work studies why oversmoothing happens at a relatively shallow depth in GNNs. By carefully analyzing the oversmoothing mechanisms in a stylized formulation, we distinguish between adverse mixing that homogenizes nodes across different classes and beneficial denoising within the same class. We quantify these two effects on random graphs sampled from the Contextual Stochastic Block Model (CSBM) and show that oversmoothing occurs once the mixing effect starts to dominate the denoising effect. We establish that the number of layers required for this transition is O(logN/log(logN)) for sufficiently dense graphs with N nodes. We also extend our analysis to study the effects of Personalized PageRank (PPR), or equivalently, the effects of initial residual connections on oversmoothing, and shed light on when and why they might not be an ideal solution to the problem.

In the second work, we study oversmoothing in attention-based GNNs, such as Graph Attention Networks (GATs) and transformers. Treating attention-based GNNs as dynamical systems, our study demonstrates that the graph attention mechanism cannot prevent oversmoothing and loses expressive power exponentially. From a technical point of view, the proposed framework significantly extends the existing results on oversmoothing, and can account for asymmetric, state-dependent and time-varying aggregation operators and a wide range of common nonlinear activation functions, such as ReLU, LeakyReLU, GELU and SiLU.

The talk is based on the following papers: https://arxiv.org/abs/2212.10701, https://arxiv.org/abs/2305.16102. Joint works with Amir Ajorlou (MIT), Zhengdao Chen (NYU/Google), William Wang (MIT), Zihui Wu (Caltech) and Ali Jadbabaie (MIT).

ABOUT THE SPEAKER: Xinyi Wu is a fourth-year Ph.D. student in the Institute for Data, Systems, and Society (IDSS) at Massachusetts Institute of Technology (MIT), advised by Professor Ali Jadbabaie. She is affiliated with the Laboratory for Information and Decision Systems (LIDS). She is a recipient of the MIT Michael Hammer Fellowship. She is interested in applied graph theory, dynamical systems, networks, and machine learning on graphs. Her work on oversmoothing in GNNs has been awarded as Spotlight paper in NeurIPS 2023.




Other Videos By Google TechTalks


2024-04-22Design is Testability
2024-04-12Charles Hoskinson | CEO of Input Output Global | web3 talks | Apr 4th 2024 | MC: Marlon Ruiz
2024-04-09Limitations of Stochastic Selection with Pairwise Independent Priors
2024-04-02NASA CARA - Air Traffic Control in Spaaaaaaaace
2024-03-28How Your Brain Processes Code
2024-03-25Fixed-point Error Bounds for Mean-payoff Markov Decision Processes
2024-03-19One Tree to Rule Them All: Polylogarithmic Universal Steiner Trees
2024-01-26Understanding Oversmoothing in Graph Neural Networks (GNNs): Insights from Two Theoretical Studies
2023-12-05Socially Responsible Software Development (Teaching Software Design Systematically)
2023-12-04Understanding and Mitigating Copying in Diffusion Models
2023-12-04Efficient Training Image Extraction from Diffusion Models Ryan Webs
2023-11-30High-Dimensional Prediction for Sequential Decision Making
2023-09-01Representational Strengths and Limitations of Transformers
2023-09-01Steven Goldfeder | CEO Offchain Labs / Arbitrum | web3 talks | Aug 24 2023 | MC: Marlon Ruiz
2023-08-29Differentially Private Sampling from Distributions
2023-07-14Revisiting Nearest Neighbors from a Sparse Signal Approximation View
2023-07-042023 Blockly Developer Summit Day 2-5: Plug-ins Demonstration
2023-07-042023 Blockly Developer Summit DAY 1-5: The Future of Computational Thinking
2023-07-042023 Blockly Developer Summit DAY 1-7: Cubi - Extending Blockly for Teachers
2023-07-042023 Blockly Developer Summit DAY 1-12: Serialization and Visual Diff
2023-07-042023 Blockly Developer Summit Day 2-2: Blockly Themes for Accessibility