Skip to yearly menu bar Skip to main content


Poster

Spectral Graph Pruning Against Over-Squashing and Over-Smoothing

Adarsh Jamadandi · Celia Rubio-Madrigal · Rebekka Burkholz

[ ]
Wed 11 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

Message Passing Graph Neural Networks are known to suffer from two problems that are sometimes believed to be diametrically opposed: over-squashing and over-smoothing. The former results from topological bottlenecks that hamper the information flow from distant nodes and are mitigated by spectral gap maximization, primarily, by means of edge additions. However, such additions often promote over-smoothing that renders nodes of different classes less distinguishable. Inspired by the Braess phenomenon, we argue that deleting edges can address over-squashing and over-smoothing simultaneously. This insight explains how edge deletions can improve generalization, thus connecting spectral gap optimization to a seemingly disconnected objective of reducing computational resources by pruning graphs for lottery tickets. To this end, we propose a computationally effective spectral gap optimization framework to add or delete edges and demonstrate its effectiveness on the long range graph benchmark and on larger heterophilous datasets.

Live content is unavailable. Log in and register to view live content