Skip to yearly menu bar Skip to main content


Poster

Revisiting Self-Supervised Heterogeneous Graph Learning from Spectral Clustering Perspective

YUJIE MO · Zhihe Lu · Runpeng Yu · Xiaofeng Zhu · Xinchao Wang

East Exhibit Hall A-C #3006
[ ]
Fri 13 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

Self-supervised heterogeneous graph learning (SHGL) has shown promising potential in diverse scenarios. However, while existing SHGL methods share a similar essential with clustering approaches, they encounter two significant limitations: (i) noise in graph structures is often introduced during the message-passing process to weaken node representations, and (ii) cluster-level information may be inadequately captured and leveraged, diminishing the performance in downstream tasks. In this paper, we address these limitations by theoretically revisiting SHGL from the spectral clustering perspective and introducing a novel framework enhanced by rank and dual consistency constraints. Specifically, our framework incorporates a rank-constrained spectral clustering method that refines the affinity matrix to exclude noise effectively. Additionally, we integrate node-level and cluster-level consistency constraints that concurrently capture invariant and clustering information to facilitate learning in downstream tasks. We theoretically demonstrate that the learned representations are divided into distinct partitions based on the number of classes and exhibit enhanced generalization ability across tasks. Experimental results affirm the superiority of our method, showcasing remarkable improvements in several downstream tasks compared to existing methods.

Live content is unavailable. Log in and register to view live content