Timezone: »
Poster
Transformers Generalize DeepSets and Can be Extended to Graphs & Hypergraphs
Jinwoo Kim · Saeyoon Oh · Seunghoon Hong
We present a generalization of Transformers to any-order permutation invariant data (sets, graphs, and hypergraphs). We begin by observing that Transformers generalize DeepSets, or first-order (set-input) permutation invariant MLPs. Then, based on recently characterized higher-order invariant MLPs, we extend the concept of self-attention to higher orders and propose higher-order Transformers for order-$k$ data ($k=2$ for graphs and $k>2$ for hypergraphs). Unfortunately, higher-order Transformers turn out to have prohibitive complexity $\mathcal{O}(n^{2k})$ to the number of input nodes $n$. To address this problem, we present sparse higher-order Transformers that have quadratic complexity to the number of input hyperedges, and further adopt the kernel attention approach to reduce the complexity to linear. In particular, we show that the sparse second-order Transformers with kernel attention are theoretically more expressive than message passing operations while having an asymptotically identical complexity. Our models achieve significant performance improvement over invariant MLPs and message-passing graph neural networks in large-scale graph regression and set-to-(hyper)graph prediction tasks. Our implementation is available at https://github.com/jw9730/hot.
Author Information
Jinwoo Kim (KAIST)
Saeyoon Oh (Korea Advanced Institute of Science and Technology)
Seunghoon Hong (University of Michigan)
More from the Same Authors
-
2023 Poster: Learning Probabilistic Symmetrization for Architecture Agnostic Equivariance »
Jinwoo Kim · Dat Nguyen · Ayhan Suleymanzade · Hyeokjun An · Seunghoon Hong -
2022 Poster: Pure Transformers are Powerful Graph Learners »
Jinwoo Kim · Dat Nguyen · Seonwoo Min · Sungjun Cho · Moontae Lee · Honglak Lee · Seunghoon Hong -
2022 Poster: Transformers meet Stochastic Block Models: Attention with Data-Adaptive Sparsity and Cost »
Sungjun Cho · Seonwoo Min · Jinwoo Kim · Moontae Lee · Honglak Lee · Seunghoon Hong -
2021 Poster: Multi-View Representation Learning via Total Correlation Objective »
HyeongJoo Hwang · Geon-Hyeong Kim · Seunghoon Hong · Kee-Eung Kim -
2018 Poster: Learning Hierarchical Semantic Image Manipulation through Structured Representations »
Seunghoon Hong · Xinchen Yan · Thomas Huang · Honglak Lee