Timezone: »
This paper is concerned with learning to solve tasks that require a chain of interde- pendent steps of relational inference, like answering complex questions about the relationships between objects, or solving puzzles where the smaller elements of a solution mutually constrain each other. We introduce the recurrent relational net- work, a general purpose module that operates on a graph representation of objects. As a generalization of Santoro et al. [2017]’s relational network, it can augment any neural network model with the capacity to do many-step relational reasoning. We achieve state of the art results on the bAbI textual question-answering dataset with the recurrent relational network, consistently solving 20/20 tasks. As bAbI is not particularly challenging from a relational reasoning point of view, we introduce Pretty-CLEVR, a new diagnostic dataset for relational reasoning. In the Pretty- CLEVR set-up, we can vary the question to control for the number of relational reasoning steps that are required to obtain the answer. Using Pretty-CLEVR, we probe the limitations of multi-layer perceptrons, relational and recurrent relational networks. Finally, we show how recurrent relational networks can learn to solve Sudoku puzzles from supervised training data, a challenging task requiring upwards of 64 steps of relational reasoning. We achieve state-of-the-art results amongst comparable methods by solving 96.6% of the hardest Sudoku puzzles.
Author Information
Rasmus Berg Palm (Technical University Denmark)
Ulrich Paquet (DeepMind)
Ole Winther (Technical University of Denmark)
More from the Same Authors
-
2020 Meetup: MeetUp: Copenhagen, Denmark »
Ole Winther -
2021 : Hierarchical Few-Shot Generative Models »
Giorgio Giannone · Ole Winther -
2022 : Human-AI Interaction in Selective Prediction Systems »
Elizabeth Bondi-Kelly · Raphael Koster · Hannah Sheahan · Martin Chadwick · Yoram Bachrach · Taylan Cemgil · Ulrich Paquet · Krishnamurthy Dvijotham -
2022 : Identifying endogenous peptide receptors by combining structure and transmembrane topology prediction »
Felix Teufel · Jan Christian Refsgaard · Christian Toft Madsen · Carsten Stahlhut · Mads Grønborg · Dennis Madsen · Ole Winther -
2022 : Few-Shot Diffusion Models »
Giorgio Giannone · Didrik Nielsen · Ole Winther -
2020 : Contributed talk - Testing the Genomic Bottleneck Hypothesis in Hebbian Meta-Learning »
Rasmus Berg Palm -
2019 Poster: BIVA: A Very Deep Hierarchy of Latent Variables for Generative Modeling »
Lars Maaløe · Marco Fraccaro · Valentin Liévin · Ole Winther -
2017 : Panel Session »
Neil Lawrence · Finale Doshi-Velez · Zoubin Ghahramani · Yann LeCun · Max Welling · Yee Whye Teh · Ole Winther -
2017 Poster: A Disentangled Recognition and Nonlinear Dynamics Model for Unsupervised Learning »
Marco Fraccaro · Simon Kamronn · Ulrich Paquet · Ole Winther -
2017 Spotlight: A Disentangled Recognition and Nonlinear Dynamics Model for Unsupervised Learning »
Marco Fraccaro · Simon Kamronn · Ulrich Paquet · Ole Winther -
2017 Poster: Hash Embeddings for Efficient Word Representations »
Dan Tito Svenstrup · Jonas Hansen · Ole Winther