Timezone: »
A disentangled representation encodes information about the salient factors of variation in the data independently. Although it is often argued that this representational format is useful in learning to solve many real-world down-stream tasks, there is little empirical evidence that supports this claim. In this paper, we conduct a large-scale study that investigates whether disentangled representations are more suitable for abstract reasoning tasks. Using two new tasks similar to Raven's Progressive Matrices, we evaluate the usefulness of the representations learned by 360 state-of-the-art unsupervised disentanglement models. Based on these representations, we train 3600 abstract reasoning models and observe that disentangled representations do in fact lead to better down-stream performance. In particular, they enable quicker learning using fewer samples.
Author Information
Sjoerd van Steenkiste (The Swiss AI Lab - IDSIA)
Francesco Locatello (ETH Zürich - MPI Tübingen)
Jürgen Schmidhuber (Swiss AI Lab, IDSIA (USI & SUPSI) - NNAISENSE)
Since age 15 or so, the main goal of professor Jürgen Schmidhuber has been to build a self-improving Artificial Intelligence (AI) smarter than himself, then retire. His lab's Deep Learning Neural Networks based on ideas published in the "Annus Mirabilis" 1990-1991 have revolutionised machine learning and AI. By the mid 2010s, they were on 3 billion devices, and used billions of times per day through users of the world's most valuable public companies, e.g., for greatly improved (CTC-LSTM-based) speech recognition on all Android phones, greatly improved machine translation through Google Translate and Facebook (over 4 billion LSTM-based translations per day), Apple's Siri and Quicktype on all iPhones, the answers of Amazon's Alexa, and numerous other applications. In 2011, his team was the first to win official computer vision contests through deep neural nets, with superhuman performance. In 2012, they had the first deep NN to win a medical imaging contest (on cancer detection). All of this attracted enormous interest from industry. His research group also established the fields of mathematically rigorous universal AI and recursive self-improvement in metalearning machines that learn to learn (since 1987). In 1990, he introduced unsupervised adversarial neural networks that fight each other in a minimax game to achieve artificial curiosity (GANs are a special case). In 1991, he introduced very deep learning through unsupervised pre-training, and neural fast weight programmers formally equivalent to what's now called linear Transformers. His formal theory of creativity & curiosity & fun explains art, science, music, and humor. He also generalized algorithmic information theory and the many-worlds theory of physics, and introduced the concept of Low-Complexity Art, the information age's extreme form of minimal art. He is recipient of numerous awards, author of over 350 peer-reviewed papers, and Chief Scientist of the company NNAISENSE, which aims at building the first practical general purpose AI. He is a frequent keynote speaker, and advising various governments on AI strategies.
Olivier Bachem (Google Brain)
More from the Same Authors
-
2021 : Brax - A Differentiable Physics Engine for Large Scale Rigid Body Simulation »
Daniel Freeman · Erik Frey · Anton Raichuk · Sertan Girgin · Igor Mordatch · Olivier Bachem -
2021 : Learning Adaptive Control Flow in Transformers for Improved Systematic Generalization »
Róbert Csordás · Kazuki Irie · Jürgen Schmidhuber -
2021 : Augmenting Classic Algorithms with Neural Components for Strong Generalisation on Ambiguous and High-Dimensional Data »
Imanol Schlag · Jürgen Schmidhuber -
2021 : Improving Baselines in the Wild »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : A Modern Self-Referential Weight Matrix That Learns to Modify Itself »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2022 : Learning to Control Rapidly Changing Synaptic Connections: An Alternative Type of Memory in Sequence Processing Artificial Neural Networks »
Kazuki Irie · Jürgen Schmidhuber -
2022 : On Narrative Information and the Distillation of Stories »
Dylan Ashley · Vincent Herrmann · Zachary Friggstad · Jürgen Schmidhuber -
2022 : The Benefits of Model-Based Generalization in Reinforcement Learning »
Kenny Young · Aditya Ramesh · Louis Kirsch · Jürgen Schmidhuber -
2022 : Learning gaze control, external attention, and internal attention since 1990-91 »
Jürgen Schmidhuber -
2022 Poster: Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules »
Kazuki Irie · Francesco Faccio · Jürgen Schmidhuber -
2022 Poster: Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Panel Discussion 1 »
Megan Peters · Jürgen Schmidhuber · Simona Ghetti · Nick Roy · Oiwi Parker Jones · Ingmar Posner -
2021 : Credit Assignment & Meta-Learning in a Single Lifelong Trial »
Jürgen Schmidhuber -
2021 Poster: Going Beyond Linear Transformers with Recurrent Fast Weight Programmers »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 Poster: What Matters for Adversarial Imitation Learning? »
Manu Orsini · Anton Raichuk · Leonard Hussenot · Damien Vincent · Robert Dadashi · Sertan Girgin · Matthieu Geist · Olivier Bachem · Olivier Pietquin · Marcin Andrychowicz -
2021 Poster: Meta Learning Backpropagation And Improving It »
Louis Kirsch · Jürgen Schmidhuber -
2020 Workshop: Object Representations for Learning and Reasoning »
William Agnew · Rim Assouel · Michael Chang · Antonia Creswell · Eliza Kosoy · Aravind Rajeswaran · Sjoerd van Steenkiste -
2020 Poster: Object-Centric Learning with Slot Attention »
Francesco Locatello · Dirk Weissenborn · Thomas Unterthiner · Aravindh Mahendran · Georg Heigold · Jakob Uszkoreit · Alexey Dosovitskiy · Thomas Kipf -
2020 Spotlight: Object-Centric Learning with Slot Attention »
Francesco Locatello · Dirk Weissenborn · Thomas Unterthiner · Aravindh Mahendran · Georg Heigold · Jakob Uszkoreit · Alexey Dosovitskiy · Thomas Kipf -
2019 : Panel Discussion »
Jacob Andreas · Edward Gibson · Stefan Lee · Noga Zaslavsky · Jason Eisner · Jürgen Schmidhuber -
2019 Poster: On the Fairness of Disentangled Representations »
Francesco Locatello · Gabriele Abbati · Thomas Rainforth · Stefan Bauer · Bernhard Schölkopf · Olivier Bachem -
2019 Poster: On the Transfer of Inductive Bias from Simulation to the Real World: a New Disentanglement Dataset »
Muhammad Waleed Gondal · Manuel Wuethrich · Djordje Miladinovic · Francesco Locatello · Martin Breidt · Valentin Volchkov · Joel Akpo · Olivier Bachem · Bernhard Schölkopf · Stefan Bauer -
2019 Poster: Stochastic Frank-Wolfe for Composite Convex Minimization »
Francesco Locatello · Alp Yurtsever · Olivier Fercoq · Volkan Cevher -
2018 : Invited Speaker #4 Juergen Schmidhuber »
Jürgen Schmidhuber -
2018 Poster: Boosting Black Box Variational Inference »
Francesco Locatello · Gideon Dresdner · Rajiv Khanna · Isabel Valera · Gunnar Ratsch -
2018 Spotlight: Boosting Black Box Variational Inference »
Francesco Locatello · Gideon Dresdner · Rajiv Khanna · Isabel Valera · Gunnar Ratsch -
2018 Poster: Recurrent World Models Facilitate Policy Evolution »
David Ha · Jürgen Schmidhuber -
2018 Poster: Assessing Generative Models via Precision and Recall »
Mehdi S. M. Sajjadi · Olivier Bachem · Mario Lucic · Olivier Bousquet · Sylvain Gelly -
2018 Oral: Recurrent World Models Facilitate Policy Evolution »
David Ha · Jürgen Schmidhuber -
2018 Poster: Learning to Reason with Third Order Tensor Products »
Imanol Schlag · Jürgen Schmidhuber -
2017 : Morning panel discussion »
Jürgen Schmidhuber · Noah Goodman · Anca Dragan · Pushmeet Kohli · Dhruv Batra -
2017 : HRL with gradient-based subgoal generators, asymptotically optimal incremental problem solvers, various meta-learners, and PowerPlay (Jürgen Schmidhuber) »
Jürgen Schmidhuber -
2017 : Relational neural expectation maximization »
Sjoerd van Steenkiste -
2017 : Poster Spotlights »
Francesco Locatello · Ari Pakman · Da Tang · Thomas Rainforth · Zalan Borsos · Marko Järvenpää · Eric Nalisnick · Gabriele Abbati · XIAOYU LU · Jonathan Huggins · Rachit Singh · Rui Luo -
2017 : Invited Talk »
Jürgen Schmidhuber -
2017 Poster: Greedy Algorithms for Cone Constrained Optimization with Convergence Guarantees »
Francesco Locatello · Michael Tschannen · Gunnar Ratsch · Martin Jaggi -
2017 Poster: Neural Expectation Maximization »
Klaus Greff · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2016 : Juergen Schmidhuber (Scientific Director of the Swiss AI Lab IDSIA) »
Jürgen Schmidhuber -
2016 Symposium: Recurrent Neural Networks and Other Machines that Learn Algorithms »
Jürgen Schmidhuber · Sepp Hochreiter · Alex Graves · Rupesh K Srivastava -
2016 Poster: Tagger: Deep Unsupervised Perceptual Grouping »
Klaus Greff · Antti Rasmus · Mathias Berglund · Hotloo Xiranood · Harri Valpola · Jürgen Schmidhuber -
2015 : Deep Learning RNNaissance »
Jürgen Schmidhuber -
2015 : On General Problem Solving and How to Learn an Algorithm »
Jürgen Schmidhuber -
2015 Poster: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Spotlight: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Poster: Parallel Multi-Dimensional LSTM, With Application to Fast Biomedical Volumetric Image Segmentation »
Marijn F Stollenga · Wonmin Byeon · Marcus Liwicki · Jürgen Schmidhuber -
2014 Poster: Deep Networks with Internal Selective Attention through Feedback Connections »
Marijn F Stollenga · Jonathan Masci · Faustino Gomez · Jürgen Schmidhuber -
2013 Poster: Compete to Compute »
Rupesh K Srivastava · Jonathan Masci · Sohrob Kazerounian · Faustino Gomez · Jürgen Schmidhuber -
2012 Poster: Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images »
Dan Ciresan · Alessandro Giusti · luca Maria Gambardella · Jürgen Schmidhuber -
2010 Poster: Improving the Asymptotic Performance of Markov Chain Monte-Carlo by Inserting Vortices »
Yi Sun · Faustino Gomez · Jürgen Schmidhuber -
2008 Poster: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2008 Spotlight: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2007 Poster: Unconstrained On-line Handwriting Recognition with Recurrent Neural Networks »
Alex Graves · Santiago Fernandez · Marcus Liwicki · Horst Bunke · Jürgen Schmidhuber