Timezone: »
A generative recurrent neural network is quickly trained in an unsupervised manner to model popular reinforcement learning environments through compressed spatio-temporal representations. The world model's extracted features are fed into compact and simple policies trained by evolution, achieving state of the art results in various environments. We also train our agent entirely inside of an environment generated by its own internal world model, and transfer this policy back into the actual environment. Interactive version of this paper is available at https://worldmodels.github.io
Author Information
David Ha (Google Brain)
Jürgen Schmidhuber (Swiss AI Lab, IDSIA (USI & SUPSI) - NNAISENSE)
Since age 15 or so, the main goal of professor Jürgen Schmidhuber has been to build a self-improving Artificial Intelligence (AI) smarter than himself, then retire. His lab's Deep Learning Neural Networks based on ideas published in the "Annus Mirabilis" 1990-1991 have revolutionised machine learning and AI. By the mid 2010s, they were on 3 billion devices, and used billions of times per day through users of the world's most valuable public companies, e.g., for greatly improved (CTC-LSTM-based) speech recognition on all Android phones, greatly improved machine translation through Google Translate and Facebook (over 4 billion LSTM-based translations per day), Apple's Siri and Quicktype on all iPhones, the answers of Amazon's Alexa, and numerous other applications. In 2011, his team was the first to win official computer vision contests through deep neural nets, with superhuman performance. In 2012, they had the first deep NN to win a medical imaging contest (on cancer detection). All of this attracted enormous interest from industry. His research group also established the fields of mathematically rigorous universal AI and recursive self-improvement in metalearning machines that learn to learn (since 1987). In 1990, he introduced unsupervised adversarial neural networks that fight each other in a minimax game to achieve artificial curiosity (GANs are a special case). In 1991, he introduced very deep learning through unsupervised pre-training, and neural fast weight programmers formally equivalent to what's now called linear Transformers. His formal theory of creativity & curiosity & fun explains art, science, music, and humor. He also generalized algorithmic information theory and the many-worlds theory of physics, and introduced the concept of Low-Complexity Art, the information age's extreme form of minimal art. He is recipient of numerous awards, author of over 350 peer-reviewed papers, and Chief Scientist of the company NNAISENSE, which aims at building the first practical general purpose AI. He is a frequent keynote speaker, and advising various governments on AI strategies.
Related Events (a corresponding poster, oral, or spotlight)
-
2018 Oral: Recurrent World Models Facilitate Policy Evolution »
Wed. Dec 5th 09:25 -- 09:40 PM Room Room 220 CD
More from the Same Authors
-
2021 : Learning Adaptive Control Flow in Transformers for Improved Systematic Generalization »
Róbert Csordás · Kazuki Irie · Jürgen Schmidhuber -
2021 : Augmenting Classic Algorithms with Neural Components for Strong Generalisation on Ambiguous and High-Dimensional Data »
Imanol Schlag · Jürgen Schmidhuber -
2021 : Improving Baselines in the Wild »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : A Modern Self-Referential Weight Matrix That Learns to Modify Itself »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 : Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2021 : Unsupervised Learning of Temporal Abstractions using Slot-based Transformers »
Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber · Sjoerd van Steenkiste -
2022 : Learning to Control Rapidly Changing Synaptic Connections: An Alternative Type of Memory in Sequence Processing Artificial Neural Networks »
Kazuki Irie · Jürgen Schmidhuber -
2022 : On Narrative Information and the Distillation of Stories »
Dylan Ashley · Vincent Herrmann · Zachary Friggstad · Jürgen Schmidhuber -
2022 : The Benefits of Model-Based Generalization in Reinforcement Learning »
Kenny Young · Aditya Ramesh · Louis Kirsch · Jürgen Schmidhuber -
2023 Poster: Contrastive Training of Complex-Valued Autoencoders for Object Discovery »
Aleksandar Stanić · Anand Gopalakrishnan · Kazuki Irie · Jürgen Schmidhuber -
2022 : Learning gaze control, external attention, and internal attention since 1990-91 »
Jürgen Schmidhuber -
2022 Poster: Neural Differential Equations for Learning to Program Neural Nets Through Continuous Learning Rules »
Kazuki Irie · Francesco Faccio · Jürgen Schmidhuber -
2022 Poster: Exploring through Random Curiosity with General Value Functions »
Aditya Ramesh · Louis Kirsch · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2021 : Panel Discussion 1 »
Megan Peters · Jürgen Schmidhuber · Simona Ghetti · Nick Roy · Oiwi Parker Jones · Ingmar Posner -
2021 : Credit Assignment & Meta-Learning in a Single Lifelong Trial »
Jürgen Schmidhuber -
2021 Poster: Going Beyond Linear Transformers with Recurrent Fast Weight Programmers »
Kazuki Irie · Imanol Schlag · Róbert Csordás · Jürgen Schmidhuber -
2021 Poster: Meta Learning Backpropagation And Improving It »
Louis Kirsch · Jürgen Schmidhuber -
2019 : Panel Discussion »
Jacob Andreas · Edward Gibson · Stefan Lee · Noga Zaslavsky · Jason Eisner · Jürgen Schmidhuber -
2019 : Innate Bodies, Innate Brains, and Innate World Models »
David Ha -
2019 Poster: Are Disentangled Representations Helpful for Abstract Visual Reasoning? »
Sjoerd van Steenkiste · Francesco Locatello · Jürgen Schmidhuber · Olivier Bachem -
2019 Poster: Weight Agnostic Neural Networks »
Adam Gaier · David Ha -
2019 Poster: Learning to Predict Without Looking Ahead: World Models Without Forward Prediction »
Daniel Freeman · David Ha · Luke Metz -
2019 Spotlight: Weight Agnostic Neural Networks »
Adam Gaier · David Ha -
2018 : David Ha »
David Ha -
2018 : Invited Speaker #4 Juergen Schmidhuber »
Jürgen Schmidhuber -
2018 Poster: Learning to Reason with Third Order Tensor Products »
Imanol Schlag · Jürgen Schmidhuber -
2017 : Morning panel discussion »
Jürgen Schmidhuber · Noah Goodman · Anca Dragan · Pushmeet Kohli · Dhruv Batra -
2017 : HRL with gradient-based subgoal generators, asymptotically optimal incremental problem solvers, various meta-learners, and PowerPlay (Jürgen Schmidhuber) »
Jürgen Schmidhuber -
2017 : Invited Talk »
Jürgen Schmidhuber -
2017 Workshop: Machine Learning for Creativity and Design »
Douglas Eck · David Ha · S. M. Ali Eslami · Sander Dieleman · Rebecca Fiebrink · Luba Elliott -
2017 Poster: Neural Expectation Maximization »
Klaus Greff · Sjoerd van Steenkiste · Jürgen Schmidhuber -
2016 : Juergen Schmidhuber (Scientific Director of the Swiss AI Lab IDSIA) »
Jürgen Schmidhuber -
2016 Symposium: Recurrent Neural Networks and Other Machines that Learn Algorithms »
Jürgen Schmidhuber · Sepp Hochreiter · Alex Graves · Rupesh K Srivastava -
2016 Poster: Tagger: Deep Unsupervised Perceptual Grouping »
Klaus Greff · Antti Rasmus · Mathias Berglund · Hotloo Xiranood · Harri Valpola · Jürgen Schmidhuber -
2015 : Deep Learning RNNaissance »
Jürgen Schmidhuber -
2015 : On General Problem Solving and How to Learn an Algorithm »
Jürgen Schmidhuber -
2015 Poster: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Spotlight: Training Very Deep Networks »
Rupesh K Srivastava · Klaus Greff · Jürgen Schmidhuber -
2015 Poster: Parallel Multi-Dimensional LSTM, With Application to Fast Biomedical Volumetric Image Segmentation »
Marijn F Stollenga · Wonmin Byeon · Marcus Liwicki · Jürgen Schmidhuber -
2014 Poster: Deep Networks with Internal Selective Attention through Feedback Connections »
Marijn F Stollenga · Jonathan Masci · Faustino Gomez · Jürgen Schmidhuber -
2013 Poster: Compete to Compute »
Rupesh K Srivastava · Jonathan Masci · Sohrob Kazerounian · Faustino Gomez · Jürgen Schmidhuber -
2012 Poster: Deep Neural Networks Segment Neuronal Membranes in Electron Microscopy Images »
Dan Ciresan · Alessandro Giusti · luca Maria Gambardella · Jürgen Schmidhuber -
2010 Poster: Improving the Asymptotic Performance of Markov Chain Monte-Carlo by Inserting Vortices »
Yi Sun · Faustino Gomez · Jürgen Schmidhuber -
2008 Poster: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2008 Spotlight: Offline Handwriting Recognition with Multidimensional Recurrent Neural Networks »
Alex Graves · Jürgen Schmidhuber -
2007 Poster: Unconstrained On-line Handwriting Recognition with Recurrent Neural Networks »
Alex Graves · Santiago Fernandez · Marcus Liwicki · Horst Bunke · Jürgen Schmidhuber