Timezone: »
Reinforcement learning (RL) has succeeded in many challenging tasks such as Atari, Go, and Chess and even in high dimensional continuous domains such as robotics. Most impressive successes are in tasks where the agent observes the task features fully. However, in real world problems, the agent usually can only rely on partial observations. In real time games the agent makes only local observations; in robotics the agent has to cope with noisy sensors, occlusions, and unknown dynamics. Even more fundamentally, any agent without a full a priori world model or without full access to the system state, has to make decisions based on partial knowledge about the environment and its dynamics.
Reinforcement learning under partial observability has been tackled in the operations research, control, planning, and machine learning communities. One of the goals of the workshop is to bring researchers from different backgrounds together. Moreover, the workshop aims to highlight future applications. In addition to robotics where partial observability is a well known challenge, many diverse applications such as wireless networking, human-robot interaction and autonomous driving require taking partial observability into account.
Partial observability introduces unique challenges: the agent has to remember the past but also connect the present with potential futures requiring memory, exploration, and value propagation techniques that can handle partial observability. Current model-based methods can handle discrete values and take long term information gathering into account while model-free methods can handle high-dimensional continuous problems but often assume that the state space has been created for the problem at hand such that there is sufficient information for optimal decision making or just add memory to the policy without taking partial observability explicitly into account.
In this workshop, we want to go further and ask among others the following questions.
* How can we extend deep RL methods to robustly solve partially observable problems?
* Can we learn concise abstractions of history that are sufficient for high-quality decision-making?
* There have been several successes in decision making under partial observability despite the inherent challenges. Can we characterize problems where computing good policies is feasible?
* Since decision making is hard under partial observability do we want to use more complex models and solve them approximately or use (inaccurate) simple models and solve them exactly? Or not use models at all?
* How can we use control theory together with reinforcement learning to advance decision making under partial observability?
* Can we combine the strengths of model-based and model-free methods under partial observability?
* Can recent method improvements in general RL already tackle some partially observable applications which were not previously possible?
* How do we scale up reinforcement learning in multi-agent systems with partial observability?
* Do hierarchical models / temporal abstraction improve RL efficiency under partial observability?
Sat 5:30 a.m. - 5:40 a.m.
|
Opening Remarks
|
🔗 |
Sat 5:40 a.m. - 6:05 a.m.
|
Joelle Pineau
(
Talk
)
|
Joelle Pineau 🔗 |
Sat 6:05 a.m. - 6:30 a.m.
|
Leslie Kaelbling
(
Talk
)
|
Leslie Kaelbling 🔗 |
Sat 6:30 a.m. - 6:45 a.m.
|
Contributed Talk 1: High-Level Strategy Selection under Partial Observability in StarCraft: Brood War
(
Talk
)
|
🔗 |
Sat 6:45 a.m. - 7:10 a.m.
|
David Silver
(
Talk
)
|
David Silver 🔗 |
Sat 7:10 a.m. - 7:25 a.m.
|
Contributed Talk 2: Joint Belief Tracking and Reward Optimization through Approximate Inference
(
Talk
)
|
🔗 |
Sat 8:00 a.m. - 8:15 a.m.
|
Contributed Talk 3: Learning Dexterous In-Hand Manipulation
(
Talk
)
|
🔗 |
Sat 8:15 a.m. - 8:40 a.m.
|
Pieter Abbeel
(
Talk
)
|
Pieter Abbeel 🔗 |
Sat 8:40 a.m. - 9:00 a.m.
|
Spotlights & Poster Session
(
Spotlights
)
|
James A Preiss · Alexander Grishin · Ville Kyrki · Pol Moreno Comellas · Akshay Narayan · Tze-Yun Leong · Yongxi Tan · Lilian Weng · Toshiharu Sugawara · Kenny Young · Tianmin Shu · Jonas Gehring · Ahmad Beirami · Chris Amato · sammie katt · Andrea Baisero · Arseny Kuznetsov · Jan Humplik · Vladimír Petrík
|
Sat 11:00 a.m. - 11:25 a.m.
|
Peter Stone
(
Talk
)
|
Peter Stone 🔗 |
Sat 11:25 a.m. - 11:40 a.m.
|
Contributed Talk 4: Differentiable Algorithm Networks: Learning Wrong Models for Wrong Algorithms
(
Talk
)
|
🔗 |
Sat 11:40 a.m. - 12:05 p.m.
|
Jilles Dibangoye
(
Talk
)
|
Jilles Dibangoye 🔗 |
Sat 12:35 p.m. - 1:00 p.m.
|
Anca Dragan
(
Talk
)
|
Anca Dragan 🔗 |
Sat 1:00 p.m. - 2:00 p.m.
|
Panel Discussion
|
🔗 |
Sat 2:00 p.m. - 3:30 p.m.
|
Poster Session
|
🔗 |
Author Information
Joni Pajarinen (TU Darmstadt)
Chris Amato (Northeastern University)
Pascal Poupart (University of Waterloo & RBC Borealis AI)
David Hsu (National University of Singapore)
More from the Same Authors
-
2022 Poster: Optimality and Stability in Non-Convex Smooth Games »
Guojun Zhang · Pascal Poupart · Yaoliang Yu -
2022 : Attribute Controlled Dialogue Prompting »
Runcheng Liu · Ahmad Rashid · Ivan Kobyzev · Mehdi Rezaghoizadeh · Pascal Poupart -
2022 : Deep Transformer Q-Networks for Partially Observable Reinforcement Learning »
Kevin Esslinger · Robert Platt · Christopher Amato -
2022 : Geometric attacks on batch normalization »
Amur Ghose · Apurv Gupta · Yaoliang Yu · Pascal Poupart -
2022 Spotlight: Optimality and Stability in Non-Convex Smooth Games »
Guojun Zhang · Pascal Poupart · Yaoliang Yu -
2022 : Attribute Controlled Dialogue Prompting »
Runcheng Liu · Ahmad Rashid · Ivan Kobyzev · Mehdi Rezaghoizadeh · Pascal Poupart -
2022 Workshop: Second Workshop on Efficient Natural Language and Speech Processing (ENLSP-II) »
Mehdi Rezagholizadeh · Peyman Passban · Yue Dong · Lili Mou · Pascal Poupart · Ali Ghodsi · Qun Liu -
2022 Poster: Asynchronous Actor-Critic for Multi-Agent Reinforcement Learning »
Yuchen Xiao · Weihao Tan · Christopher Amato -
2022 Poster: Receding Horizon Inverse Reinforcement Learning »
Yiqing Xu · Wei Gao · David Hsu -
2022 Poster: Shield Decentralization for Safe Multi-Agent Reinforcement Learning »
Daniel Melcer · Christopher Amato · Stavros Tripakis -
2022 Poster: Uncertainty-Aware Reinforcement Learning for Risk-Sensitive Player Evaluation in Sports Game »
Guiliang Liu · Yudong Luo · Oliver Schulte · Pascal Poupart -
2021 : (Live) Panel Discussion: Cooperative AI »
Kalesha Bullard · Allan Dafoe · Fei Fang · Chris Amato · Elizabeth M. Adams -
2021 : Best Papers and Closing Remarks »
Ali Ghodsi · Pascal Poupart -
2021 : Panel Discussion »
Pascal Poupart · Ali Ghodsi · Luke Zettlemoyer · Sameer Singh · Kevin Duh · Yejin Choi · Lu Hou -
2021 Workshop: Efficient Natural Language and Speech Processing (Models, Training, and Inference) »
Mehdi Rezaghoizadeh · Lili Mou · Yue Dong · Pascal Poupart · Ali Ghodsi · Qun Liu -
2021 : Opening Speech »
Pascal Poupart -
2021 Poster: Quantifying and Improving Transferability in Domain Generalization »
Guojun Zhang · Han Zhao · Yaoliang Yu · Pascal Poupart -
2021 Poster: Learning Tree Interpretation from Object Representation for Deep Reinforcement Learning »
Guiliang Liu · Xiangyu Sun · Oliver Schulte · Pascal Poupart -
2020 Poster: Self-Paced Deep Reinforcement Learning »
Pascal Klink · Carlo D'Eramo · Jan Peters · Joni Pajarinen -
2020 Oral: Self-Paced Deep Reinforcement Learning »
Pascal Klink · Carlo D'Eramo · Jan Peters · Joni Pajarinen -
2020 Poster: Learning Agent Representations for Ice Hockey »
Guiliang Liu · Oliver Schulte · Pascal Poupart · Mike Rudd · Mehrsan Javan -
2020 Poster: Learning Dynamic Belief Graphs to Generalize on Text-Based Games »
Ashutosh Adhikari · Xingdi Yuan · Marc-Alexandre Côté · Mikuláš Zelinka · Marc-Antoine Rondeau · Romain Laroche · Pascal Poupart · Jian Tang · Adam Trischler · Will Hamilton -
2019 Poster: Reconciling λ-Returns with Experience Replay »
Brett Daley · Christopher Amato -
2018 : Spotlights & Poster Session »
James A Preiss · Alexander Grishin · Ville Kyrki · Pol Moreno Comellas · Akshay Narayan · Tze-Yun Leong · Yongxi Tan · Lilian Weng · Toshiharu Sugawara · Kenny Young · Tianmin Shu · Jonas Gehring · Ahmad Beirami · Chris Amato · sammie katt · Andrea Baisero · Arseny Kuznetsov · Jan Humplik · Vladimír Petrík -
2018 Poster: Deep Homogeneous Mixture Models: Representation, Separation, and Approximation »
Priyank Jaini · Pascal Poupart · Yaoliang Yu -
2018 Poster: Online Structure Learning for Feed-Forward and Recurrent Sum-Product Networks »
Agastya Kalra · Abdullah Rashwan · Wei-Shou Hsu · Pascal Poupart · Prashant Doshi · George Trimponias -
2018 Poster: Unsupervised Video Object Segmentation for Deep Reinforcement Learning »
Vikash Goel · Jameson Weng · Pascal Poupart -
2018 Poster: Monte-Carlo Tree Search for Constrained POMDPs »
Jongmin Lee · Geon-Hyeong Kim · Pascal Poupart · Kee-Eung Kim -
2017 Poster: QMDP-Net: Deep Learning for Planning under Partial Observability »
Peter Karkus · David Hsu · Wee Sun Lee -
2016 Poster: Online Bayesian Moment Matching for Topic Modeling with Unknown Number of Topics »
Wei-Shou Hsu · Pascal Poupart -
2016 Poster: A Unified Approach for Learning the Parameters of Sum-Product Networks »
Han Zhao · Pascal Poupart · Geoffrey Gordon -
2015 Poster: Adaptive Stochastic Optimization: From Sets to Paths »
Zhan Wei Lim · David Hsu · Wee Sun Lee -
2013 Poster: DESPOT: Online POMDP Planning with Regularization »
Adhiraj Somani · Nan Ye · David Hsu · Wee Sun Lee -
2011 Poster: Periodic Finite State Controllers for Efficient POMDP and DEC-POMDP Planning »
Joni Pajarinen · Jaakko Peltonen -
2011 Poster: Monte Carlo Value Iteration with Macro-Actions »
Zhan Wei Lim · David Hsu · Wee Sun Lee -
2007 Spotlight: What makes some POMDP problems easy to approximate? »
David Hsu · Wee Sun Lee · Nan Rong -
2007 Poster: What makes some POMDP problems easy to approximate? »
David Hsu · Wee Sun Lee · Nan Rong