Workshop
|
|
Rewards Encoding Environment Dynamics Improves Preference-based Reinforcement Learning
Katherine Metcalf · Miguel Sarabia · Barry-John Theobald
|
|
Workshop
|
|
Efficient Preference-Based Reinforcement Learning Using Learned Dynamics Models
Yi Liu · Gaurav Datta · Ellen Novoseller · Daniel Brown
|
|
Workshop
|
|
PD-MORL: Preference-Driven Multi-Objective Reinforcement Learning Algorithm
Toygun Basaklar · Suat Gumussoy · Umit Ogras
|
|
Workshop
|
|
Towards customizable reinforcement learning agents: Enabling preference specification through online vocabulary expansion
Utkarsh Soni · Sarath Sreedharan · Mudit Verma · Lin Guan · Matthew Marquez · Subbarao Kambhampati
|
|
Poster
|
|
Efficient Meta Reinforcement Learning for Preference-based Fast Adaptation
Zhizhou Ren · Anji Liu · Yitao Liang · Jian Peng · Jianzhu Ma
|
|
Poster
|
|
Meta-Reward-Net: Implicitly Differentiable Reward Learning for Preference-based Reinforcement Learning
Runze Liu · Fengshuo Bai · Yali Du · Yaodong Yang
|
|
Affinity Workshop
|
|
Modelling non-reinforced preferences using selective attention
Noor Sajid · Panagiotis Tigas · Zafeirios Fountas · Qinghai Guo · Alexey Zakharov · Lancelot Da Costa
|
|
Workshop
|
|
Memory-Efficient Reinforcement Learning with Priority based on Surprise and On-policyness
Ryosuke Unno · Yoshimasa Tsuruoka
|
|
Poster
|
|
I2Q: A Fully Decentralized Q-Learning Algorithm
Jiechuan Jiang · Zongqing Lu
|
|
Workshop
|
|
Meta-learning of Black-box Solvers Using Deep Reinforcement Learning
Cedric Malherbe · Aladin Virmaux · Ludovic Dos Santos · Sofian Chaybouti
|
|
Poster
|
|
Mildly Conservative Q-Learning for Offline Reinforcement Learning
Jiafei Lyu · Xiaoteng Ma · Xiu Li · Zongqing Lu
|
|
Poster
|
|
Multiagent Q-learning with Sub-Team Coordination
Wenhan Huang · Kai Li · Kun Shao · Tianze Zhou · Matthew Taylor · Jun Luo · Dongge Wang · Hangyu Mao · Jianye Hao · Jun Wang · Xiaotie Deng
|
|