Timezone: »
Neural networks often learn task-specific latent representations that fail to generalize to novel settings or tasks. Conversely, humans learn discrete representations (i.e., concepts or words) at a variety of abstraction levels (e.g., "bird" vs. "sparrow'") and use the appropriate abstraction based on tasks. Inspired by this, we train neural models to generate a spectrum of discrete representations, and control the complexity of the representations (roughly, how many bits are allocated for encoding inputs) by tuning the entropy of the distribution over representations. In finetuning experiments, using only a small number of labeled examples for a new task, we show that (1) tuning the representation to a task-appropriate complexity level supports the greatest finetuning performance, and (2) in a human-participant study, users were able to identify the appropriate complexity level for a downstream task via visualizations of discrete representations. Our results indicate a promising direction for rapid model finetuning by leveraging human insight.
Author Information
Andi Peng (MIT)
Mycal Tucker (Massachusetts Institute of Technology)
Eoin Kenny (MIT)

I am an explainable AI reseracher (XAI). Previously, I did my Ph.D. at University College Dublin, Ireland. There I worked on post-hoc explanation-by-example with my supervisor Mark Keane. Currently I am researching XAI at MIT alongside Julie Shah, with a focus on contrastive explanation and interpretable deep reinforcement learning. I envision AI systems which can be successfully deployed with useful, human-friendly explanations, so that we can clearly see what they are doing in a way that everyone can understand (not just ML experts). To do this, I use example-based XAI, because it is similar to how humans are thought to reason and has much support in user testing that it is useful and understandable to people. My strongest contributions to the field have been (1) the introduction of Semi-Factual explanation, and (2) designing the first interpretable Deep RL system.
Noga Zaslavsky (UCI)
Pulkit Agrawal (MIT)
Julie A Shah (MIT)
More from the Same Authors
-
2021 : Self-supervised pragmatic reasoning »
Jennifer Hu · Roger Levy · Noga Zaslavsky -
2021 : 3D Neural Scene Representations for Visuomotor Control »
Yunzhu Li · Shuang Li · Vincent Sitzmann · Pulkit Agrawal · Antonio Torralba -
2021 : 3D Neural Scene Representations for Visuomotor Control »
Yunzhu Li · Shuang Li · Vincent Sitzmann · Pulkit Agrawal · Antonio Torralba -
2022 : Trading off Utility, Informativeness, and Complexity in Emergent Communication »
Mycal Tucker · Julie A Shah · Roger Levy · Noga Zaslavsky -
2022 : Is Conditional Generative Modeling all you need for Decision-Making? »
Anurag Ajay · Yilun Du · Abhi Gupta · Josh Tenenbaum · Tommi Jaakkola · Pulkit Agrawal -
2022 : Learning to Extrapolate: A Transductive Approach »
Aviv Netanyahu · Abhishek Gupta · Max Simchowitz · Kaiqing Zhang · Pulkit Agrawal -
2022 : Towards True Lossless Sparse Communication in Multi-Agent Systems »
Seth Karten · Mycal Tucker · Siva Kailas · Katia Sycara -
2022 : Fast Adaptation via Human Diagnosis of Task Distribution Shift »
Andi Peng · Mark Ho · Aviv Netanyahu · Julie A Shah · Pulkit Agrawal -
2022 : Temporal Logic Imitation: Learning Plan-Satisficing Motion Policies from Demonstrations »
Felix Yanwei Wang · Nadia Figueroa · Shen Li · Ankit Shah · Julie A Shah -
2022 : Aligning Robot Representations with Humans »
Andreea Bobu · Andi Peng · Pulkit Agrawal · Julie A Shah · Anca Dragan -
2022 : Generalization and Translatability in Emergent Communication via Informational Constraints »
Mycal Tucker · Roger Levy · Julie A Shah · Noga Zaslavsky -
2023 : Neuro-Inspired Fragmentation and Recall to Overcome Catastrophic Forgetting in Curiosity »
Jaedong Hwang · Zhang-Wei Hong · Eric Chen · Akhilan Boopathy · Pulkit Agrawal · Ila Fiete -
2023 : Increasing Brain-LLM Alignment via Information-Theoretic Compression »
Mycal Tucker · Greta Tuckute -
2023 : In Pursuit of Regulatable LLMs »
Eoin Kenny · Julie A Shah -
2023 : Compositional Foundation Models for Hierarchical Planning »
Anurag Ajay · Seungwook Han · Yilun Du · Shuang Li · Abhi Gupta · Tommi Jaakkola · Josh Tenenbaum · Leslie Kaelbling · Akash Srivastava · Pulkit Agrawal -
2023 : Compositional Foundation Models for Hierarchical Planning »
Anurag Ajay · Seungwook Han · Yilun Du · Shuang Li · Abhi Gupta · Tommi Jaakkola · Josh Tenenbaum · Leslie Kaelbling · Akash Srivastava · Pulkit Agrawal -
2023 : Increasing Brain-LLM Alignment via Information-Theoretic Compression »
Mycal Tucker · Greta Tuckute -
2023 : Information theory, cognition, and deep learning: Challenges and opportunities »
Sarah Marzen · Stephan Mandt · Noah Goodman · Danielle S Bassett · Noga Zaslavsky · Rava Azeredo da Silveira · Ron M. Hecht · Ronit Bustin -
2023 : Noisy Population Dynamics Lead to Efficiently Compressed Semantic Systems »
Nathaniel Imel · Noga Zaslavsky · Michael Franke · Richard Futrell -
2023 Workshop: Goal-Conditioned Reinforcement Learning »
Benjamin Eysenbach · Ishan Durugkar · Jason Ma · Andi Peng · Tongzhou Wang · Amy Zhang -
2023 : Opening Remarks »
Noga Zaslavsky -
2023 Workshop: Information-Theoretic Principles in Cognitive Systems (InfoCog) »
Noga Zaslavsky · Rava Azeredo da Silveira · Ronit Bustin · Ron M. Hecht -
2023 Poster: Self-Supervised Reinforcement Learning that Transfers using Random Features »
Boyuan Chen · Chuning Zhu · Pulkit Agrawal · Kaiqing Zhang · Abhishek Gupta -
2023 Poster: Breadcrumbs to the Goal: Supervised Goal Selection from Human-in-the-Loop Feedback »
Marcel Torne Villasevil · Max Balsells I Pamies · Zihan Wang · Samedh Desai · Tao Chen · Pulkit Agrawal · Abhishek Gupta -
2023 Poster: Compositional Foundation Models for Hierarchical Planning »
Anurag Ajay · Seungwook Han · Yilun Du · Shuang Li · Abhi Gupta · Tommi Jaakkola · Josh Tenenbaum · Leslie Kaelbling · Akash Srivastava · Pulkit Agrawal -
2023 Poster: Beyond Uniform Sampling: Offline Reinforcement Learning with Imbalanced Datasets »
Zhang-Wei Hong · Aviral Kumar · Sathwik Karnik · Abhishek Bhandwaldar · Akash Srivastava · Joni Pajarinen · Romain Laroche · Abhishek Gupta · Pulkit Agrawal -
2023 Poster: The Utility of “Even if” Semifactual Explanation to Optimise Positive Outcomes »
Eoin Kenny · Weipeng Huang -
2022 : Generalization and Translatability in Emergent Communication via Informational Constraints »
Mycal Tucker · Roger Levy · Julie A Shah · Noga Zaslavsky -
2022 Workshop: Information-Theoretic Principles in Cognitive Systems »
Noga Zaslavsky · Mycal Tucker · Sarah Marzen · Irina Higgins · Stephanie Palmer · Samuel J Gershman -
2022 : Opening Remarks »
Noga Zaslavsky -
2022 : Visual Pre-training for Navigation: What Can We Learn from Noise? »
Felix Yanwei Wang · Ching-Yun Ko · Pulkit Agrawal -
2022 Poster: Redeeming intrinsic rewards via constrained optimization »
Eric Chen · Zhang-Wei Hong · Joni Pajarinen · Pulkit Agrawal -
2022 Poster: Distributionally Adaptive Meta Reinforcement Learning »
Anurag Ajay · Abhishek Gupta · Dibya Ghosh · Sergey Levine · Pulkit Agrawal -
2022 Poster: Trading off Utility, Informativeness, and Complexity in Emergent Communication »
Mycal Tucker · Roger Levy · Julie Shah · Noga Zaslavsky -
2021 : [O5] Do Feature Attribution Methods Correctly Attribute Features? »
Yilun Zhou · Serena Booth · Marco Tulio Ribeiro · Julie A Shah -
2021 : 3D Neural Scene Representations for Visuomotor Control »
Yunzhu Li · Shuang Li · Vincent Sitzmann · Pulkit Agrawal · Antonio Torralba -
2021 Workshop: 2nd Workshop on Self-Supervised Learning: Theory and Practice »
Pengtao Xie · Ishan Misra · Pulkit Agrawal · Abdelrahman Mohamed · Shentong Mo · Youwei Liang · Jeannette Bohg · Kristina N Toutanova -
2021 Workshop: Meaning in Context: Pragmatic Communication in Humans and Machines »
Jennifer Hu · Noga Zaslavsky · Aida Nematzadeh · Michael Franke · Roger Levy · Noah Goodman -
2021 : Opening remarks »
Jennifer Hu · Noga Zaslavsky · Aida Nematzadeh · Michael Franke · Roger Levy · Noah Goodman -
2021 Poster: Emergent Discrete Communication in Semantic Spaces »
Mycal Tucker · Huao Li · Siddharth Agrawal · Dana Hughes · Katia Sycara · Michael Lewis · Julie A Shah -
2020 Workshop: Self-Supervised Learning -- Theory and Practice »
Pengtao Xie · Shanghang Zhang · Pulkit Agrawal · Ishan Misra · Cynthia Rudin · Abdelrahman Mohamed · Wenzhen Yuan · Barret Zoph · Laurens van der Maaten · Xingyi Yang · Eric Xing -
2020 Session: Orals & Spotlights Track 09: Reinforcement Learning »
Pulkit Agrawal · Mohammad Ghavamzadeh -
2019 : Panel Discussion »
Jacob Andreas · Edward Gibson · Stefan Lee · Noga Zaslavsky · Jason Eisner · Jürgen Schmidhuber -
2019 : Invited Talk - 2 »
Noga Zaslavsky -
2019 Poster: Superposition of many models into one »
Brian Cheung · Alexander Terekhov · Yubei Chen · Pulkit Agrawal · Bruno Olshausen -
2018 Poster: Bayesian Inference of Temporal Task Specifications from Demonstrations »
Ankit Shah · Pritish Kamath · Julie A Shah · Shen Li -
2017 : Efficient human-like semantic representations via the information bottleneck principle »
Noga Zaslavsky -
2016 : What makes ImageNet good for Transfer Learning? »
Jacob MY Huh · Pulkit Agrawal · Alexei Efros -
2016 : Jitendra Malik and Pulkit Agrawal »
Jitendra Malik · Pulkit Agrawal -
2016 Workshop: The Future of Interactive Machine Learning »
Kory Mathewson @korymath · Kaushik Subramanian · Mark Ho · Robert Loftin · Joseph L Austerweil · Anna Harutyunyan · Doina Precup · Layla El Asri · Matthew Gombolay · Jerry Zhu · Sonia Chernova · Charles Isbell · Patrick M Pilarski · Weng-Keen Wong · Manuela Veloso · Julie A Shah · Matthew Taylor · Brenna Argall · Michael Littman -
2016 Poster: Learning to Poke by Poking: Experiential Learning of Intuitive Physics »
Pulkit Agrawal · Ashvin Nair · Pieter Abbeel · Jitendra Malik · Sergey Levine -
2016 Oral: Learning to Poke by Poking: Experiential Learning of Intuitive Physics »
Pulkit Agrawal · Ashvin Nair · Pieter Abbeel · Jitendra Malik · Sergey Levine -
2015 Poster: Mind the Gap: A Generative Approach to Interpretable Feature Selection and Extraction »
Been Kim · Julie A Shah · Finale Doshi-Velez -
2014 Poster: Fairness in Multi-Agent Sequential Decision-Making »
Chongjie Zhang · Julie A Shah -
2014 Poster: The Bayesian Case Model: A Generative Approach for Case-Based Reasoning and Prototype Classification »
Been Kim · Cynthia Rudin · Julie A Shah