Timezone: »
Emergent communication (EC) research often focuses on optimizing task-specific utility as a driver for communication. However, there is increasing evidence that human languages are shaped by task-general communicative constraints and evolve under pressure to optimize the Information Bottleneck (IB) tradeoff between the informativeness and complexity of the lexicon. Here, we integrate these two approaches by trading off utility, informativeness, and complexity in EC. To this end, we propose Vector-Quantized Variational Information Bottleneck (VQ-VIB), a method for training neural agents to encode inputs into discrete signals embedded in a continuous space. We evaluate our approach in multi-agent reinforcement learning settings and in color reference games and show that: (1) VQ-VIB agents can continuously adapt to changing communicative needs and, in the color domain, align with human languages; (2) the emergent VQ-VIB embedding spaces are semantically meaningful and perceptually grounded; and (3) encouraging informativeness leads to faster convergence rates and improved utility, both in VQ-VIB and in prior neural architectures for symbolic EC, with VQ-VIB achieving higher utility for any given complexity. This work offers a new framework for EC that is grounded in information-theoretic principles that are believed to characterize human language evolution and that may facilitate human-agent interaction.
Author Information
Mycal Tucker (Massachusetts Institute of Technology)
Roger Levy (Massachusetts Institute of Technology)
Julie Shah (Association for the Advancement of Artificial Intelligence)
Noga Zaslavsky (MIT)
More from the Same Authors
-
2021 : Self-supervised pragmatic reasoning »
Jennifer Hu · Roger Levy · Noga Zaslavsky -
2022 : Trading off Utility, Informativeness, and Complexity in Emergent Communication »
Mycal Tucker · Julie A Shah · Roger Levy · Noga Zaslavsky -
2022 : Towards True Lossless Sparse Communication in Multi-Agent Systems »
Seth Karten · Mycal Tucker · Siva Kailas · Katia Sycara -
2022 : Generalization and Translatability in Emergent Communication via Informational Constraints »
Mycal Tucker · Roger Levy · Julie A Shah · Noga Zaslavsky -
2023 Poster: Human-Guided Complexity-Controlled Abstractions »
Andi Peng · Mycal Tucker · Eoin Kenny · Noga Zaslavsky · Pulkit Agrawal · Julie A Shah -
2023 Workshop: Information-Theoretic Principles in Cognitive Systems (InfoCog) »
Noga Zaslavsky · Rava Azeredo da Silveira · Ronit Bustin · Ron M. Hecht -
2023 : Opening Remarks »
Noga Zaslavsky -
2022 : Generalization and Translatability in Emergent Communication via Informational Constraints »
Mycal Tucker · Roger Levy · Julie A Shah · Noga Zaslavsky -
2022 Workshop: Information-Theoretic Principles in Cognitive Systems »
Noga Zaslavsky · Mycal Tucker · Sarah Marzen · Irina Higgins · Stephanie Palmer · Samuel J Gershman -
2022 : Opening Remarks »
Noga Zaslavsky -
2021 Workshop: Meaning in Context: Pragmatic Communication in Humans and Machines »
Jennifer Hu · Noga Zaslavsky · Aida Nematzadeh · Michael Franke · Roger Levy · Noah Goodman -
2021 : Opening remarks »
Jennifer Hu · Noga Zaslavsky · Aida Nematzadeh · Michael Franke · Roger Levy · Noah Goodman -
2021 Poster: Grammar-Based Grounded Lexicon Learning »
Jiayuan Mao · Freda Shi · Jiajun Wu · Roger Levy · Josh Tenenbaum -
2021 Poster: Emergent Discrete Communication in Semantic Spaces »
Mycal Tucker · Huao Li · Siddharth Agrawal · Dana Hughes · Katia Sycara · Michael Lewis · Julie A Shah -
2019 : Panel Discussion »
Jacob Andreas · Edward Gibson · Stefan Lee · Noga Zaslavsky · Jason Eisner · Jürgen Schmidhuber -
2019 : Invited Talk - 2 »
Noga Zaslavsky -
2017 : Efficient human-like semantic representations via the information bottleneck principle »
Noga Zaslavsky