Workshop: Cooperative AI

Thore Graepel, Dario Amodei, Vincent Conitzer, Allan Dafoe, Gillian Hadfield, Eric Horvitz, Sarit Kraus, Kate Larson, Yoram Bachrach

2020-12-12T05:20:00-08:00 - 2020-12-12T12:55:00-08:00
Abstract: https://www.CooperativeAI.com/

Problems of cooperation—in which agents seek ways to jointly improve their welfare—are ubiquitous and important. They can be found at all scales ranging from our daily routines—such as highway driving, communication via shared language, division of labor, and work collaborations—to our global challenges—such as disarmament, climate change, global commerce, and pandemic preparedness. Arguably, the success of the human species is rooted in our ability to cooperate, in our social intelligence and skills. Since machines powered by artificial intelligence and machine learning are playing an ever greater role in our lives, it will be important to equip them with the skills necessary to cooperate and to foster cooperation.

We see an opportunity for the field of AI, and particularly machine learning, to explicitly focus effort on this class of problems which we term Cooperative AI. The goal of this research would be to study the many aspects of the problem of cooperation, and innovate in AI to contribute to solving these problems. Central questions include how to build machine agents with the capabilities needed for cooperation, and how advances in machine learning can help foster cooperation in populations of agents (of machines and/or humans), such as through improved mechanism design and mediation.

Research could be organized around key capabilities necessary for cooperation, including: understanding other agents, communicating with other agents, constructing cooperative commitments, and devising and negotiating suitable bargains and institutions. Since artificial agents will often act on behalf of particular humans and in ways that are consequential for humans, this research will need to consider how machines can adequately learn human preferences, and how best to integrate human norms and ethics into cooperative arrangements.

We are planning to bring together scholars from diverse backgrounds to discuss how AI research can contribute to the field of cooperation.


Call for Papers
We invite high-quality paper submissions on the following topics (broadly construed, this is not an exhaustive list):

-Multi-agent learning
-Agent cooperation
-Agent communication
-Resolving commitment problems
-Agent societies, organizations and institutions
-Trust and reputation
-Theory of mind and peer modelling
-Markets, mechanism design and and economics based cooperation
-Negotiation and bargaining agents
-Team formation problems

Accepted papers will be presented during joint virtual poster sessions and be made publicly available as non archival reports, allowing future submissions to archival conferences or journals.

Submissions should be up to eight pages excluding references, acknowledgements, and supplementary material, and should follow NeurIPS format. The review process will be double-blind.

Paper submissions: https://easychair.org/my/conference?conf=coopai2020#

Video

Chat

Chat is not available.

Schedule

2020-12-12T05:20:00-08:00 - 2020-12-12T05:30:00-08:00
Welcome: Yoram Bachrach (DeepMind) and Gillian Hadfield (University of Toronto)
Yoram Bachrach, Gillian Hadfield
2020-12-12T05:30:00-08:00 - 2020-12-12T06:00:00-08:00
Open Problems in Cooperative AI: Thore Graepel (DeepMind) and Allan Dafoe (University of Oxford)
Thore Graepel, Allan Dafoe
2020-12-12T06:00:00-08:00 - 2020-12-12T06:30:00-08:00
Invited Speaker: Peter Stone (The University of Texas at Austin) on Ad Hoc Autonomous Agent Teams: Collaboration without Pre-Coordination
Peter Stone
As autonomous agents proliferate in the real world, both in software and robotic settings, they will increasingly need to band together for cooperative activities with previously unfamiliar teammates. In such "ad hoc" team settings, team strategies cannot be developed a priori. Rather, an agent must be prepared to cooperate with many types of teammates: it must collaborate without pre-coordination. This talk will cover past and ongoing research on the challenge of building autonomous agents that are capable of robust ad hoc teamwork.
2020-12-12T06:30:00-08:00 - 2020-12-12T07:00:00-08:00
Invited Speaker: Gillian Hadfield (University of Toronto) on The Normative Infrastructure of Cooperation
Gillian Hadfield
In this talk, I will present the case for the critical role played by third-party enforced rules in the extensive forms of cooperation we see in humans. Cooperation, I’ll argue, cannot be adequately accounted for—or modeled for AI—within the framework of human preferences, coordination incentives or bilateral commitments and reciprocity alone. Cooperation is a group phenomenon and requires group infrastructure to maintain. This insight is critical for training AI agents that can cooperate with humans and, likely, other AI agents. Training environments need to be built with normative infrastructure that enables AI agents to learn and participate in cooperative activities—including the cooperative activity that undergirds all others: collective punishment of agents that violate community norms.
2020-12-12T07:00:00-08:00 - 2020-12-12T07:30:00-08:00
Invited Speaker: James Fearon (Stanford University) on Two Kinds of Cooperative AI Challenges: Game Play and Game Design
James Fearon
Humans routinely face two types of cooperation problems: How to get to a collectively good outcome given some set of preferences and structural constraints; and how to design, shape, or shove structural constraints and preferences to induce agents to make choices that bring about better collective outcomes. In the terminology of economic theory, the first is a problem of equilibrium selection given a game structure, and the second is a problem of mechanism design by a “social planner.” These two types of problems have been distinguished in and are central to a much longer tradition of political philosophy (e.g., state of nature arguments). It is fairly clear how AI can and might be constructively applied to the first type of problem, while less clear for the second type. How to think about using AI to contribute to optimal design of the terms and parameters – the rules of a game – for other agents? Put differently, could there be an AI of constitutional design?
2020-12-12T07:30:00-08:00 - 2020-12-12T08:00:00-08:00
Invited Speaker: Sarit Kraus (Bar-Ilan University) on Agent-Human Collaboration and Learning for Improving Human Satisfaction
Sarit Kraus
We consider environments where a set of human workers needs to handle a large set of tasks while interacting with human users. The arriving tasks vary: they may differ in their urgency, their difficulty and the required knowledge and time duration in which to perform them. Our goal is to decrease the number of workers, which we refer to as operators that are handling the tasks while increasing the users’ satisfaction. We present automated intelligent agents that will work together with the human operators in order to improve the overall performance of such systems and increase both operators' and users’ satisfaction. Examples include: home hospitalization environment where remote specialists will instruct and supervise treatments that are carried out at the patients' homes; operators that tele-operate autonomous vehicles when human intervention is needed and bankers that provide online service to customers. The automated agents could support the operators: the machine learning-based agent follows the operator’s work and makes recommendations, helping him interact proficiently with the users. The agents can also learn from the operators and eventually replace the operators in many of their tasks.
2020-12-12T08:00:00-08:00 - 2020-12-12T08:30:00-08:00
Invited Speaker: William Isaac (DeepMind) on Can Cooperation make AI (and Society) Fairer?
William Isaac
2020-12-12T08:30:00-08:00 - 2020-12-12T08:45:00-08:00
Q&A: Open Problems in Cooperative AI with Thore Graepel (DeepMind), Allan Dafoe (University of Oxford), Yoram Bachrach (DeepMind), and Natasha Jaques (Google) [moderator]
Thore Graepel, Yoram Bachrach, Allan Dafoe, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/ambolxqi
2020-12-12T08:45:00-08:00 - 2020-12-12T09:00:00-08:00
Q&A: Gillian Hadfield (University of Toronto): The Normative Infrastructure of Cooperation, with Natasha Jaques (Google) [moderator]
Gillian Hadfield, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/02lguhzy
2020-12-12T09:00:00-08:00 - 2020-12-12T09:15:00-08:00
Q&A: William Isaac (DeepMind): Can Cooperative Make AI (and Society) Fairer?, with Natasha Jaques (Google) [moderator]
William Isaac, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/riko0stp
2020-12-12T09:15:00-08:00 - 2020-12-12T09:30:00-08:00
Q&A: Peter Stone (The University of Texas at Austin): Ad Hoc Autonomous Agent Teams: Collaboration without Pre-Coordination, with Natasha Jaques (Google) [moderator]
Peter Stone, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/50mlx6cq
2020-12-12T09:30:00-08:00 - 2020-12-12T09:45:00-08:00
Q&A: Sarit Kraus (Bar-Ilan University): Agent-Human Collaboration and Learning for Improving Human Satisfaction, with Natasha Jaques (Google) [moderator]
Sarit Kraus, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/9opzmndo
2020-12-12T09:45:00-08:00 - 2020-12-12T10:00:00-08:00
Q&A: James Fearon (Stanford University): Cooperation Inside and Over the Rules of the Game, with Natasha Jaques (Google) [moderator]
James Fearon, Natasha Jaques
Participants can send questions via Sli.do using this link: https://app.sli.do/event/uqh9pktn
2020-12-12T10:00:00-08:00 - 2020-12-12T11:00:00-08:00
Poster Sessions (hosted in GatherTown)
Gather Town link: https://neurips.gather.town/app /1l0kNMMpqLZvr9Co/CooperativeAI
2020-12-12T11:00:00-08:00 - 2020-12-12T11:45:00-08:00
Panel: Kate Larson (DeepMind) [moderator], Natasha Jaques (Google), Jeffrey Rosenschein (The Hebrew University of Jerusalem), Michael Wooldridge (University of Oxford)
Kate Larson, Natasha Jaques, Jeff S Rosenschein, Michael Wooldridge
2020-12-12T11:45:00-08:00 - 2020-12-12T12:00:00-08:00
Spotlight Talk: Too many cooks: Bayesian inference for coordinating multi-agent collaboration
Rose Wang
Authors: Rose Wang, Sarah Wu, James Evans, Joshua Tenenbaum, David Parkes and Max Kleiman-Weiner
2020-12-12T12:00:00-08:00 - 2020-12-12T12:15:00-08:00
Spotlight Talk: Learning Social Learning
Kamal Ndousse
Authors: Kamal Ndousse, Douglas Eck, Sergey Levine and Natasha Jaques
2020-12-12T12:15:00-08:00 - 2020-12-12T12:30:00-08:00
Spotlight Talk: Benefits of Assistance over Reward Learning
Rohin Shah
Authors: Rohin Shah, Pedro Freire, Neel Alex, Rachel Freedman, Dmitrii Krasheninnikov, Lawrence Chan, Michael Dennis, Pieter Abbeel, Anca Dragan and Stuart Russell
2020-12-12T12:30:00-08:00 - 2020-12-12T12:45:00-08:00
Spotlight Talk: Watch-And-Help: A Challenge for Social Perception and Human-AI Collaboration
Xavier Puig
Authors: Xavier Puig, Tianmin Shu, Shuang Li, Zilin Wang, Josh Tenenbaum, Sanja Fidler and Antonio Torralba
2020-12-12T12:45:00-08:00 - 2020-12-12T12:55:00-08:00
Closing Remarks: Eric Horvitz (Microsoft)
Eric Horvitz