Skip to yearly menu bar Skip to main content


Causal Bandits: Learning Good Interventions via Causal Inference

Finnian Lattimore · Tor Lattimore · Mark Reid

Area 5+6+7+8 #25

Keywords: [ Causality ] [ Bandit Algorithms ]


We study the problem of using causal models to improve the rate at which good interventions can be learned online in a stochastic environment. Our formalism combines multi-arm bandits and causal inference to model a novel type of bandit feedback that is not exploited by existing approaches. We propose a new algorithm that exploits the causal feedback and prove a bound on its simple regret that is strictly better (in all quantities) than algorithms that do not use the additional causal information.

Live content is unavailable. Log in and register to view live content