Poster
Transition Constrained Bayesian Optimization via Markov Decision Processes
Jose Pablo Folch · Calvin Tsay · Robert Lee · Behrang Shafei · Weronika Ormaniec · Andreas Krause · Mark van der Wilk · Ruth Misener · Mojmir Mutny
West Ballroom A-D #6005
Bayesian optimization is a methodology to optimize black-box functions. Traditionally, it focuses on the setting where you can arbitrarily query the search space. However, many real-life problems do not offer this flexibility; in particular, the search space of the next query may depend on previous ones. Example challenges arise in the physical sciences in the form of local movement constraints, required monotonicity in certain variables, and transitions influencing the accuracy of measurements. Altogether, such transition constraints necessitate a form of planning. This work extends classical Bayesian optimization via the framework of Markov Decision Processes. We iteratively solve a tractable linearization of our utility function using reinforcement learning to obtain a policy that plans ahead for the entire horizon. This is a parallel to the optimization of an acquisition function in policy space. The resulting policy is potentially history-dependent and non-Markovian. We showcase applications in chemical reactor optimization, informative path planning, machine calibration, and other synthetic examples.
Live content is unavailable. Log in and register to view live content