Timezone: »
Solving optimization problems with unknown parameters often requires learning a predictive model to predict the values of the unknown parameters and then solving the problem using these values. Recent work has shown that including the optimization problem as a layer in the model training pipeline results in predictions of the unobserved parameters that lead to higher decision quality. Unfortunately, this process comes at a large computational cost because the optimization problem must be solved and differentiated through in each training iteration; furthermore, it may also sometimes fail to improve solution quality due to non-smoothness issues that arise when training through a complex optimization layer. To address these shortcomings, we learn a low-dimensional surrogate model of a large optimization problem by representing the feasible space in terms of meta-variables, each of which is a linear combination of the original variables. By training a low-dimensional surrogate model end-to-end, and jointly with the predictive model, we achieve: i) a large reduction in training and inference time; and ii) improved performance by focusing attention on the more important variables in the optimization and learning in a smoother space. Empirically, we demonstrate these improvements on a non-convex adversary modeling task, a submodular recommendation task and a convex portfolio optimization task.
Author Information
Kai Wang (Harvard University)
Bryan Wilder (Harvard University)
Andrew Perrault (Harvard University)
Milind Tambe (Harvard University/Google Research India)
Related Events (a corresponding poster, oral, or spotlight)
-
2020 Poster: Automatically Learning Compact Quality-aware Surrogates for Optimization Problems »
Wed. Dec 9th 05:00 -- 07:00 PM Room Poster Session 3 #1044
More from the Same Authors
-
2021 Spotlight: Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning »
Kai Wang · Sanket Shah · Haipeng Chen · Andrew Perrault · Finale Doshi-Velez · Milind Tambe -
2021 : Your Bandit Model is Not Perfect: Introducing Robustness to Restless Bandits Enabled by Deep Reinforcement Learning »
Jackson Killian · Lily Xu · Arpita Biswas · Milind Tambe -
2022 : Fuzzy c-Means Clustering in Persistence Diagram Space for Deep Learning Model Selection »
Thomas Davies · Jack Aspinall · Bryan Wilder · Long Tran-Thanh -
2022 : Case Study: Applying Decision Focused Learning in the Real World »
Shresth Verma · Aditya Mate · Kai Wang · Aparna Taneja · Milind Tambe -
2022 : Invited Talk: Milind Tambe »
Milind Tambe -
2022 Poster: Decision-Focused Learning without Decision-Making: Learning Locally Optimized Decision Losses »
Sanket Shah · Kai Wang · Bryan Wilder · Andrew Perrault · Milind Tambe -
2021 : Invite Talk Q&A »
Milind Tambe · Tejumade Afonja · Paula Rodriguez Diaz -
2021 : Invited Talk: AI for Social Impact: Results from Deployments for Public Health »
Milind Tambe -
2021 Poster: Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning »
Kai Wang · Sanket Shah · Haipeng Chen · Andrew Perrault · Finale Doshi-Velez · Milind Tambe -
2020 : Q/A and Panel Discussion for People-Earth with Dan Kammen and Milind Tambe »
Daniel Kammen · Milind Tambe · Giulio De Leo · Mayur Mudigonda · Surya Karthik Mukkavilli -
2020 : Q/A and Discussion »
Surya Karthik Mukkavilli · Mayur Mudigonda · Milind Tambe -
2020 : Milind Tambe »
Milind Tambe -
2020 Poster: Collapsing Bandits and Their Application to Public Health Intervention »
Aditya Mate · Jackson Killian · Haifeng Xu · Andrew Perrault · Milind Tambe