Timezone: »
Decision-Focused Learning (DFL) is a paradigm for tailoring a predictive model to a downstream optimization task that uses its predictions in order to perform better \textit{on that specific task}. The main technical challenge associated with DFL is that it requires being able to differentiate through the optimization problem, which is difficult due to discontinuous solutions and other challenges. Past work has largely gotten around this this issue by \textit{handcrafting} task-specific surrogates to the original optimization problem that provide informative gradients when differentiated through. However, the need to handcraft surrogates for each new task limits the usability of DFL. In addition, there are often no guarantees about the convexity of the resulting surrogates and, as a result, training a predictive model using them can lead to inferior local optima. In this paper, we do away with surrogates altogether and instead \textit{learn} loss functions that capture task-specific information. To the best of our knowledge, ours is the first approach that entirely replaces the optimization component of decision-focused learning with a loss that is automatically learned. Our approach (a) only requires access to a black-box oracle that can solve the optimization problem and is thus \textit{generalizable}, and (b) can be \textit{convex by construction} and so can be easily optimized over. We evaluate our approach on three resource allocation problems from the literature and find that our approach outperforms learning without taking into account task-structure in all three domains, and even hand-crafted surrogates from the literature.
Author Information
Sanket Shah (Harvard University)

I am a third-year PhD student at Harvard University advised by Prof. Milind Tambe. My current work focuses on Decision-Focused Learning, a paradigm for tailoring a predictive model for a downstream optimization task that uses its predictions.
Kai Wang (Harvard University)
Bryan Wilder (Harvard University)
Andrew Perrault (The Ohio State University)
Milind Tambe (Harvard University/Google Research)
More from the Same Authors
-
2021 Spotlight: Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning »
Kai Wang · Sanket Shah · Haipeng Chen · Andrew Perrault · Finale Doshi-Velez · Milind Tambe -
2021 : Your Bandit Model is Not Perfect: Introducing Robustness to Restless Bandits Enabled by Deep Reinforcement Learning »
Jackson Killian · Lily Xu · Arpita Biswas · Milind Tambe -
2022 : Fuzzy c-Means Clustering in Persistence Diagram Space for Deep Learning Model Selection »
Thomas Davies · Jack Aspinall · Bryan Wilder · Long Tran-Thanh -
2022 : Case Study: Applying Decision Focused Learning in the Real World »
Shresth Verma · Aditya Mate · Kai Wang · Aparna Taneja · Milind Tambe -
2022 : Invited Talk: Milind Tambe »
Milind Tambe -
2021 : Invite Talk Q&A »
Milind Tambe · Tejumade Afonja · Paula Rodriguez Diaz -
2021 : Invited Talk: AI for Social Impact: Results from Deployments for Public Health »
Milind Tambe -
2021 Poster: Learning MDPs from Features: Predict-Then-Optimize for Sequential Decision Making by Reinforcement Learning »
Kai Wang · Sanket Shah · Haipeng Chen · Andrew Perrault · Finale Doshi-Velez · Milind Tambe -
2020 : Q/A and Panel Discussion for People-Earth with Dan Kammen and Milind Tambe »
Daniel Kammen · Milind Tambe · Giulio De Leo · Mayur Mudigonda · Surya Karthik Mukkavilli -
2020 : Q/A and Discussion »
Surya Karthik Mukkavilli · Mayur Mudigonda · Milind Tambe -
2020 : Milind Tambe »
Milind Tambe -
2020 Poster: Automatically Learning Compact Quality-aware Surrogates for Optimization Problems »
Kai Wang · Bryan Wilder · Andrew Perrault · Milind Tambe -
2020 Spotlight: Automatically Learning Compact Quality-aware Surrogates for Optimization Problems »
Kai Wang · Bryan Wilder · Andrew Perrault · Milind Tambe -
2020 Poster: Collapsing Bandits and Their Application to Public Health Intervention »
Aditya Mate · Jackson Killian · Haifeng Xu · Andrew Perrault · Milind Tambe