Timezone: »
Descent directions such as movement towards Frank-Wolfe vertices, away steps, in-face away steps and pairwise directions have been an important design consideration in conditional gradient descent (CGD) variants. In this work, we attempt to demystify the impact of movement in these directions towards attaining constrained minimizers. The best local direction of descent is the directional derivative of the projection of the gradient, which we refer to as the "shadow" of the gradient. We show that the continuous-time dynamics of moving in the shadow are equivalent to those of PGD however non-trivial to discretize. By projecting gradients in PGD, one not only ensures feasibility but also is able to "wrap" around the convex region. We show that Frank-Wolfe (FW) vertices in fact recover the maximal wrap one can obtain by projecting gradients, thus providing a new perspective to these steps. We also claim that the shadow steps give the best direction of descent emanating from the convex hull of all possible away-vertices. Opening up the PGD movements in terms of shadow steps gives linear convergence, dependent on the number of faces. We combine these insights into a novel Shadow-CG method that uses FW steps (i.e., wrap around the polytope) and shadow steps (i.e., optimal local descent direction), while enjoying linear convergence. Our analysis develops properties of directional derivatives of projections (which may be of independent interest), while providing a unifying view of various descent directions in the CGD literature.
Author Information
Hassan Mortagy (Georgia Institute of Technology)
Swati Gupta (Georgia Institute of Technology)
Sebastian Pokutta (Zuse Institute Berlin)
More from the Same Authors
-
2022 : Accelerated Riemannian Optimization: Handling Constraints to Bound Geometric Penalties »
David Martinez-Rubio · Sebastian Pokutta -
2022 Poster: Fast Algorithms for Packing Proportional Fairness and its Dual »
Francisco Criado · David Martinez-Rubio · Sebastian Pokutta -
2021 : Machine Learning for Combinatorial Optimization + Q&A »
Maxime Gasse · Simon Bowly · Chris Cameron · Quentin Cappart · Jonas Charfreitag · Laurent Charlin · Shipra Agrawal · Didier Chetelat · Justin Dumouchelle · Ambros Gleixner · Aleksandr Kazachkov · Elias Khalil · Pawel Lichocki · Andrea Lodi · Miles Lubin · Christopher Morris · Dimitri Papageorgiou · Augustin Parjadis · Sebastian Pokutta · Antoine Prouvost · Yuandong Tian · Lara Scavuzzo · Giulia Zarpellon -
2021 Poster: Learning to Schedule Heuristics in Branch and Bound »
Antonia Chmiela · Elias Khalil · Ambros Gleixner · Andrea Lodi · Sebastian Pokutta -
2021 Poster: Simple steps are all you need: Frank-Wolfe and generalized self-concordant functions »
Alejandro Carderera · Mathieu Besançon · Sebastian Pokutta -
2021 Poster: Reusing Combinatorial Structure: Faster Iterative Projections over Submodular Base Polytopes »
Jai Moondra · Hassan Mortagy · Swati Gupta -
2020 Poster: Group-Fair Online Allocation in Continuous Time »
Semih Cayci · Swati Gupta · Atilla Eryilmaz -
2018 Poster: Limited Memory Kelley's Method Converges for Composite Convex and Submodular Objectives »
Song Zhou · Swati Gupta · Madeleine Udell -
2018 Spotlight: Limited Memory Kelley's Method Converges for Composite Convex and Submodular Objectives »
Song Zhou · Swati Gupta · Madeleine Udell