Skip to yearly menu bar Skip to main content


Poster

An Embedding Framework for Consistent Polyhedral Surrogates

Jessica Finocchiaro · Rafael Frongillo · Bo Waggoner

East Exhibition Hall B + C #221

Keywords: [ Optimization ] [ Convex Optimization ] [ Theory ] [ Learning Theory ]


Abstract:

We formalize and study the natural approach of designing convex surrogate loss functions via embeddings for problems such as classification or ranking. In this approach, one embeds each of the finitely many predictions (e.g. classes) as a point in \reals^d, assigns the original loss values to these points, and convexifies the loss in some way to obtain a surrogate. We prove that this approach is equivalent, in a strong sense, to working with polyhedral (piecewise linear convex) losses. Moreover, given any polyhedral loss L, we give a construction of a link function through which L is a consistent surrogate for the loss it embeds. We go on to illustrate the power of this embedding framework with succinct proofs of consistency or inconsistency of various polyhedral surrogates in the literature.

Live content is unavailable. Log in and register to view live content