Skip to yearly menu bar Skip to main content


Poster

Realizable H-Consistent and Bayes-Consistent Loss Functions for Learning to Defer

Anqi Mao · Mehryar Mohri · Yutao Zhong

West Ballroom A-D #5607
[ ]
Wed 11 Dec 4:30 p.m. PST — 7:30 p.m. PST

Abstract: We present a comprehensive study of surrogate loss functions for learning to defer. We introduce a broad family of surrogate losses, parameterized by a non-increasing function Ψ, and establish their realizable H-consistency under mild conditions. For cost functions based on classification error, we further show that these losses admit H-consistency bounds when the hypothesis set is symmetric and complete, a property satisfied by common neural network and linear function hypothesis sets. Our results also resolve an open question raised in previous work [Mozannar et al., 2023] by proving the realizable H-consistency and Bayes-consistency of a specific surrogate loss. Furthermore, we identify choices of Ψ that lead to H-consistent surrogate losses for *any general cost function*, thus achieving Bayes-consistency, realizable H-consistency, and H-consistency bounds *simultaneously*. We also investigate the relationship between H-consistency bounds and realizable H-consistency in learning to defer, highlighting key differences from standard classification. Finally, we empirically evaluate our proposed surrogate losses and compare them with existing baselines.

Chat is not available.