Poster
in
Workshop: Workshop on Distribution Shifts: Connecting Methods and Applications
Tailored Overlap for Learning Under Distribution Shift
David Bruns-Smith · Alexander D'Amour · Avi Feller · Steve Yadlowsky
Abstract:
Distributional overlap is a critical determinant of learnability in domain adaptation. The standard theory quantifies overlap in terms of divergence, as this factors directly into variance and generalization bounds agnostic to the functional form of the - relationship. However, in many modern settings, we cannot afford this agnosticism; we often wish to transfer across distributions with disjoint support, where these standard divergence measures are infinite. In this note, we argue that tailored'' divergences that are restricted to measuring overlap in a particular function class are more appropriate. We show how (and other) divergences can be generalized to this restricted function class setting via a variational representation, and use this to motivate balancing weight-based methods that have been proposed before, but, we believe, should be more widely used.
Chat is not available.