Skip to yearly menu bar Skip to main content


Poster

Class Distribution Shifts in Zero-Shot Learning: Learning Robust Representations

Yuli Slavutsky · Yuval Benjamini

East Exhibit Hall A-C #2204
[ ]
Fri 13 Dec 11 a.m. PST — 2 p.m. PST

Abstract:

Zero-shot learning methods typically assume that the new, unseen classes encountered during deployment come from the same distribution as the the classes in the training set. However, real-world scenarios often involve class distribution shifts (e.g., in age or gender for person identification), posing challenges for zero-shot classifiers that rely on learned representations from training classes. In this work, we propose and analyze a model that assumes that the attribute responsible for the shift is unknown in advance. We show that in this setting, standard training may lead to non-robust representations. To mitigate this, we develop an algorithm for learning robust representations in which (a) synthetic data environments are constructed via hierarchical sampling, and (b) environment balancing penalization, inspired by out-of-distribution problems, is applied. We show that our algorithm improves generalization to diverse class distributions in both simulations and experiments on real-world datasets.

Live content is unavailable. Log in and register to view live content