Timezone: »
During training, models can exploit spurious correlations as shortcuts, resulting in poor generalization performance when shortcuts do not persist. In this work, assuming access to a representation based on domain knowledge (i.e., known concepts) that is invariant to shortcuts, we aim to learn robust and accurate models from biased training data. In contrast to previous work, we do not rely solely on known concepts, but allow the model to also learn unknown concepts. We propose two approaches for mitigating shortcuts that incorporate domain knowledge, while accounting for potentially important yet unknown concepts. The first approach is two-staged. After fitting a model using known concepts, it accounts for the residual using unknown concepts. While flexible, we show that this approach is vulnerable when shortcuts are correlated with the unknown concepts. This limitation is addressed by our second approach that extends a recently proposed regularization penalty. Applied to two real-world datasets, we demonstrate that both approaches can successfully mitigate shortcut learning.
Author Information
Jiaxuan Wang (Meta)
Sarah Jabbour (University of Michigan - Ann Arbor)
Maggie Makar (University of Michigan)
Michael Sjoding (University of Michigan - Ann Arbor)
Jenna Wiens (University of Michigan)
More from the Same Authors
-
2021 : Sequential Decision Making with Limited Resources »
Hallee Wong · Maggie Makar · Aniruddh Raghu · John Guttag -
2022 : Conditional differential measurement error: partial identifiability and estimation »
Pengrun Huang · Maggie Makar -
2023 Poster: Counterfactual-Augmented Importance Sampling for Semi-Offline Policy Evaluation »
Shengpu Tang · Jenna Wiens -
2022 Poster: Leveraging Factored Action Spaces for Efficient Offline Reinforcement Learning in Healthcare »
Shengpu Tang · Maggie Makar · Michael Sjoding · Finale Doshi-Velez · Jenna Wiens -
2022 Poster: Causally motivated multi-shortcut identification and removal »
Jiayun Zheng · Maggie Makar