Timezone: »
The ability to estimate epistemic uncertainty is often crucial when deploying machine learning in the real world, but modern methods often produce overconfident, uncalibrated uncertainty predictions. A common approach to quantify epistemic uncertainty, usable across a wide class of prediction models, is to train a model ensemble. In a naive implementation, the ensemble approach has high computational cost and high memory demand. This challenges in particular modern deep learning, where even a single deep network is already demanding in terms of compute and memory, and has given rise to a number of attempts to emulate the model ensemble without actually instantiating separate ensemble members. We introduce FiLM-Ensemble, a deep, implicit ensemble method based on the concept of Feature-wise Linear Modulation (FiLM). That technique was originally developed for multi-task learning, with the aim of decoupling different tasks. We show that the idea can be extended to uncertainty quantification: by modulating the network activations of a single deep network with FiLM, one obtains a model ensemble with high diversity, and consequently well-calibrated estimates of epistemic uncertainty, with low computational overhead in comparison. Empirically, FiLM-Ensemble outperforms other implicit ensemble methods, and it comes very close to the upper bound of an explicit ensemble of networks (sometimes even beating it), at a fraction of the memory cost.
Author Information
Mehmet Ozgur Turkoglu (ETH Zurich)
Alexander Becker (ETH Zurich)
Hüseyin Anil Gündüz (LMU Munich)
Mina Rezaei (Ludwig-Maximilian University)
Bernd Bischl (LMU)
Rodrigo Caye Daudt (ETH Zurich)
Stefano D'Aronco (Swiss Federal Institute of Technology)
Jan Wegner (University of Zurich)
Konrad Schindler (ETH Zürich)
More from the Same Authors
-
2021 : Survival-oriented embeddings for improving accessibility to complex data structures »
Tobias Weber · Bernd Bischl · David Ruegamer -
2022 Poster: tntorch: Tensor Network Learning with PyTorch »
Mikhail Usvyatsov · Rafael Ballester-Ripoll · Konrad Schindler -
2022 : Transformer Model for Genome Sequence Analysis »
Noah Hurmer · Xiao-Yin To · Martin Binder · Hüseyin Anil Gündüz · Philipp Münch · René Mreches · Alice McHardy · Bernd Bischl · Mina Rezaei -
2022 : What cleaves? Is proteasomal cleavage prediction reaching a ceiling? »
Ingo Ziegler · Bolei Ma · Ercong Nie · Bernd Bischl · David Rügamer · Benjamin Schubert · Emilio Dorigatti -
2022 Spotlight: tntorch: Tensor Network Learning with PyTorch »
Mikhail Usvyatsov · Rafael Ballester-Ripoll · Konrad Schindler -
2022 Spotlight: Lightning Talks 1A-1 »
Siba Smarak Panigrahi · Xuhong Li · Mikhail Usvyatsov · Shaohan Chen · Sohan Patnaik · Haoyi Xiong · Nikolaos V Sahinidis · Rafael Ballester-Ripoll · Chuanhou Gao · Xingjian Li · Konrad Schindler · Xuanyu Wu · Zeyu Chen · Dejing Dou