Timezone: »
Multi-Task Learning (MTL) is appealing for deep learning regularization. In this paper, we tackle a specific MTL context denoted as primary MTL, where the ultimate goal is to improve the performance of a given primary task by leveraging several other auxiliary tasks. Our main methodological contribution is to introduce ROCK, a new generic multi-modal fusion block for deep learning tailored to the primary MTL context. ROCK architecture is based on a residual connection, which makes forward prediction explicitly impacted by the intermediate auxiliary representations. The auxiliary predictor's architecture is also specifically designed to our primary MTL context, by incorporating intensive pooling operators for maximizing complementarity of intermediate representations. Extensive experiments on NYUv2 dataset (object detection with scene classification, depth prediction, and surface normal estimation as auxiliary tasks) validate the relevance of the approach and its superiority to flat MTL approaches. Our method outperforms state-of-the-art object detection models on NYUv2 dataset by a large margin, and is also able to handle large-scale heterogeneous inputs (real and synthetic images) with missing annotation modalities.
Author Information
Taylor Mordan (Sorbonne Université, LIP6)
Nicolas THOME (Cnam)
Gilles Henaff (Thales Optronique S.A.S.)
Matthieu Cord (Sorbonne University)
More from the Same Authors
-
2020 : Paper 16: Driving Behavior Explanation with Multi-level Fusion »
Matthieu Cord · Patrick Pérez -
2022 : A generic diffusion-based approach for 3D human pose prediction in the wild »
Saeed Saadatnejad · Ali Rasekh · Mohammadreza Mofayezi · Yasamin Medghalchi · Sara Rajabzadeh · Taylor Mordan · Alexandre Alahi -
2021 Poster: RED : Looking for Redundancies for Data-FreeStructured Compression of Deep Neural Networks »
Edouard YVINEC · Arnaud Dapogny · Matthieu Cord · Kevin Bailly -
2021 Poster: Robust and Decomposable Average Precision for Image Retrieval »
Elias Ramzi · Nicolas THOME · Clément Rambour · Nicolas Audebert · Xavier Bitot -
2021 Poster: Look at the Variance! Efficient Black-box Explanations with Sobol-based Sensitivity Analysis »
Thomas FEL · Remi Cadene · Mathieu Chalvidal · Matthieu Cord · David Vigouroux · Thomas Serre -
2021 Poster: TTT++: When Does Self-Supervised Test-Time Training Fail or Thrive? »
Yuejiang Liu · Parth Kothari · Bastien van Delft · Baptiste Bellot-Gurlet · Taylor Mordan · Alexandre Alahi -
2020 Poster: Probabilistic Time Series Forecasting with Shape and Temporal Diversity »
Vincent LE GUEN · Nicolas THOME -
2019 Poster: Shape and Time Distortion Loss for Training Deep Time Series Forecasting Models »
Vincent LE GUEN · Nicolas THOME -
2019 Poster: RUBi: Reducing Unimodal Biases for Visual Question Answering »
Remi Cadene · Corentin Dancette · Hedi Ben younes · Matthieu Cord · Devi Parikh -
2019 Poster: Zero-Shot Semantic Segmentation »
Maxime Bucher · Tuan-Hung VU · Matthieu Cord · Patrick Pérez -
2019 Poster: Addressing Failure Prediction by Learning Model Confidence »
Charles Corbière · Nicolas THOME · Avner Bar-Hen · Matthieu Cord · Patrick Pérez -
2019 Poster: Riemannian batch normalization for SPD neural networks »
Daniel Brooks · Olivier Schwander · Frederic Barbaresco · Jean-Yves Schneider · Matthieu Cord -
2013 Poster: Top-Down Regularization of Deep Belief Networks »
Hanlin Goh · Nicolas Thome · Matthieu Cord · Joo-Hwee Lim