Timezone: »

Efficient Output Kernel Learning for Multiple Tasks
Pratik Kumar Jawanpuria · Maksim Lapin · Matthias Hein · Bernt Schiele

Wed Dec 09 04:00 PM -- 08:59 PM (PST) @ 210 C #54

The paradigm of multi-task learning is that one can achieve better generalization by learning tasks jointly and thus exploiting the similarity between the tasks rather than learning them independently of each other. While previously the relationship between tasks had to be user-defined in the form of an output kernel, recent approaches jointly learn the tasks and the output kernel. As the output kernel is a positive semidefinite matrix, the resulting optimization problems are not scalable in the number of tasks as an eigendecomposition is required in each step. Using the theory of positive semidefinite kernels we show in this paper that for a certain class of regularizers on the output kernel, the constraint of being positive semidefinite can be dropped as it is automatically satisfied for the relaxed problem. This leads to an unconstrained dual problem which can be solved efficiently. Experiments on several multi-task and multi-class data sets illustrate the efficacy of our approach in terms of computational efficiency as well as generalization performance.

Author Information

Pratik Kumar Jawanpuria (Saarlanduniversity)
Maksim Lapin (Max Planck Institute for Informatics)
Matthias Hein (Saarland University)
Bernt Schiele (Max Planck Institute for Informatics)

More from the Same Authors