Timezone: »
Self-supervised learning allows AI systems to learn effective representations from large amounts of data using tasks that do not require costly labeling. Mode collapse, i.e., the model producing identical representations for all inputs, is a central problem to many self-supervised learning approaches, making self-supervised tasks, such as matching distorted variants of the inputs, ineffective. In this article, we argue that a straightforward application of information maximization among alternative latent representations of the same input naturally solves the collapse problem and achieves competitive empirical results. We propose a self-supervised learning method, CorInfoMax, that uses a second-order statistics-based mutual information measure that reflects the level of correlation among its arguments. Maximizing this correlative information measure between alternative representations of the same input serves two purposes: (1) it avoids the collapse problem by generating feature vectors with non-degenerate covariances; (2) it establishes relevance among alternative representations by increasing the linear dependence among them. An approximation of the proposed information maximization objective simplifies to a Euclidean distance-based objective function regularized by the log-determinant of the feature covariance matrix. The regularization term acts as a natural barrier against feature space degeneracy. Consequently, beyond avoiding complete output collapse to a single point, the proposed approach also prevents dimensional collapse by encouraging the spread of information across the whole feature space. Numerical experiments demonstrate that CorInfoMax achieves better or competitive performance results relative to the state-of-the-art SSL approaches.
Author Information
Serdar Ozsoy (Koç University)
Shadi Hamdan (Koç University)
Sercan Arik (Google)
Deniz Yuret (Koç University, İstanbul)
Alper Erdogan (Koç University)

Alper T. Erdogan (Senior Member, IEEE) was born in Ankara, Turkey, in 1971. He received the B.S. degree from the Middle East Technical University, Ankara, Turkey, in 1993, and the M.S. and Ph.D. degrees from Stanford University, Stanford, CA, USA, in 1995 and 1999, respectively. He was a Principal Research Engineer with the Globespan-Virata Corporation (formerly Excess Bandwidth and Virata Corporations) from September 1999 to November 2001. He joined the Electrical and Electronics Engineering Department, Koc University, Istanbul, Turkey, in January 2002, where he is currently a Professor. His research interests include adaptive signal processing, machine learning, physical layer communications, computational neuroscience, optimization, system theory and control, and information theory. Dr. Erdogan was the recipient of several awards including TUBITAK Career Award (2005), Werner Von Siemens Excellence Award (2007), TUBA GEBIP Outstanding Young Scientist Award (2008), TUBITAK Encouragement Award (2010), and Outstanding Teaching Award (2017). He was an Associate Editor for the IEEE Transactions on Signal Processing, and he was a member of IEEE Signal Processing Theory and Methods Technical Committee.
More from the Same Authors
-
2022 : Provable Re-Identification Privacy »
Zachary Izzo · Jinsung Yoon · Sercan Arik · James Zou -
2022 Poster: Biologically-Plausible Determinant Maximization Neural Networks for Blind Separation of Correlated Sources »
Bariscan Bozkurt · Cengiz Pehlevan · Alper Erdogan -
2021 Poster: Controlling Neural Networks with Rule Representations »
Sungyong Seo · Sercan Arik · Jinsung Yoon · Xiang Zhang · Kihyuk Sohn · Tomas Pfister -
2020 Poster: On Completeness-aware Concept-Based Explanations in Deep Neural Networks »
Chih-Kuan Yeh · Been Kim · Sercan Arik · Chun-Liang Li · Tomas Pfister · Pradeep Ravikumar -
2020 : Interpretable Sequence Learning for Covid-19 Forecasting »
Sercan Arik -
2020 Poster: Interpretable Sequence Learning for Covid-19 Forecasting »
Sercan Arik · Chun-Liang Li · Jinsung Yoon · Rajarishi Sinha · Arkady Epshteyn · Long Le · Vikas Menon · Shashank Singh · Leyou Zhang · Martin Nikoltchev · Yash Sonthalia · Hootan Nakhost · Elli Kanal · Tomas Pfister -
2020 Spotlight: Interpretable Sequence Learning for Covid-19 Forecasting »
Sercan Arik · Chun-Liang Li · Jinsung Yoon · Rajarishi Sinha · Arkady Epshteyn · Long Le · Vikas Menon · Shashank Singh · Leyou Zhang · Martin Nikoltchev · Yash Sonthalia · Hootan Nakhost · Elli Kanal · Tomas Pfister