Self-Supervised Learning with an Information Maximization Criterion

Published in NeurIPS, 2022

Recommended citation: Ozsoy, S., Hamdan, S.S., Arik, S.Ö., Yuret, D., & Erdogan, A.T. (2022). Self-Supervised Learning with an Information Maximization Criterion. ArXiv, abs/2209.07999. https://arxiv.org/abs/2209.07999

This article proposes a self-supervised learning method that uses a second-order statistics-based mutual information measure that reflects the level of correlation among its arguments and prevents dimensional collapse by encouraging the spread of information across the whole feature space.

For this project, I worked heavily on the technical side of the project, focusing on the implementation, optimizing it, and parallelizing it as well as implementing new ideas and carrying out experiments to test them. I directly contributed to the following:

  • Implementing the CorInfoMax self-supervised approach in PyTorch.
  • Implementing multi-GPU support for the CorInfoMax loss function and ensuring the exact same performance as in the single GPU case.
  • Identified and addressed bottlenecks in dataloading that were heavily affecting performance.
  • Identified other improvements in the model implementation and design that led to a two-fold speedup with minimal to no impact in performance.

Talk

Access

Access the code and paper here

  • Ozsoy, S., Hamdan, S.S., Arik, S.Ö., Yuret, D., & Erdogan, A.T. (2022). Self-Supervised Learning with an Information Maximization Criterion. ArXiv, abs/2209.07999.