ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2201.12680
29
26

Understanding Deep Contrastive Learning via Coordinate-wise Optimization

29 January 2022
Yuandong Tian
ArXivPDFHTML
Abstract

We show that Contrastive Learning (CL) under a broad family of loss functions (including InfoNCE) has a unified formulation of coordinate-wise optimization on the network parameter θ\boldsymbol{\theta}θ and pairwise importance α\alphaα, where the \emph{max player} θ\boldsymbol{\theta}θ learns representation for contrastiveness, and the \emph{min player} α\alphaα puts more weights on pairs of distinct samples that share similar representations. The resulting formulation, called α\alphaα-CL, unifies not only various existing contrastive losses, which differ by how sample-pair importance α\alphaα is constructed, but also is able to extrapolate to give novel contrastive losses beyond popular ones, opening a new avenue of contrastive loss design. These novel losses yield comparable (or better) performance on CIFAR10, STL-10 and CIFAR-100 than classic InfoNCE. Furthermore, we also analyze the max player in detail: we prove that with fixed α\alphaα, max player is equivalent to Principal Component Analysis (PCA) for deep linear network, and almost all local minima are global and rank-1, recovering optimal PCA solutions. Finally, we extend our analysis on max player to 2-layer ReLU networks, showing that its fixed points can have higher ranks.

View on arXiv
Comments on this paper