ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2106.02933
28
17

k-Mixup Regularization for Deep Learning via Optimal Transport

5 June 2021
Kristjan Greenewald
Anming Gu
Mikhail Yurochkin
Justin Solomon
Edward Chien
ArXivPDFHTML
Abstract

Mixup is a popular regularization technique for training deep neural networks that improves generalization and increases robustness to certain distribution shifts. It perturbs input training data in the direction of other randomly-chosen instances in the training set. To better leverage the structure of the data, we extend mixup in a simple, broadly applicable way to \emph{kkk-mixup}, which perturbs kkk-batches of training points in the direction of other kkk-batches. The perturbation is done with displacement interpolation, i.e. interpolation under the Wasserstein metric. We demonstrate theoretically and in simulations that kkk-mixup preserves cluster and manifold structures, and we extend theory studying the efficacy of standard mixup to the kkk-mixup case. Our empirical results show that training with kkk-mixup further improves generalization and robustness across several network architectures and benchmark datasets of differing modalities. For the wide variety of real datasets considered, the performance gains of kkk-mixup over standard mixup are similar to or larger than the gains of mixup itself over standard ERM after hyperparameter optimization. In several instances, in fact, kkk-mixup achieves gains in settings where standard mixup has negligible to zero improvement over ERM.

View on arXiv
Comments on this paper