ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 2308.11511
11
1

Mode Combinability: Exploring Convex Combinations of Permutation Aligned Models

22 August 2023
Adrián Csiszárik
M. Kiss
Péter Korösi-Szabó
Márton Muntag
Gergely Papp
D. Varga
    MoMe
ArXivPDFHTML
Abstract

We explore element-wise convex combinations of two permutation-aligned neural network parameter vectors ΘA\Theta_AΘA​ and ΘB\Theta_BΘB​ of size ddd. We conduct extensive experiments by examining various distributions of such model combinations parametrized by elements of the hypercube [0,1]d[0,1]^{d}[0,1]d and its vicinity. Our findings reveal that broad regions of the hypercube form surfaces of low loss values, indicating that the notion of linear mode connectivity extends to a more general phenomenon which we call mode combinability. We also make several novel observations regarding linear mode connectivity and model re-basin. We demonstrate a transitivity property: two models re-based to a common third model are also linear mode connected, and a robustness property: even with significant perturbations of the neuron matchings the resulting combinations continue to form a working model. Moreover, we analyze the functional and weight similarity of model combinations and show that such combinations are non-vacuous in the sense that there are significant functional differences between the resulting models.

View on arXiv
Comments on this paper