5
0

Structure-aware Semantic Discrepancy and Consistency for 3D Medical Image Self-supervised Learning

Tan Pan
Zhaorui Tan
Kaiyu Guo
Dongli Xu
Weidi Xu
Chen Jiang
Xin Guo
Yuan Qi
Yuan Cheng
Main:8 Pages
7 Figures
Bibliography:3 Pages
8 Tables
Abstract

3D medical image self-supervised learning (mSSL) holds great promise for medical analysis. Effectively supporting broader applications requires considering anatomical structure variations in location, scale, and morphology, which are crucial for capturing meaningful distinctions. However, previous mSSL methods partition images with fixed-size patches, often ignoring the structure variations. In this work, we introduce a novel perspective on 3D medical images with the goal of learning structure-aware representations. We assume that patches within the same structure share the same semantics (semantic consistency) while those from different structures exhibit distinct semantics (semantic discrepancy). Based on this assumption, we propose an mSSL framework named S2DCS^2DC, achieving Structure-aware Semantic Discrepancy and Consistency in two steps. First, S2DCS^2DC enforces distinct representations for different patches to increase semantic discrepancy by leveraging an optimal transport strategy. Second, S2DCS^2DC advances semantic consistency at the structural level based on neighborhood similarity distribution. By bridging patch-level and structure-level representations, S2DCS^2DC achieves structure-aware representations. Thoroughly evaluated across 10 datasets, 4 tasks, and 3 modalities, our proposed method consistently outperforms the state-of-the-art methods in mSSL.

View on arXiv
@article{pan2025_2507.02581,
  title={ Structure-aware Semantic Discrepancy and Consistency for 3D Medical Image Self-supervised Learning },
  author={ Tan Pan and Zhaorui Tan and Kaiyu Guo and Dongli Xu and Weidi Xu and Chen Jiang and Xin Guo and Yuan Qi and Yuan Cheng },
  journal={arXiv preprint arXiv:2507.02581},
  year={ 2025 }
}
Comments on this paper