212
v1v2v3v4 (latest)

Unifying Multiple Foundation Models for Advanced Computational Pathology

Ruiying Li
Zhengqun Jiang
Fang Yan
Xiaofan Zhang
Shaoting Zhang
Main:47 Pages
5 Figures
Bibliography:3 Pages
3 Tables
Abstract

Foundation models have substantially advanced computational pathology by learning transferable visual representations from large histological datasets, yet their performance varies widely across tasks due to differences in training data composition and reliance on proprietary datasets that cannot be cumulatively expanded. Existing efforts to combine foundation models through offline distillation partially mitigate this issue but require dedicated distillation data and repeated retraining to integrate new models. Here we present Shazam, an online integration model that adaptively combines multiple pretrained pathology foundation models within a unified and scalable representation learning paradigm. Our findings show that fusing multi-level features through adaptive expert weighting and online distillation enables efficient consolidation of complementary model strengths without additional pretraining. Across spatial transcriptomics prediction, survival prognosis, tile-level classification, and visual question answering, Shazam consistently outperforms strong individual models, demonstrating that online model integration provides a practical and extensible strategy for advancing computational pathology.

View on arXiv
Comments on this paper