36
0

Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features

Abstract

Recently, images that distort or fabricate facts using generative models have become a social concern. To cope with continuous evolution of generative artificial intelligence (AI) models, model attribution (MA) is necessary beyond just detection of synthetic images. However, current deep learning-based MA methods must be trained from scratch with new data to recognize unseen models, which is time-consuming and data-intensive. This work proposes a new strategy to deal with persistently emerging generative models. We adapt few-shot class-incremental learning (FSCIL) mechanisms for MA problem to uncover novel generative AI models. Unlike existing FSCIL approaches that focus on object classification using high-level information, MA requires analyzing low-level details like color and texture in synthetic images. Thus, we utilize a learnable representation from different levels of CLIP-ViT features. To learn an effective representation, we propose Adaptive Integration Module (AIM) to calculate a weighted sum of CLIP-ViT block features for each image, enhancing the ability to identify generative models. Extensive experiments show our method effectively extends from prior generative models to recent ones.

View on arXiv
@article{lee2025_2503.08148,
  title={ Few-Shot Class-Incremental Model Attribution Using Learnable Representation From CLIP-ViT Features },
  author={ Hanbyul Lee and Juneho Yi },
  journal={arXiv preprint arXiv:2503.08148},
  year={ 2025 }
}
Comments on this paper