42
0

Comparison of Different Deep Neural Network Models in the Cultural Heritage Domain

Abstract

The integration of computer vision and deep learning is an essential part of documenting and preserving cultural heritage, as well as improving visitor experiences. In recent years, two deep learning paradigms have been established in the field of computer vision: convolutional neural networks and transformer architectures. The present study aims to make a comparative analysis of some representatives of these two techniques of their ability to transfer knowledge from generic dataset, such as ImageNet, to cultural heritage specific tasks. The results of testing examples of the architectures VGG, ResNet, DenseNet, Visual Transformer, Swin Transformer, and PoolFormer, showed that DenseNet is the best in terms of efficiency-computability ratio.

View on arXiv
@article{boyadzhiev2025_2504.21387,
  title={ Comparison of Different Deep Neural Network Models in the Cultural Heritage Domain },
  author={ Teodor Boyadzhiev and Gabriele Lagani and Luca Ciampi and Giuseppe Amato and Krassimira Ivanova },
  journal={arXiv preprint arXiv:2504.21387},
  year={ 2025 }
}
Comments on this paper