34
0

Enhancing Layer Attention Efficiency through Pruning Redundant Retrievals

Abstract

Growing evidence suggests that layer attention mechanisms, which enhance interaction among layers in deep neural networks, have significantly advanced network architectures. However, existing layer attention methods suffer from redundancy, as attention weights learned by adjacent layers often become highly similar. This redundancy causes multiple layers to extract nearly identical features, reducing the model's representational capacity and increasing training time. To address this issue, we propose a novel approach to quantify redundancy by leveraging the Kullback-Leibler (KL) divergence between adjacent layers. Additionally, we introduce an Enhanced Beta Quantile Mapping (EBQM) method that accurately identifies and skips redundant layers, thereby maintaining model stability. Our proposed Efficient Layer Attention (ELA) architecture, improves both training efficiency and overall performance, achieving a 30% reduction in training time while enhancing performance in tasks such as image classification and object detection.

View on arXiv
@article{li2025_2503.06473,
  title={ Enhancing Layer Attention Efficiency through Pruning Redundant Retrievals },
  author={ Hanze Li and Xiande Huang },
  journal={arXiv preprint arXiv:2503.06473},
  year={ 2025 }
}
Comments on this paper