29
3

iiANET: Inception Inspired Attention Hybrid Network for efficient Long-Range Dependency

Abstract

The recent emergence of hybrid models has introduced a transformative approach to computer vision, gradually moving beyond conventional convolutional neural net-works and vision transformers. However, efficiently combining these two paradigms to better capture long-range dependencies in complex images remains a challenge. In this paper, we present iiANET (Inception Inspired Attention Network), an efficient hybrid visual backbone designed to improve the modeling of long-range dependen-cies. The core innovation of iiANET is the iiABlock, a unified building block that in-tegrates global r-MHSA (Multi-Head Self-Attention) and convolutional layers in paral-lel. This design enables iiABlock to simultaneously capture global context and local details, making it highly effective for extracting rich and diverse features. By effi-ciently fusing these complementary representations, iiABlock allows iiANET to achieve strong feature interaction while maintaining computational efficiency. Exten-sive qualitative and quantitative evaluations across various benchmarks show im-proved performance over several state-of-the-art models.

View on arXiv
@article{yunusa2025_2407.07603,
  title={ iiANET: Inception Inspired Attention Hybrid Network for efficient Long-Range Dependency },
  author={ Haruna Yunusa and Qin Shiyin and Abdulrahman Hamman Adama Chukkol and Adamu Lawan and Abdulganiyu Abdu Yusuf and Isah Bello },
  journal={arXiv preprint arXiv:2407.07603},
  year={ 2025 }
}
Comments on this paper