26
0

ATP: Adaptive Threshold Pruning for Efficient Data Encoding in Quantum Neural Networks

Abstract

Quantum Neural Networks (QNNs) offer promising capabilities for complex data tasks, but are often constrained by limited qubit resources and high entanglement, which can hinder scalability and efficiency. In this paper, we introduce Adaptive Threshold Pruning (ATP), an encoding method that reduces entanglement and optimizes data complexity for efficient computations in QNNs. ATP dynamically prunes non-essential features in the data based on adaptive thresholds, effectively reducing quantum circuit requirements while preserving high performance. Extensive experiments across multiple datasets demonstrate that ATP reduces entanglement entropy and improves adversarial robustness when combined with adversarial training methods like FGSM. Our results highlight ATPs ability to balance computational efficiency and model resilience, achieving significant performance improvements with fewer resources, which will help make QNNs more feasible in practical, resource-constrained settings.

View on arXiv
@article{afane2025_2503.21815,
  title={ ATP: Adaptive Threshold Pruning for Efficient Data Encoding in Quantum Neural Networks },
  author={ Mohamed Afane and Gabrielle Ebbrecht and Ying Wang and Juntao Chen and Junaid Farooq },
  journal={arXiv preprint arXiv:2503.21815},
  year={ 2025 }
}
Comments on this paper