Benchmarking Ultra-Low-Power NPUs
- BDL
Main:11 Pages
7 Figures
Bibliography:3 Pages
6 Tables
Appendix:1 Pages
Abstract
Efficient on-device neural network (NN) inference has various advantages over cloud-based processing, including predictable latency, enhanced privacy, greater reliability, and reduced operating costs for vendors. This has sparked the recent rapid development of microcontroller-scale NN accelerators, often referred to as neural processing units (NPUs), designed specifically for ultra-low-power applications.
View on arXivComments on this paper
