218

Benchmarking Ultra-Low-Power μμNPUs

Main:11 Pages
7 Figures
Bibliography:3 Pages
6 Tables
Appendix:1 Pages
Abstract

Efficient on-device neural network (NN) inference has various advantages over cloud-based processing, including predictable latency, enhanced privacy, greater reliability, and reduced operating costs for vendors. This has sparked the recent rapid development of microcontroller-scale NN accelerators, often referred to as neural processing units (μ\muNPUs), designed specifically for ultra-low-power applications.

View on arXiv
Comments on this paper