416
v1v2 (latest)

Beyond Low-rank Decomposition: A Shortcut Approach for Efficient On-Device Learning

Main:9 Pages
6 Figures
Bibliography:3 Pages
4 Tables
Appendix:3 Pages
Abstract

On-device learning has emerged as a promising direction for AI development, particularly because of its potential to reduce latency issues and mitigate privacy risks associated with device-server communication, while improving energy efficiency. Despite these advantages, significant memory and computational constraints still represent major challenges for its deployment. Drawing on previous studies on low-rank decomposition methods that address activation memory bottlenecks in backpropagation, we propose a novel shortcut approach as an alternative. Our analysis and experiments demonstrate that our method can reduce activation memory usage, even up to 120.09×120.09\times compared to vanilla training, while also reducing overall training FLOPs up to 1.86×1.86\times when evaluated on traditional benchmarks.

View on arXiv
Comments on this paper