529
v1v2v3 (latest)

NdLinear: Preserving Multi-Dimensional Structure for Parameter-Efficient Neural Networks

Main:13 Pages
7 Figures
Bibliography:6 Pages
19 Tables
Appendix:12 Pages
Abstract

In deep learning, processing multidimensional inputs (e.g., images, medical scans, and time series) is an important task that often requires flattening the inputs. We introduce NdLinear\mathit{NdLinear}, a drop-in replacement for linear layers that operates directly on tensors, requiring no flattening. By applying transformations separately along each dimension, NdLinear preserves native data structure while achieving dramatic parameter reductions, often by orders of magnitude, with minimal memory overhead. We prove NdLinear maintains expressivity through structured Tucker decomposition while preserving VC-dimension scaling. Extensive experiments demonstrate NdLinear's capacity to achieve significant parameter reductions with substantial wall-clock efficiency gains and minimal memory overhead. For instance, our NdLinearLoRA\mathit{NdLinear-LoRA} matches or exceeds standard LoRA on language reasoning tasks using up to 9×9\times fewer parameters. Experiments across CNNs, RNNs, Transformers, and MLPs on vision, language, time-series, and tabular tasks consistently demonstrate NdLinear's efficiency gains. While excelling at axis-separable tasks, NdLinear has limitations with entangled spatial interactions. By processing data in its original N-dimensional form, NdLinear provides a theoretically grounded, practical component for building more efficient neural architectures.

View on arXiv
Comments on this paper