Feature selection is a critical step in data-driven applications, reducing input dimensionality to enhance learning accuracy, computational efficiency, and interpretability. Existing state-of-the-art methods often require post-selection retraining and extensive hyperparameter tuning, complicating their adoption. We introduce a novel, non-intrusive feature selection layer that, given a target feature count , automatically identifies and selects the most informative features during neural network training. Our method is uniquely simple, requiring no alterations to the loss function, network architecture, or post-selection retraining. The layer is mathematically elegant and can be fully described by: \begin{align} \nonumber \tilde{x}_i = a_i x_i + (1-a_i)z_i \end{align} where is the input feature, the output, a Gaussian noise, and trainable gain such that . This formulation induces an automatic clustering effect, driving of the gains to (selecting informative features) and the rest to (discarding redundant ones) via weighted noise distortion and gain normalization. Despite its extreme simplicity, our method delivers state-of-the-art performance on standard benchmark datasets and a novel real-world dataset, outperforming or matching existing approaches without requiring hyperparameter search for or retraining. Theoretical analysis in the context of linear regression further validates its efficacy. Our work demonstrates that simplicity and performance are not mutually exclusive, offering a powerful yet straightforward tool for feature selection in machine learning.
View on arXiv@article{pad2025_2505.03923, title={ SAND: One-Shot Feature Selection with Additive Noise Distortion }, author={ Pedram Pad and Hadi Hammoud and Mohamad Dia and Nadim Maamari and L. Andrea Dunbar }, journal={arXiv preprint arXiv:2505.03923}, year={ 2025 } }