In this paper we study the support recovery problem for single index models , where is an unknown link function, and is an -sparse unit vector such that . In particular, we look into the performance of two computationally inexpensive algorithms: (a) the diagonal thresholding sliced inverse regression (DT-SIR) introduced by Lin et al. (2015); and (b) a semi-definite programming (SDP) approach inspired by Amini & Wainwright (2008). When for some , we demonstrate that both procedures can succeed in recovering the support of as long as the rescaled sample size is larger than a certain critical threshold. On the other hand, when is smaller than a critical value, any algorithm fails to recover the support with probability at least asymptotically. In other words, we demonstrate that both DT-SIR and the SDP approach are optimal (up to a scalar) for recovering the support of in terms of sample size. We provide extensive simulations, as well as a real dataset application to help verify our theoretical observations.
View on arXiv