We analyze the convergence behaviour of a recently proposed algorithm for sparse learning called Dual Augmented Lagrangian (DAL). We theoretically analyze under some conditions that DAL converges super-linearly in a non-asymptotic and global sense. We experimentally confirm our analysis in a large scale -regularized logistic regression problem and compare the efficiency of DAL algorithm to existing algorithms.
View on arXiv