ResearchTrend.AI
  • Papers
  • Communities
  • Events
  • Blog
  • Pricing
Papers
Communities
Social Events
Terms and Conditions
Pricing
Parameter LabParameter LabTwitterGitHubLinkedInBlueskyYoutube

© 2025 ResearchTrend.AI, All rights reserved.

  1. Home
  2. Papers
  3. 1805.01870
8
4

Hedging parameter selection for basis pursuit

4 May 2018
Stéphane Chrétien
A. Gibberd
S. Roy
ArXiv (abs)PDFHTML
Abstract

In Compressed Sensing and high dimensional estimation, signal recovery often relies on sparsity assumptions and estimation is performed via ℓ1\ell_1ℓ1​-penalized least-squares optimization, a.k.a. LASSO. The ℓ1\ell_1ℓ1​ penalisation is usually controlled by a weight, also called "relaxation parameter", denoted by λ\lambdaλ. It is commonly thought that the practical efficiency of the LASSO for prediction crucially relies on accurate selection of λ\lambdaλ. In this short note, we propose to consider the hyper-parameter selection problem from a new perspective which combines the Hedge online learning method by Freund and Shapire, with the stochastic Frank-Wolfe method for the LASSO. Using the Hedge algorithm, we show that a our simple selection rule can achieve prediction results comparable to Cross Validation at a potentially much lower computational cost.

View on arXiv
Comments on this paper