2008
DOI: 10.3103/s1066530708040030
|View full text |Cite
|
Sign up to set email alerts
|

Some theoretical results on the Grouped Variables Lasso

Abstract: We consider the linear regression model with Gaussian error. We estimate the unknown parameters by a procedure inspired from the Group Lasso estimator introduced in [21]. We show that this estimator satisfies a sparsity oracle inequality, i.e., a bound in terms of the number of non-zero components of the oracle vector. We prove that this bound is better, in some cases, than the one achieved by the Lasso and the Dantzig selector.

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
4
1

Citation Types

0
25
0

Year Published

2009
2009
2023
2023

Publication Types

Select...
8
1

Relationship

0
9

Authors

Journals

citations
Cited by 33 publications
(25 citation statements)
references
References 22 publications
0
25
0
Order By: Relevance
“…Our techniques of proofs build upon and extend those in these papers. Several papers analyzing statistical properties of the Group Lasso estimator appeared quite recently [3,10,16,19,24,25,26,31]. Most of them are focused on the Group Lasso for additive models [16,19,25,31] or generalized linear models [24].…”
Section: Previous Workmentioning
confidence: 99%
“…Our techniques of proofs build upon and extend those in these papers. Several papers analyzing statistical properties of the Group Lasso estimator appeared quite recently [3,10,16,19,24,25,26,31]. Most of them are focused on the Group Lasso for additive models [16,19,25,31] or generalized linear models [24].…”
Section: Previous Workmentioning
confidence: 99%
“…Other methods have been introduced that have concave and continuous differentiable penalties, such as a smoothly clipped absolute deviation (SCAD) penalty defined as pitalicλfalse(false|italicβfalse|false)=normalΣj=1pitalicλfalse[Ifalse(false|italicβjfalse|italicλfalse)+Ifalse(false|italicβjfalse|>italicλfalse)(aλ|βjfalse|false)+false/{false(a1false)λ}false]0.166667emfor some a0.166667em>2 for some a > 2 (Fan & Li ) and a minimax concave penalty (MCP) with pitalicλfalse(false|italicβfalse|false)=normalΣj=1pfalse(aitalicλfalse|italicβj|)+/a (Zhang ). Most of these methods are consistent when the sample size is much larger than the number of parameters (Zhao & Yu ; Meinshausen & Bühlmann ; Bunea, Tsybakov & Wegkamp ; Huang, Ma & Zhang ; Chesneau & Hebiri ; Wainwright ).…”
Section: Introductionmentioning
confidence: 99%
“…Other methods have been introduced that have concave and continuous differentiable penalties, such as a smoothly clipped absolute deviation (SCAD) penalty defined as p (| |) = p j=1 [I (| j | ) + I (| j | > )(a − | j |) + ={(a − 1) }] for some a > 2 for some a > 2 (Fan & Li 2001) and a minimax concave penalty (MCP) with p (| |) = p j=1 (a − | j |) + =a (Zhang 2010). Most of these methods are consistent when the sample size is much larger than the number of parameters (Zhao & Yu 2006;Meinshausen & Bühlmann 2006;Bunea, Tsybakov & Wegkamp 2007;Huang, Ma & Zhang 2008;Chesneau & Hebiri 2008;Wainwright 2009).…”
Section: Introductionmentioning
confidence: 99%
“…Those results lead to the refinements of their previous results for multi-task learning (see [17]). The behavior of the Lasso and Group Lasso regarding their selection and estimation properties have been studied in : [16,24,40,29,39,25] for Lasso in linear regression; [9,26] for Group Lasso in linear regression; [31,22,13] for additive models. Few results on the Lasso and Group Lasso concern logistic regression model.…”
Section: Introductionmentioning
confidence: 99%