2014
DOI: 10.1109/tnnls.2013.2286696
|View full text |Cite
|
Sign up to set email alerts
|

Nonconvex Regularizations for Feature Selection in Ranking With Sparse SVM

Abstract: Abstract-Feature selection in learning to rank has recently emerged as a crucial issue. Whereas several preprocessing approaches have been proposed, only a few works have been focused on integrating the feature selection into the learning process. In this work, we propose a general framework for feature selection in learning to rank using SVM with a sparse regularization term. We investigate both classical convex regularizations such as 1 or weighted 1 and non-convex regularization terms such as log penalty, M… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
67
0

Year Published

2015
2015
2024
2024

Publication Types

Select...
4
3
2

Relationship

1
8

Authors

Journals

citations
Cited by 94 publications
(68 citation statements)
references
References 34 publications
1
67
0
Order By: Relevance
“…Zhang et al [2] established a unified theory for SCAD-and MCP-penalized SVM in the high-dimensional setting. Laporte et al [16] proposed a general framework for feature selection in learning to rank using SVM with nonconvex regularizations such as log penalty, MCP and ℓ p norm with 0 < p < 1. Recently, Zhang et al [17] have established an unified theory for a general class of nonconvex penalized SVMs in the high-dimensional setting.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Zhang et al [2] established a unified theory for SCAD-and MCP-penalized SVM in the high-dimensional setting. Laporte et al [16] proposed a general framework for feature selection in learning to rank using SVM with nonconvex regularizations such as log penalty, MCP and ℓ p norm with 0 < p < 1. Recently, Zhang et al [17] have established an unified theory for a general class of nonconvex penalized SVMs in the high-dimensional setting.…”
Section: Related Workmentioning
confidence: 99%
“…• the successive quadratic algorithm for the SCADpenalized hinge loss function (SCAD SVM 2 ) [11]. • the reweighted ℓ 1 scheme for the MCP-penalized squared hinge loss function (RankSVM-MCP 3 ) [16]. • the generative shrinkage and thresholding (GIST 4 ) algorithm [12].…”
Section: Experimental Evaluation a Experimental Setupmentioning
confidence: 99%
“…11] for p with p = 1/2). Note that efficient methods which estimate the Hessian matrix [61], [62] exist, as well as a wide range of methods based on DC programming, which have shown to work very well in practice [62], [63] and can handle the general case p ∈ (0, 1] for the p pseudonorm (see [64] for an implementation).…”
Section: E Optimization Algorithmsmentioning
confidence: 99%
“…Existing feature selection algorithm can be categorized as supervised feature selection (on data with full class labels) [5]- [9], unsupervised feature selection (on data without class labels) [10]- [15], and semisupervised feature selection (on data with partial labels) [14], [16], [17]. Feature selection in unsupervised context is considered to be more difficult than the other two cases, since there is no target information available for training.…”
Section: Introductionmentioning
confidence: 99%