2021
DOI: 10.1007/s10994-021-06085-5
|View full text |Cite
|
Sign up to set email alerts
|

Sparse classification: a scalable discrete optimization perspective

Abstract: We formulate the sparse classification problem of n samples with p features as a binary convex optimization problem and propose a outer-approximation algorithm to solve it exactly. For sparse logistic regression and sparse SVM, our algorithm finds optimal solutions for n and p in the 10,000 s within minutes. On synthetic data our algorithm achieves perfect support recovery in the large sample regime. Namely, there exists an n 0 such that the algorithm takes a long time to find an optimal solution and does not … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
39
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
2
1

Relationship

1
5

Authors

Journals

citations
Cited by 26 publications
(39 citation statements)
references
References 65 publications
0
39
0
Order By: Relevance
“…Therefore, new research in numerical algorithms for solving the exact formulation (2) directly has flourished. Leveraging recent advances in mixed-integer solvers [6,4], Lagrangian relaxation [39] or cutting-plane methods [5,7], these works have demonstrated significant improvement over existing Lasso-based heuristics. To the best of our knowledge, the exact algorithm proposed by Bertsimas and Van Parys [5], Bertsimas et al [7] is the most scalable method providing provably optimal solutions to the optimization problem (2), at the expense of potentially significant computational time and the use of a commercial integer optimization solver.…”
Section: Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…Therefore, new research in numerical algorithms for solving the exact formulation (2) directly has flourished. Leveraging recent advances in mixed-integer solvers [6,4], Lagrangian relaxation [39] or cutting-plane methods [5,7], these works have demonstrated significant improvement over existing Lasso-based heuristics. To the best of our knowledge, the exact algorithm proposed by Bertsimas and Van Parys [5], Bertsimas et al [7] is the most scalable method providing provably optimal solutions to the optimization problem (2), at the expense of potentially significant computational time and the use of a commercial integer optimization solver.…”
Section: Methodsmentioning
confidence: 99%
“…Thanks to both hardware improvement and advances in mixed-integer optimization solvers, Bertsimas et al [6], Bertsimas and King [4] successfully used discrete optimization techniques to solve instances with n, p in the 1, 000s within minutes. More recently, Bertsimas and Van Parys [5], Bertsimas et al [7] proposed a cutting plane approach which scales to data sizes of with n, p in the 100, 000s for ordinary least square and n, p in the 10, 000s for logistic regression. To the best of our knowledge, our approach is the only method which scales to instances of such sizes, while provably solving such an NP-hard problem.…”
Section: Integer Optimization Formulationmentioning
confidence: 99%
See 3 more Smart Citations