Proceedings of the 2021 SIAM International Conference on Data Mining (SDM) 2021
DOI: 10.1137/1.9781611976700.40
|View full text |Cite
|
Sign up to set email alerts
|

Better Short than Greedy: Interpretable Models through Optimal Rule Boosting

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
6
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
4
1
1

Relationship

0
6

Authors

Journals

citations
Cited by 8 publications
(9 citation statements)
references
References 0 publications
0
6
0
Order By: Relevance
“…This approach was pioneered by algorithms such as Lri (Weiss and Indurkhya, 2000) and SLipper (Cohen and Singer, 1999), the general framework of gradient boosting for rule learning was most clearly defined in ender (Dembczyński et al, 2010). Recent additions to this family include Boomer (Rapp et al, 2020), which generalizes this approach to learning multi-label rules, and the algorithm of Boley et al (2021), which replaced the greedy search for the best addition to the rule set with an efficient exhaustive search.…”
Section: Covering Algorithmsmentioning
confidence: 99%
“…This approach was pioneered by algorithms such as Lri (Weiss and Indurkhya, 2000) and SLipper (Cohen and Singer, 1999), the general framework of gradient boosting for rule learning was most clearly defined in ender (Dembczyński et al, 2010). Recent additions to this family include Boomer (Rapp et al, 2020), which generalizes this approach to learning multi-label rules, and the algorithm of Boley et al (2021), which replaced the greedy search for the best addition to the rule set with an efficient exhaustive search.…”
Section: Covering Algorithmsmentioning
confidence: 99%
“…In this paper, we investigate how to extend these ideas to the multi-label classiőcation setting. The problem of controlling the number of rules has also been studied for single-label rule boosting, where learned rules are combined additively [15]. An extension to multi-label classiőcation represents a possible direction of future work.…”
Section: Related Workmentioning
confidence: 99%
“…A similar issue arises in head sampling. The weight function in (15) grows exponentially with |D + |, so that cftp most likely returns the positive data records with the highest number of present features. Therefore, sampled heads tend to be very long and have small support (often 1).…”
Section: Limitations Of the Two-stage Pattern-sampling Frameworkmentioning
confidence: 99%
“…Nonetheless, the main limitation of these approaches is that they are based on a heuristic definition of a rule-based model, i.e., they add rules without a global optimal criteria. Over the past years, rule learning methods that go beyond greedy approaches have been developed, i.e., Monte-Carlo search for Bayesian rule lists (Letham et al 2015;Yang et al 2017), and branch-and-bound with tight bounds for decision lists (Angelino et al 2017) and rule sets (Boley et al 2021). However, the main limitation of these methods is that they can only be applied to small or mid-size datasets and are mostly limited to binary targets.…”
Section: Rule-based Classifiersmentioning
confidence: 99%
“…Another interesting and straightforward development would be the extension of our work to mixed targets, combining nominal numeric variables. second developments could go from upper-and-lower bounds to improvements in search methods and to study the feasibility of global search such as Markov Chain Monte Carlo methods used by Yang et al (2017) or branch-and-bound algorithms used by Boley et al (2021). In the third category, our approach could be formalised for subgroup sets, allowing for overlap between the subgroups.…”
Section: In Pattern Miningmentioning
confidence: 99%