2017
DOI: 10.48550/arxiv.1711.04992
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Feature importance scores and lossless feature pruning using Banzhaf power indices

Abstract: Understanding the influence of features in machine learning is crucial to interpreting models and selecting the best features for classification. In this work we propose the use of principles from coalitional game theory to reason about importance of features. In particular, we propose the use of the Banzhaf power index as a measure of influence of features on the outcome of a classifier. We show that features having Banzhaf power index of zero can be losslessly pruned without damage to classifier accuracy. Co… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1

Citation Types

0
3
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(3 citation statements)
references
References 2 publications
0
3
0
Order By: Relevance
“…It was reinvented by John F. Banzhaf III in 1964[Banzhaf III, 1964, and was reinvented once more by James Samuel Coleman in 1971 [Coleman, 1971] before it became part of the mainstream literature. In the field of machine learning, Banzhaf value has been previously applied to the problem of measuring feature importance [Datta et al, 2015, Kulynych and Troncoso, 2017, Sliwinski et al, 2019, Patel et al, 2021, Karczmarz et al, 2021. While these works suggest that Banzhaf value could be an alternative to the popular Shapley value-based model interpretation methods (e.g., [Lundberg and Lee, 2017]), it remains unclear in which settings the Banzhaf value may be preferable to the Shapley value.…”
Section: A Related Workmentioning
confidence: 99%
“…It was reinvented by John F. Banzhaf III in 1964[Banzhaf III, 1964, and was reinvented once more by James Samuel Coleman in 1971 [Coleman, 1971] before it became part of the mainstream literature. In the field of machine learning, Banzhaf value has been previously applied to the problem of measuring feature importance [Datta et al, 2015, Kulynych and Troncoso, 2017, Sliwinski et al, 2019, Patel et al, 2021, Karczmarz et al, 2021. While these works suggest that Banzhaf value could be an alternative to the popular Shapley value-based model interpretation methods (e.g., [Lundberg and Lee, 2017]), it remains unclear in which settings the Banzhaf value may be preferable to the Shapley value.…”
Section: A Related Workmentioning
confidence: 99%
“…The LR model's coefficients have been widely utilized for feature importance estimation [67]. Each coefficient represents a score, known as the feature importance score, which describes the significance level between the feature and the target variable.…”
Section: Feature Selectionmentioning
confidence: 99%
“…The higher the coefficient, the more relevant the feature is to the target variable. In other words, coefficients can be utilized to determine the important and unimportant features to avoid overfitting [67] and are thus useful for prediction [68]. The RFE model ranks the 104 features based on their importance scores obtained from the LR model into a list, in which the first position represents the most significant feature, while the least important feature is ranked on the last position.…”
Section: Feature Selectionmentioning
confidence: 99%