2016
DOI: 10.1016/j.patcog.2015.11.007
|View full text |Cite
|
Sign up to set email alerts
|

Can high-order dependencies improve mutual information based feature selection?

Abstract: Mutual information (MI) based approaches are a popular paradigm for feature selection. Most previous methods have made use of low-dimensional MI quantities that are only effective at detecting low-order dependencies between variables. Several works have considered the use of higher dimensional mutual information, but the theoretical underpinning of these approaches is not yet comprehensive. To fill this gap, in this paper, we systematically investigate the issues of employing high-order dependencies for mutual… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
82
0

Year Published

2018
2018
2024
2024

Publication Types

Select...
5
1

Relationship

0
6

Authors

Journals

citations
Cited by 95 publications
(82 citation statements)
references
References 15 publications
0
82
0
Order By: Relevance
“…For example, JMI [19] criteria can be derived setting the value of β = γ = 1 |S| . In [17], the authors propose a new criterion by relaxing the the first assumption. They show under the relaxed assumption that the selected features are conditionally independent given the f m and another feature f i in S, the redundancy term can be approximated as the following…”
Section: Information Theoretic Feature Selection Methodsmentioning
confidence: 99%
See 3 more Smart Citations
“…For example, JMI [19] criteria can be derived setting the value of β = γ = 1 |S| . In [17], the authors propose a new criterion by relaxing the the first assumption. They show under the relaxed assumption that the selected features are conditionally independent given the f m and another feature f i in S, the redundancy term can be approximated as the following…”
Section: Information Theoretic Feature Selection Methodsmentioning
confidence: 99%
“…Otherwise, f m is discarded considering that it does not contribute to the score significantly. While selecting a new feature, its discretization level is also shifted by a small value δ from its original value (as selected previously based on J rel as shown in line [16][17][18][19][20][21]. This process helps to select the discretization level of features dynamically considering its dependency with other feature.…”
Section: Discretization and Feature Selection Based On Bias Correctedmentioning
confidence: 99%
See 2 more Smart Citations
“…107 Since MI based filter methods do not extract new features and thus are more 108 interpretable, parallel to the development in Deep learning, there has been a lot of effort 109 to better approximate MI measures such as relevancy and redundancy. New Information 110 theoretic measures such as complementary information, the additional information that 111 a gene has about the class, which is not found in the already selected subset of genes 112 have been proposed [15,27]. These methods attempt to estimate the joint mutual 113 information of a feature subset with the class.…”
Section: Introduction 18mentioning
confidence: 99%