2014
DOI: 10.1371/journal.pone.0087357
|View full text |Cite
|
Sign up to set email alerts
|

Mutual Information between Discrete and Continuous Data Sets

Abstract: Mutual information (MI) is a powerful method for detecting relationships between data sets. There are accurate methods for estimating MI that avoid problems with “binning” when both data sets are discrete or when both data sets are continuous. We present an accurate, non-binning MI estimator for the case of one discrete data set and one continuous data set. This case applies when measuring, for example, the relationship between base sequence and gene expression level, or the effect of a cancer drug on patient … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
360
0
4

Year Published

2016
2016
2024
2024

Publication Types

Select...
7

Relationship

0
7

Authors

Journals

citations
Cited by 627 publications
(364 citation statements)
references
References 3 publications
0
360
0
4
Order By: Relevance
“…Here we extend the GCMI estimate introduced above to this problem—estimating MI between a (univariate) discrete and a (potentially multivariate) continuous variable. Despite the wide applicability of such a measure, the practical issue of computing MI between discrete and continuous variables has so far received little attention [Lefakis and Fleuret, 2014; Magri et al, 2009; Ross, 2014]. One approach would be to discretize the continuous variable as described in Section 2.3 and use standard binned methods.…”
Section: A Novel Methods For MI Estimation Using a Gaussian Copulamentioning
confidence: 99%
“…Here we extend the GCMI estimate introduced above to this problem—estimating MI between a (univariate) discrete and a (potentially multivariate) continuous variable. Despite the wide applicability of such a measure, the practical issue of computing MI between discrete and continuous variables has so far received little attention [Lefakis and Fleuret, 2014; Magri et al, 2009; Ross, 2014]. One approach would be to discretize the continuous variable as described in Section 2.3 and use standard binned methods.…”
Section: A Novel Methods For MI Estimation Using a Gaussian Copulamentioning
confidence: 99%
“…We assume ergodicity of the processes when evaluating TE estimates. In all simulations, when using the nearest neighbours method, we set the number of neighbours as k = 3, as recommended in the literature [19], [24]. Besides, when using the binning method, we set the number of bins as…”
Section: Resultsmentioning
confidence: 99%
“…Inspired in the work of Kraskov et al, Ross developed a similar method to estimate mutual information for a mixed case, that is, to estimate mutual information between discrete and continuous random variables [19]. Ross estimator has been indicated as more efficient than popular binning estimator, even when binning is applied with bias correction [26].…”
Section: B Nearest Neighbours Methodsmentioning
confidence: 99%
See 2 more Smart Citations