2020
DOI: 10.1103/physreve.101.042304
|View full text |Cite
|
Sign up to set email alerts
|

Improved mutual information measure for clustering, classification, and community detection

Abstract: The information theoretic quantity known as mutual information finds wide use in classification and community detection analyses to compare two classifications of the same set of objects into groups. In the context of classification algorithms, for instance, it is often used to compare discovered classes to known ground truth and hence to quantify algorithm performance. Here we argue that the standard mutual information, as commonly defined, omits a crucial term which can become large under real-world conditio… Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
62
0
2

Year Published

2020
2020
2024
2024

Publication Types

Select...
6
1

Relationship

0
7

Authors

Journals

citations
Cited by 57 publications
(64 citation statements)
references
References 22 publications
0
62
0
2
Order By: Relevance
“…Similarly, it may result in artificially large values, even when two non-random partitions are compared if these have a large number of groups. To counter balance for such bias, several metrics alternative to the NMI were introduced [59][60][61]. It seems that Significance tends to favor the detection of smallscale structures, potentially returning partitions with more communities (i.e.…”
Section: Experimental Results Of Resolution Limitmentioning
confidence: 99%
“…Similarly, it may result in artificially large values, even when two non-random partitions are compared if these have a large number of groups. To counter balance for such bias, several metrics alternative to the NMI were introduced [59][60][61]. It seems that Significance tends to favor the detection of smallscale structures, potentially returning partitions with more communities (i.e.…”
Section: Experimental Results Of Resolution Limitmentioning
confidence: 99%
“…For instance, Meilȃ [9] proves that the rescaling performed by some measures for normalization purposes, such as N M I, have the effect of breaking the Convex Additivity property. Moreover, Newman et al [30], like others [5], [10], [29], show that N M I tend to favor partitions with more clusters when compared with a reference partition (cf. no k-invariance), and that this behavior can be smoothed by correcting NMI for chance.…”
Section: E: Nmimentioning
confidence: 92%
“…In this category, the most frequent property is probably kinvariance. Certain measures such as the Normalized Mutual Information tend to favor partitions depending on the number of clusters they contain when compared with a reference partition [5], a bias that a number of authors want to avoid [3], [26]- [30]. For example, suppose that one compares a ground truth partition to two estimated partitions differing only in their number of clusters.…”
Section: ) Sensitivity To Partition Characteristicsmentioning
confidence: 99%
See 2 more Smart Citations