Proceedings of the 18th ACM SIGKDD International Conference on Knowledge Discovery and Data Mining 2012
DOI: 10.1145/2339530.2339647
|View full text |Cite
|
Sign up to set email alerts
|

Batch mode active sampling based on marginal probability distribution matching

Abstract: Active Learning is a machine learning and data mining technique that selects the most informative samples for labeling and uses them as training data; it is especially useful when there are large amount of unlabeled data and labeling them is expensive. Recently, batch-mode active learning, where a set of samples are selected concurrently for labeling, based on their collective merit, has attracted a lot of attention. The objective of batch-mode active learning is to select a set of informative samples so that … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
75
0

Year Published

2015
2015
2022
2022

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 53 publications
(75 citation statements)
references
References 21 publications
0
75
0
Order By: Relevance
“…from the same distribution [23]. Based on this, Chattopadhyay et al proposed a novel selection strategy for single-label active learning in [20], which iteratively selects a set of query samples Q from U, such that P(X, Y) represented by L ∪ Q and U\Q are similar to each other. Since the conditional probability P(Y|X) remains the same for both L and U as they are drawn from the same underlying distribution, the problem reduces to selecting Q such that the marginal probability P L∪Q (X) is similar to P U\Q (X).…”
Section: The Proposed Methodsmentioning
confidence: 99%
See 4 more Smart Citations
“…from the same distribution [23]. Based on this, Chattopadhyay et al proposed a novel selection strategy for single-label active learning in [20], which iteratively selects a set of query samples Q from U, such that P(X, Y) represented by L ∪ Q and U\Q are similar to each other. Since the conditional probability P(Y|X) remains the same for both L and U as they are drawn from the same underlying distribution, the problem reduces to selecting Q such that the marginal probability P L∪Q (X) is similar to P U\Q (X).…”
Section: The Proposed Methodsmentioning
confidence: 99%
“…Similar to [24,25], we employ maximal mean discrepancy (MMD) to empirically measure the difference between the marginal probability distributions of two sets, which has been widely used for this purpose [20,26,27]. It is a statistical test based on the fact that two distributions are different if and only if there exists at least one function in a universal reproducing kernel Hilbert space (RKHS) having different expectations on the two distributions [28].…”
Section: The Proposed Methodsmentioning
confidence: 99%
See 3 more Smart Citations