2008 International Symposiums on Information Processing 2008
DOI: 10.1109/isip.2008.69
|View full text |Cite
|
Sign up to set email alerts
|

Tri-Training Based Learning from Positive and Unlabeled Data

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
3
2
1

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(3 citation statements)
references
References 10 publications
0
3
0
Order By: Relevance
“…The simplest assumption is SCAR (Selected Completely at Random Assumption), according to which the propensity score function, i.e. the probability of a labeling a positive observation, is constant [2,1,9,10,11]. Under the SCAR assumption, a possible approach is to estimate label frequency [12,13,14,15] and then use it to scale the posterior probabilities obtained from the naive method or, alternatively, optimize weighted empirical risk function with weights depending on the label frequency [1,16].…”
Section: Introductionmentioning
confidence: 99%
“…The simplest assumption is SCAR (Selected Completely at Random Assumption), according to which the propensity score function, i.e. the probability of a labeling a positive observation, is constant [2,1,9,10,11]. Under the SCAR assumption, a possible approach is to estimate label frequency [12,13,14,15] and then use it to scale the posterior probabilities obtained from the naive method or, alternatively, optimize weighted empirical risk function with weights depending on the label frequency [1,16].…”
Section: Introductionmentioning
confidence: 99%
“…Finally, select a good classifier from the set. For this step, we trained three classifiers: Support Vector Machine (SVM), J48 Decision Tree, and K-Nearest Neighbors (KNN) simultaneously using a modified version of the Co-Training algorithm [44] called Tri-Training [45]. In Tri-training, the learning process of each classifier is greatly influenced by the other two classifiers, akin to a majority-vote scheme.…”
Section: Pu-learningmentioning
confidence: 99%
“…The Tri-training algorithm [47,45] incorporates the E step from the EM algorithm and a modified version of the idea of Co-Training in order to accommodate a third classifier. In the original algorithm, three SVM classifiers were used.…”
Section: Tri-trainingmentioning
confidence: 99%