2020
DOI: 10.48550/arxiv.2012.00150
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

MUSCLE: Strengthening Semi-Supervised Learning Via Concurrent Unsupervised Learning Using Mutual Information Maximization

Abstract: Deep neural networks are powerful, massively parameterized machine learning models that have been shown to perform well in supervised learning tasks. However, very large amounts of labeled data are usually needed to train deep neural networks. Several semi-supervised learning approaches have been proposed to train neural networks using smaller amounts of labeled data with a large amount of unlabeled data. The performance of these semisupervised methods significantly degrades as the size of labeled data decreas… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
6
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(6 citation statements)
references
References 27 publications
0
6
0
Order By: Relevance
“…The right plot in Figure 1 shows the results for the CIFAR-100 model with the state-of-the-art neural network architecture and training algorithm as in [18]. For this model, the coreset upper bound-I is nonvacuous at all training checkpoints, while the coreset upper bound-II is non-vacuous for training set sizes larger than around 8000.…”
Section: E Experimental Resultsmentioning
confidence: 99%
See 4 more Smart Citations
“…The right plot in Figure 1 shows the results for the CIFAR-100 model with the state-of-the-art neural network architecture and training algorithm as in [18]. For this model, the coreset upper bound-I is nonvacuous at all training checkpoints, while the coreset upper bound-II is non-vacuous for training set sizes larger than around 8000.…”
Section: E Experimental Resultsmentioning
confidence: 99%
“…The above problem is similar to the Hilbert coreset problem posed in (8) with J(w) in ( 8) replaced by a slightly different quantity J( w, p). In contrast to (8), the objective function J( w, p) in (18) depends on two vectors { w, p}. Since the goal is to provide a bound for all D ∈ D (K) CH , one can consider the following optimization problem by adapting the problem in (8) to this case: max…”
Section: Risk Bounds Via Frank-wolfe Hilbert Coresetsmentioning
confidence: 99%
See 3 more Smart Citations