2021
DOI: 10.48550/arxiv.2110.13221
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

On Learning Prediction-Focused Mixtures

Abstract: Probabilistic models help us encode latent structures that both model the data and are ideally also useful for specific downstream tasks. Among these, mixture models and their time-series counterparts, hidden Markov models, identify discrete components in the data. In this work, we focus on a constrained capacity setting, where we want to learn a model with relatively few components (e.g. for interpretability purposes). To maintain prediction performance, we introduce prediction-focused modeling for mixtures, … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

1
0

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 5 publications
0
1
0
Order By: Relevance
“…The upstream setting allows us to implicitly train our classifier and topic model in a one-stage setting that is end-to-end. This has the benefit of allowing us to tune the trade-off between our classifier and topic model performance in a predictionconstrained framework, which has been shown to achieve better empirical results when latent variable models are used as a dimensionality reduction tool (Hughes et al, 2018;Sharma et al, 2021). Furthermore, the upstream setting allows us to introduce the document label classifier as a latent variable, enabling our model to work in semisupervised settings.…”
Section: Downstream Supervised Topic Modelsmentioning
confidence: 99%
“…The upstream setting allows us to implicitly train our classifier and topic model in a one-stage setting that is end-to-end. This has the benefit of allowing us to tune the trade-off between our classifier and topic model performance in a predictionconstrained framework, which has been shown to achieve better empirical results when latent variable models are used as a dimensionality reduction tool (Hughes et al, 2018;Sharma et al, 2021). Furthermore, the upstream setting allows us to introduce the document label classifier as a latent variable, enabling our model to work in semisupervised settings.…”
Section: Downstream Supervised Topic Modelsmentioning
confidence: 99%