2019
DOI: 10.1609/aaai.v33i01.33014763
|View full text |Cite
|
Sign up to set email alerts
|

Training Complex Models with Multi-Task Weak Supervision

Abstract: As machine learning models continue to increase in complexity, collecting large hand-labeled training sets has become one of the biggest roadblocks in practice. Instead, weaker forms of supervision that provide noisier but cheaper labels are often used. However, these weak supervision sources have diverse and unknown accuracies, may output correlated labels, and may label different tasks or apply at different levels of granularity. We propose a framework for integrating and modeling such weak supervision sourc… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
177
0

Year Published

2019
2019
2023
2023

Publication Types

Select...
5
1
1
1

Relationship

2
6

Authors

Journals

citations
Cited by 121 publications
(177 citation statements)
references
References 23 publications
0
177
0
Order By: Relevance
“…For information extraction tasks such as relation extraction and entity typing, distant supervision (Mintz et al, 2009) is a powerful approach for adding more data, using a knowledge base (Del Corro et al, 2015;Rabinovich and Klein, 2017) or heuristics (Ratner et al, 2016;Hancock et al, 2018) to automatically label instances. One can treat this data just like any other supervised data, but it is noisy; more effective approaches employ specialized probabilistic models (Riedel et al, 2010;Ratner et al, 2018a), capturing its interaction with other supervision (Wang and Poon, 2018) or breaking down aspects of a task on which it is reliable (Ratner et al, 2018b). However, these approaches often require sophisticated probabilistic inference for training of the final model.…”
Section: Introductionmentioning
confidence: 99%
“…For information extraction tasks such as relation extraction and entity typing, distant supervision (Mintz et al, 2009) is a powerful approach for adding more data, using a knowledge base (Del Corro et al, 2015;Rabinovich and Klein, 2017) or heuristics (Ratner et al, 2016;Hancock et al, 2018) to automatically label instances. One can treat this data just like any other supervised data, but it is noisy; more effective approaches employ specialized probabilistic models (Riedel et al, 2010;Ratner et al, 2018a), capturing its interaction with other supervision (Wang and Poon, 2018) or breaking down aspects of a task on which it is reliable (Ratner et al, 2018b). However, these approaches often require sophisticated probabilistic inference for training of the final model.…”
Section: Introductionmentioning
confidence: 99%
“…Specifically, the conditional independence assumption made in IVY for the inclusion of individual SNPs is valid because the same assumption is made for methods employing allele scores (Sebastiani et al, 2012). We point out the possibility of extending beyond the assumption of conditional independence such as handling SNPs occurring physically upon the same chromosome which display some degree of genetic linkage (Ratner et al, 2018). Empirical validation of such an extension is left for future research.…”
Section: Discussionmentioning
confidence: 96%
“…In IVY, we also made use of such SNPs. We refer interested readers to (Ratner et al, 2018) for a discussion of scenarios such as handling dependencies among * 's. In this case, Ivy generalizes beyond the standard conditional independence assumption.…”
Section: Assumptions and Problem Formulationmentioning
confidence: 99%
See 1 more Smart Citation
“…Moreover, motivated by the increasing volume of digitized but unlabeled medical data, we evaluate the scaling of our approach with respect to additional unlabeled data, and find that performance consistently increases as more unlabeled data is collected in a way that is consistent with theoretical predictions. 33,34 In summary, cross-modal data programming can lower a substantial barrier to machine learning model development in medicine by serving as a fundamental new interface that reduces labeling time required from domain experts, thereby providing a stepping stone towards widespread adoption of machine learning models to provide positive, tangible clinical impact.…”
mentioning
confidence: 99%