2020
DOI: 10.48550/arxiv.2009.09796
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Multi-Task Learning with Deep Neural Networks: A Survey

Michael Crawshaw

Abstract: Multi-task learning (MTL) is a subfield of machine learning in which multiple tasks are simultaneously learned by a shared model. Such approaches offer advantages like improved data efficiency, reduced overfitting through shared representations, and fast learning by leveraging auxiliary information. However, the simultaneous learning of multiple tasks presents new design and optimization challenges, and choosing which tasks should be learned jointly is in itself a non-trivial problem. In this survey, we give a… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
144
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
5
4
1

Relationship

0
10

Authors

Journals

citations
Cited by 121 publications
(144 citation statements)
references
References 116 publications
0
144
0
Order By: Relevance
“…expert training stage, is to learn representation for each specialized task type such as image classification, object detection, and semantic segmentation. This design naturally mitigates the learning difficulty caused by task conflicts [12] in multi-task learning setups. The representation learned for each task type performs better than simple ImageNet pretraining when tested on other tasks of the same type [29] strengthened knowledge specifically for that task type.…”
Section: Easy Extensibility and Great Generalizabilitymentioning
confidence: 99%
“…expert training stage, is to learn representation for each specialized task type such as image classification, object detection, and semantic segmentation. This design naturally mitigates the learning difficulty caused by task conflicts [12] in multi-task learning setups. The representation learned for each task type performs better than simple ImageNet pretraining when tested on other tasks of the same type [29] strengthened knowledge specifically for that task type.…”
Section: Easy Extensibility and Great Generalizabilitymentioning
confidence: 99%
“…Therefore, in order to create an effective MTL architecture it is important to analyze how to combine the shared modules (layers) and task specific modules and what portion of model's parameters will be shared between tasks. In conventional MTL, the parameter sharing approach is classified as [18]-…”
Section: Multi-task Learning (Mtl)mentioning
confidence: 99%
“…Although unified QG encoding enables models to process question generation across formats, how to effectively and efficiently train a QG model across multiple datasets is still challenging. A straightforward solution is to use multitask learning [16], but it needs to retrain the QG model using all the historical data whenever a new dataset is available. As a result, it is not scalable due to the linearly increasing computation and storage [6] costs.…”
Section: Introductionmentioning
confidence: 99%