2021
DOI: 10.1609/aaai.v35i12.17323
|View full text |Cite
|
Sign up to set email alerts
|

Multi-task Learning by Leveraging the Semantic Information

Abstract: One crucial objective of multi-task learning is to align distributions across tasks so that the information between them can be transferred and shared. However, existing approaches only focused on matching the marginal feature distribution while ignoring the semantic information, which may hinder the learning performance. To address this issue, we propose to leverage the label information in multi-task learning by exploring the semantic conditional relations among tasks. We first theoretically analyze the gene… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1

Citation Types

0
2
0

Year Published

2022
2022
2025
2025

Publication Types

Select...
3
3

Relationship

0
6

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 31 publications
0
2
0
Order By: Relevance
“…Multi-tasking learning (Collobert and Weston, 2008;Zhang and Yang, 2017;Ahmad et al, 2018;Liu et al, 2019a;Zhou et al, 2021) is a common way to incorporate the knowledge of several source tasks into language models. It fine-tunes a language model with multiple objectives for source tasks at the same time.…”
Section: Comparison To Multi-tasking Trainingmentioning
confidence: 99%
See 1 more Smart Citation
“…Multi-tasking learning (Collobert and Weston, 2008;Zhang and Yang, 2017;Ahmad et al, 2018;Liu et al, 2019a;Zhou et al, 2021) is a common way to incorporate the knowledge of several source tasks into language models. It fine-tunes a language model with multiple objectives for source tasks at the same time.…”
Section: Comparison To Multi-tasking Trainingmentioning
confidence: 99%
“…We compare our proposed prefix-based training with multi-tasking training (Collobert and Weston, 2008;Zhang and Yang, 2017;Ahmad et al, 2018;Liu et al, 2019a;Zhou et al, 2021). Both approaches are trained on the same source tasks and use pre-trained RoBERTa-large (Liu et al, 2019b).…”
Section: Baselines For Comparisonmentioning
confidence: 99%