Proceedings of the 2019 Conference on Empirical Methods in Natural Language Processing and the 9th International Joint Conferen 2019
DOI: 10.18653/v1/d19-1315
|View full text |Cite
|
Sign up to set email alerts
|

Keeping Consistency of Sentence Generation and Document Classification with Multi-Task Learning

Abstract: The automated generation of information indicating the characteristics of articles such as headlines, key phrases, summaries and categories helps writers to alleviate their workload. Previous research has tackled these tasks using neural abstractive summarization and classification methods. However, the outputs may be inconsistent if they are generated individually. The purpose of our study is to generate multiple outputs consistently. We introduce a multi-task learning model with a shared encoder and multiple… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
16
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2

Relationship

0
6

Authors

Journals

citations
Cited by 11 publications
(16 citation statements)
references
References 25 publications
0
16
0
Order By: Relevance
“…[62] adds sentence-level sentiment classification and attention-level supervision to assist the primary stance detection task. [85] adds attention-level supervision to improve consistency of the two primary language generation tasks. [16] minimizes an auxiliary cosine softmax loss based on the audio encoder to learn more accurate speech-to-semantic mappings.…”
Section: Vanillamentioning
confidence: 99%
See 1 more Smart Citation
“…[62] adds sentence-level sentiment classification and attention-level supervision to assist the primary stance detection task. [85] adds attention-level supervision to improve consistency of the two primary language generation tasks. [16] minimizes an auxiliary cosine softmax loss based on the audio encoder to learn more accurate speech-to-semantic mappings.…”
Section: Vanillamentioning
confidence: 99%
“…Learning from multiple tasks makes it possible for learning models to capture generalized and complementary knowledge from the tasks at hand besides task-specific features. Tasks in MTL can be tasks with assumed relatedness [20,23,40,56,121], tasks with different styles of supervision (e.g., supervised and unsupervised tasks [41,64,73]), tasks with different types of goals (e.g., classification and generation [85]), tasks with different levels of features (e.g., token-level and sentence-level features [57,109]), and even tasks in different modalities (e.g., text and image data [66,115]). Alternatively, we can treat the same task in multiple hierarchical architecture models the hierarchical relationships between tasks.…”
Section: Introductionmentioning
confidence: 99%
“…3. To maintain uniformity between the attention weights of different tasks, we utilise consistency loss (Nishino et al, 2019) in addition to the original task-specific losses.…”
Section: Happymentioning
confidence: 99%
“…We use the "consistency loss" (Nishino et al, 2019) to reduce the difference between the attention weights from different tasks. Attention agreement favours emotional words while decoding the responses.…”
Section: Consistency Lossmentioning
confidence: 99%
“…Reason being that several relational facts may overlap in a sentence (Zhang et al, 2018). Although a conventional MTL method may learn task-specific features and has been successfully applied in a wide variety of scenarios (Zhang and Wang, 2016;Wu et al, 2016;Goo et al, 2018;Han et al, 2019;Nishino et al, 2019;Hu et al, 2019), its flat structure restricts the model to effectively learn the correlations between tasks. For example in Figure 1(a), the model cannot explicitly learn correlations between the two tasks.…”
Section: Introductionmentioning
confidence: 99%