Proceedings of the 28th International Conference on Computational Linguistics 2020
DOI: 10.18653/v1/2020.coling-main.364
|View full text |Cite
|
Sign up to set email alerts
|

Diverse dialogue generation with context dependent dynamic loss function

Abstract: Dialogue systems using deep learning have achieved generation of fluent response sentences to user utterances. Nevertheless, they tend to produce responses that are not diverse and which are less context-dependent. To address these shortcomings, we propose a new loss function, an Inverse N-gram Frequency (INF) loss, which incorporates contextual fluency and diversity at the same time by a simple formula. Our INF loss can adjust its loss dynamically by a weight using the inverse frequency of the tokens' n-gram … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1
1

Relationship

0
2

Authors

Journals

citations
Cited by 2 publications
(1 citation statement)
references
References 10 publications
0
1
0
Order By: Relevance
“…Several other approaches to promoting response diversity which involve introducing an alternate loss function such as the Maximum Mutual Information (MMI) objective (Li et al, 2016a), the Inverse Token Frequency (ITF) objective (Nakamura et al, 2018), and the Inverse N-gram Frequency (INF) objective (Ueyama and Kano, 2020), require considerable additional computation steps. On the other hand, adversarial learning-based (Li et al, 2017a) and embedding augmentation-based (Cao et al, 2021) approaches require extensive modifications to the standard training process, resulting in a significant increase in training complexity.…”
Section: Introductionmentioning
confidence: 99%
“…Several other approaches to promoting response diversity which involve introducing an alternate loss function such as the Maximum Mutual Information (MMI) objective (Li et al, 2016a), the Inverse Token Frequency (ITF) objective (Nakamura et al, 2018), and the Inverse N-gram Frequency (INF) objective (Ueyama and Kano, 2020), require considerable additional computation steps. On the other hand, adversarial learning-based (Li et al, 2017a) and embedding augmentation-based (Cao et al, 2021) approaches require extensive modifications to the standard training process, resulting in a significant increase in training complexity.…”
Section: Introductionmentioning
confidence: 99%