Proceedings of the 2nd Workshop on Natural Language Processing for Conversational AI 2020
DOI: 10.18653/v1/2020.nlp4convai-1.7
|View full text |Cite
|
Sign up to set email alerts
|

DLGNet: A Transformer-based Model for Dialogue Response Generation

Abstract: Neural dialogue models, despite their successes, still suffer from lack of relevance, diversity, and in many cases coherence in their generated responses. On the other hand, transformer-based models such as GPT-2 have demonstrated an excellent ability to capture long-range structures in language modeling tasks. In this paper, we present DLGNet, a transformer-based model for dialogue modeling. We specifically examine the use of DL-GNet for multi-turn dialogue response generation. In our experiments, we evaluate… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
11
0

Year Published

2021
2021
2023
2023

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 24 publications
(11 citation statements)
references
References 13 publications
0
11
0
Order By: Relevance
“…Nevertheless, generating such responses in dialogue is still an open problem using current state-of-the-art language models, such as LLM's. Though LLM's generate realistic text outputs, allowing free-form natural language responses in dialogue is unpredictable, often forgets context, and can lead to incorrect or, in the worst case, offensive responses [97,19,65,98]. In settings where it is not acceptable to have unpredictable and potentially harmful responses, it may not be appropriate to use dynamically generated responses by LLM's.…”
Section: Responding Appropriately In Dialoguementioning
confidence: 99%
“…Nevertheless, generating such responses in dialogue is still an open problem using current state-of-the-art language models, such as LLM's. Though LLM's generate realistic text outputs, allowing free-form natural language responses in dialogue is unpredictable, often forgets context, and can lead to incorrect or, in the worst case, offensive responses [97,19,65,98]. In settings where it is not acceptable to have unpredictable and potentially harmful responses, it may not be appropriate to use dynamically generated responses by LLM's.…”
Section: Responding Appropriately In Dialoguementioning
confidence: 99%
“…Possible reasons for getting more negative or less positive outputs from models could be from two aspects: the transformer-based model and the dataset size for fine-tuning. The OpenAI GPT-2 is a transformer-based model [102]. The transformer is a model architecture that can directly model the dependency between every two words in a sequence to allow the model to learn language representations and generate outputs like the natural language [46].…”
Section: Discussionmentioning
confidence: 99%
“…However, using length normalized log-likelihoods (Brown et al, 2020) has become standard due to its superior performance, and is commonly used for generation tasks (Mao et al, 2019;Oluwatobi and Mueller, 2020). For causal language models, e.g., GPT-2 and GPT-3, Equation 1 can be decomposed as:…”
Section: Standard Methodsmentioning
confidence: 99%