Proceedings of the 58th Annual Meeting of the Association for Computational Linguistics 2020
DOI: 10.18653/v1/2020.acl-main.183
|View full text |Cite
|
Sign up to set email alerts
|

Can You Put it All Together: Evaluating Conversational Agents’ Ability to Blend Skills

Abstract: Being engaging, knowledgeable, and empathetic are all desirable general qualities in a conversational agent. Previous work has introduced tasks and datasets that aim to help agents to learn those qualities in isolation and gauge how well they can express them. But rather than being specialized in one single quality, a good open-domain conversational agent should be able to seamlessly blend them all into one cohesive conversational flow. In this work, we investigate several ways to combine models trained toward… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1

Citation Types

0
128
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
3
3

Relationship

0
10

Authors

Journals

citations
Cited by 118 publications
(128 citation statements)
references
References 12 publications
0
128
0
Order By: Relevance
“…Transformer models represent the state-of-the-art for many natural language processing (NLP) tasks, such as question-answering (Devlin et al, 2019), dialogue (Smith et al, 2020), search results (Nayak, 2019), and more. Popular pre-trained models, such as those available from Hugging Face (Wolf et al, 2019), allow developers without extensive computation power to benefit from these models.…”
Section: Introductionmentioning
confidence: 99%
“…Transformer models represent the state-of-the-art for many natural language processing (NLP) tasks, such as question-answering (Devlin et al, 2019), dialogue (Smith et al, 2020), search results (Nayak, 2019), and more. Popular pre-trained models, such as those available from Hugging Face (Wolf et al, 2019), allow developers without extensive computation power to benefit from these models.…”
Section: Introductionmentioning
confidence: 99%
“…Transformer-based chatbots, that are trained on a large amount of dialogue data [38], [64], are capable of generating a diverse range of responses. To evaluate such chatbots, our framework considers the specificity of dialogue responses within their corresponding contexts.…”
Section: Response Sampling Policymentioning
confidence: 99%
“…Finally, human annotators label each of the remaining candidates as good or bad, with justifications (Section 2.3). We augment the Schema-Guided Dialogue (SGD) (Rastogi et al, 2020) and MultiWOZ 2.1 (Eric et al, 2020) corpora using the proposed approach. (See Figure 1 or Appendix A.4 for examples.)…”
Section: Introductionmentioning
confidence: 99%