2020
DOI: 10.48550/arxiv.2004.13637
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Recipes for building an open-domain chatbot

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

1
216
1

Year Published

2022
2022
2023
2023

Publication Types

Select...
5
2

Relationship

0
7

Authors

Journals

citations
Cited by 128 publications
(259 citation statements)
references
References 0 publications
1
216
1
Order By: Relevance
“…Our work is also closely related to recent successes in applying language models to dialog modeling (e.g., [25,26,17,18]), which built on earlier research in neural dialog modeling (e.g., [14,15,16,27,28]). One of our fine-tuning stages requires training on dialog-only data, which is related to Wolf et al [29], Dinan et al [25] and Zhang et al [30].…”
Section: Related Workmentioning
confidence: 82%
See 2 more Smart Citations
“…Our work is also closely related to recent successes in applying language models to dialog modeling (e.g., [25,26,17,18]), which built on earlier research in neural dialog modeling (e.g., [14,15,16,27,28]). One of our fine-tuning stages requires training on dialog-only data, which is related to Wolf et al [29], Dinan et al [25] and Zhang et al [30].…”
Section: Related Workmentioning
confidence: 82%
“…One of our fine-tuning stages requires training on dialog-only data, which is related to Wolf et al [29], Dinan et al [25] and Zhang et al [30]. Our use of fine-tuning on crowdworker-annotated data to improve interestingness is comparable to Roller et al [18]. However, we aim to maximize the interestingness of the model's output distinctly from its ability to engage the user in further interaction.…”
Section: Related Workmentioning
confidence: 99%
See 1 more Smart Citation
“…Adiwardana et al (2020) proposed Meena, a multi-turn open-domain chatbot trained end-to-end on data mined and filtered from public domain social media conversations. Blender (Roller et al, 2020;Xu et al, 2021) learn to provide engaging talking points and listen to their partners, as well as displaying knowledge, empathy and personality appropriately, while maintaining a consistent persona. Adapter-bot (Madotto et al, 2021) explored prompt-based few-shot learning in dialogue tasks.…”
Section: Related Workmentioning
confidence: 99%
“…Open-domain dialogue is a long-standing problem in natural language processing and has aroused the widespread interest of researchers. Many approaches have been studied, and recently, generation models trained on large-scale data have gained more attention (Adiwardana et al, 2020;Roller et al, 2020;Xu et al, 2021;Madotto et al, 2021;Bao et al, 2019Bao et al, , 2020Zhang et al, 2019;. Open-domain dialogue systems are born to deal with many diverse domains, and naturally its training data, usually crawled from online * Corresponding Author resources such as Reddit and Twitter, are heterogeneous and contain utterances with many various topics, more freedom of topic shifting, and vague responses (Kummerfeld et al, 2018).…”
Section: Introductionmentioning
confidence: 99%