2019
DOI: 10.48550/arxiv.1908.04812
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

An Effective Domain Adaptive Post-Training Method for BERT in Response Selection

Abstract: We focus on multi-turn response selection in a retrieval-based dialog system. In this paper, we utilize the powerful pre-trained language model Bi-directional Encoder Representations from Transformer (BERT) for a multi-turn dialog system and propose a highly effective post-training method on domain-specific corpus. Although BERT is easily adopted to various NLP tasks and outperforms previous baselines of each task, it still has limitations if a task corpus is too focused on a certain domain. Post-training on d… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
34
0

Year Published

2019
2019
2024
2024

Publication Types

Select...
4
4

Relationship

1
7

Authors

Journals

citations
Cited by 20 publications
(34 citation statements)
references
References 14 publications
0
34
0
Order By: Relevance
“…We fine-tune BERT (Devlin et al, 2019) (bert-basecased) for conversation response ranking using the huggingface-transformers (Wolf et al, 2019). We follow recent research in IR that employed finetuned BERT for retrieval tasks (Nogueira and Cho, 2019;, including conversation response ranking (Penha and Hauff, 2020;Vig and Ramea, 2019;Whang et al, 2019). When training BERT we employ a balanced number of relevant and non-relevant-sampled using BM25 (Robertson and Walker, 1994)-context and response pairs.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…We fine-tune BERT (Devlin et al, 2019) (bert-basecased) for conversation response ranking using the huggingface-transformers (Wolf et al, 2019). We follow recent research in IR that employed finetuned BERT for retrieval tasks (Nogueira and Cho, 2019;, including conversation response ranking (Penha and Hauff, 2020;Vig and Ramea, 2019;Whang et al, 2019). When training BERT we employ a balanced number of relevant and non-relevant-sampled using BM25 (Robertson and Walker, 1994)-context and response pairs.…”
Section: Implementation Detailsmentioning
confidence: 99%
“…Generally, most models formulate the response selection task as a dialog-response binary classification task. Whang et al (2019) first applied BERT for multi-turn response selection and obtained state-of-the-art results through further training BERT on domain-specific corpus. Subsequent researches (Lu et al, 2020;Gu et al, 2020) focused on modeling speaker information and showed its effectiveness in response retrieval.…”
Section: Related Workmentioning
confidence: 99%
“…This has shown to be effective in various tasks including review reading comprehension (Xu et al, 2019) and Su-perGLUE (Wang et al, 2019). Existing works on multi-turn response selection (Whang et al, 2019;Gu et al, 2020;Humeau et al, 2020) also adapted this post-training approach and obtained state-ofthe-art results. We also employ this post-training method in this work and show its effectiveness in improving performance (Section 5.1).…”
Section: Domain-specific Post-trainingmentioning
confidence: 99%
See 1 more Smart Citation
“…There are two types of formulation of next-utterance prediction. The first one is to generate the next utterance in a conversation given the conversation history (Zhao and Kawahara 2019;Dziri et al 2018;Hu et al 2019) and the second type is to retrieve the next utterance in the conversation from a large list of response utterances (Lowe et al 2015;Whang et al 2019). These tasks are useful for building a chatbot, aiming to generate or retrieve a good response according to the context of the conversation, while this work aims to recover the structure of the entire conversation.…”
Section: Related Workmentioning
confidence: 99%