Proceedings of the Second DialDoc Workshop on Document-Grounded Dialogue and Conversational Question Answering 2022
DOI: 10.18653/v1/2022.dialdoc-1.10
|View full text |Cite
|
Sign up to set email alerts
|

Retrieval-Free Knowledge-Grounded Dialogue Response Generation with Adapters

Abstract: To diversify and enrich generated dialogue responses, knowledge-grounded dialogue has been investigated in recent years. The existing methods tackle the knowledge grounding challenge by retrieving the relevant sentences over a large corpus and augmenting the dialogues with explicit extra information. Despite their success, however, the existing works have drawbacks on the inference efficiency. This paper proposes KnowExpert, an end-to-end framework to bypass the explicit retrieval process and inject knowledge … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
17
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
5
3
1

Relationship

1
8

Authors

Journals

citations
Cited by 18 publications
(17 citation statements)
references
References 39 publications
0
17
0
Order By: Relevance
“…S2S BERT replaces the Transformer Encoder with a pre-trained BERT (Devlin et al, 2019). KnowExpert (Xu et al, 2021) avoids the knowledge retrieval process and attempts to inject prior knowledge into the pre-trained language models for knowledge-grounded dialogue generation task. Essentially, KnowExpert stores knowledge in its parameters with lightweight adapters.…”
Section: A12 Our Supervised Methodsmentioning
confidence: 99%
“…S2S BERT replaces the Transformer Encoder with a pre-trained BERT (Devlin et al, 2019). KnowExpert (Xu et al, 2021) avoids the knowledge retrieval process and attempts to inject prior knowledge into the pre-trained language models for knowledge-grounded dialogue generation task. Essentially, KnowExpert stores knowledge in its parameters with lightweight adapters.…”
Section: A12 Our Supervised Methodsmentioning
confidence: 99%
“…In order to perform the posterior sampling of knowledge selection during joint training, some works have proposed to separately train the posterior distribution model (Paranjape et al, 2022;Lian et al, 2019) or the posterior information prediction model (Chen et al, 2020). Very recently, SPI (Xu et al, 2023) has applied short-run MCMC (Erik et al, 2019) for posterior sampling on the collaborative latent spaces.…”
Section: Unsupervised Joint Trainingmentioning
confidence: 99%
“…In specific, they have treated the retrieved knowledge, document, or passage as an unobserved latent variable and adapt latent variable modeling based on approximated marginalization (e.g. top-k) Huang et al, 2021;Cai et al, 2023;Guu et al, 2020), reinforcement learning (Zhao et al, 2020;Zhang et al, 2022;Chen et al, 2022; or variational methods (Zhan et al, 2021;Paranjape et al, 2022;Lian et al, 2019;Kim et al, 2020;Chen et al, 2020;Xu et al, 2023). However, joint training of the retriever along with the generator under this latent variable modeling has some restrictions in utilizing the retriever.…”
Section: Introductionmentioning
confidence: 99%
“…Furthermore, it also has been pointed out that using a knowledge base could reduce the problem of hallucinations (Dziri et al, 2021). Another research line tends to compress knowledge into model parameters, either by training set augmentation with template-based method (Madotto et al, 2020) or using neural architectures as domain-specific adapters (Xu et al, 2021).…”
Section: Grounded Dialogue Generationmentioning
confidence: 99%