Findings of the Association for Computational Linguistics: EMNLP 2021 2021
DOI: 10.18653/v1/2021.findings-emnlp.286
|View full text |Cite
|
Sign up to set email alerts
|

Distilling the Knowledge of Large-scale Generative Models into Retrieval Models for Efficient Open-domain Conversation

Abstract: Despite the remarkable performance of largescale generative models in open-domain conversation, they are known to be less practical for building real-time conversation systems due to high latency. On the other hand, retrieval models could return responses with much lower latency but show inferior performance to the large-scale generative models since the conversation quality is bounded by the pre-defined response set. To take advantage of both approaches, we propose a new training method called G2R (Generative… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1

Citation Types

0
4
0

Year Published

2022
2022
2024
2024

Publication Types

Select...
3
1

Relationship

1
3

Authors

Journals

citations
Cited by 4 publications
(4 citation statements)
references
References 32 publications
0
4
0
Order By: Relevance
“…Exemplar-based generative models (Wu et al, 2019;Cai et al, 2019b;Gupta et al, 2021) for open-domain conversation combine a retrieval model (Humeau et al, 2019;Mazare et al, 2018;Kim et al, 2021) and a generative model (Adiwardana et al, 2020;Roller et al, 2021; † Equal contribution * Corresponding author…”
Section: Introductionmentioning
confidence: 99%
“…Exemplar-based generative models (Wu et al, 2019;Cai et al, 2019b;Gupta et al, 2021) for open-domain conversation combine a retrieval model (Humeau et al, 2019;Mazare et al, 2018;Kim et al, 2021) and a generative model (Adiwardana et al, 2020;Roller et al, 2021; † Equal contribution * Corresponding author…”
Section: Introductionmentioning
confidence: 99%
“…Enhanced strategies were developed to detect and adjust to anomalies in task performance, ensuring a more stable learning curve and reducing the likelihood of overfitting on specific types of tasks [13][14][15]. Techniques to systematically evaluate the efficacy of adaptive learning algorithms were also explored, revealing that continuous adaptation leads to better long-term performance retention [16,17]. Strategies that incorporate elements of predictive modeling to anticipate future learning needs based on past performance data were also highlighted, showcasing a proactive approach to fine-tuning [1].…”
Section: Adaptive Learning Strategies For Dynamic Benchmarkingmentioning
confidence: 99%
“…The integration of scoring functions that incorporate ethical considerations has further enhanced the efficacy of best-of-N strategies in value alignment tasks [43,44]. Advanced optimization techniques have been employed to improve the computational efficiency of best-of-N strategies, making them more feasible for large-scale LLM applications [45,46]. The combination of best-of-N strategies with attention mechanisms has led to more contextually aware and relevant outputs, particularly in complex text generation tasks [47,48].…”
Section: Related Studiesmentioning
confidence: 99%