2021
DOI: 10.1108/dta-12-2020-0312
|View full text |Cite
|
Sign up to set email alerts
|

Robust cross-lingual knowledge base question answering via knowledge distillation

Abstract: PurposePrevious knowledge base question answering (KBQA) models only consider the monolingual scenario and cannot be directly extended to the cross-lingual scenario, in which the language of questions and that of knowledge base (KB) are different. Although a machine translation (MT) model can bridge the gap through translating questions to the language of KB, the noises of translated questions could accumulate and further sharply impair the final performance. Therefore, the authors propose a method to improve … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2022
2022
2022
2022

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 24 publications
0
1
0
Order By: Relevance
“…The idea of distilling the knowledge in a neural network was first put forth by Hinton as a way to improve the performance of a machine learning algorithm by compressing the knowledge in a complex model into a single simple model (Hinton et al , 2015). Though this idea was primarily thought of as a compression mechanism rather than as an enhancement, in recent times, this idea has evolved to fit the needs for a performance enhancement (Wang and Dang, 2021; Hongyuan et al , 2022). In datasets used for speech and object recognition tasks, training must extract structure from these very large, highly redundant datasets, which often leads to a cumbersome model.…”
Section: Related Workmentioning
confidence: 99%
“…The idea of distilling the knowledge in a neural network was first put forth by Hinton as a way to improve the performance of a machine learning algorithm by compressing the knowledge in a complex model into a single simple model (Hinton et al , 2015). Though this idea was primarily thought of as a compression mechanism rather than as an enhancement, in recent times, this idea has evolved to fit the needs for a performance enhancement (Wang and Dang, 2021; Hongyuan et al , 2022). In datasets used for speech and object recognition tasks, training must extract structure from these very large, highly redundant datasets, which often leads to a cumbersome model.…”
Section: Related Workmentioning
confidence: 99%