Interspeech 2017 2017
DOI: 10.21437/interspeech.2017-1178
|View full text |Cite
|
Sign up to set email alerts
|

ASR Error Management for Improving Spoken Language Understanding

Abstract: This paper addresses the problem of automatic speech recognition (ASR) error detection and their use for improving spoken language understanding (SLU) systems. In this study, the SLU task consists in automatically extracting, from ASR transcriptions, semantic concepts and concept/values pairs in a e.g touristic information system. An approach is proposed for enriching the set of semantic labels with error specific labels and by using a recently proposed neural approach based on word embeddings to compute well … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
2
1

Citation Types

0
34
0
3

Year Published

2017
2017
2023
2023

Publication Types

Select...
4
2
2

Relationship

1
7

Authors

Journals

citations
Cited by 30 publications
(37 citation statements)
references
References 18 publications
0
34
0
3
Order By: Relevance
“…We think the starred mode must not be applied too early in the training chain, and would be applied on sub-tasks very close to the final target. In comparison to the literature, 16.4% of CER and 20.9% of CVER are very good results, since the best results ever published on the MEDIA test, when analysing speech instead of manual transcriptions, were a CER of 19.9% and a CVER of 25.1% [8]. For fair comparison with a state-of-the-art approach, we present in the next section a pipeline system we develop that takes benefits of the very good quality of our ASR outputs.…”
Section: Performance Of Curriculum-based Transfer Learningmentioning
confidence: 62%
See 1 more Smart Citation
“…We think the starred mode must not be applied too early in the training chain, and would be applied on sub-tasks very close to the final target. In comparison to the literature, 16.4% of CER and 20.9% of CVER are very good results, since the best results ever published on the MEDIA test, when analysing speech instead of manual transcriptions, were a CER of 19.9% and a CVER of 25.1% [8]. For fair comparison with a state-of-the-art approach, we present in the next section a pipeline system we develop that takes benefits of the very good quality of our ASR outputs.…”
Section: Performance Of Curriculum-based Transfer Learningmentioning
confidence: 62%
“…For our experiments, we feed them with outputs from our end-to-end ASR system (with LM rescoring) fine tuned one the MEDIA training data. The word error rate (WER) of this system on the MEDIA data is 9.3%, that is very good in comparison to the 23.6% of WER got by the ASR used to reach the best CER/CVER values in the literacy until now [8]. Comparisons in CER/CVER values between state-of-the art pipeline approach and end-to-end system are provided in table 4.…”
Section: Performance Of Curriculum-based Transfer Learningmentioning
confidence: 79%
“…Both sets are described in section 4. Similar to [6] we report NLU performances on the VocADom@A4H test set using the concept error rate (CER) for slot labels. Intent classification is evaluated using the F1-score.…”
Section: Pipeline Slu Baseline Approachmentioning
confidence: 94%
“…More recently, acoustic word embeddings for ASR error detection were trained through a convolutional neural network (CNN) based ASR model to detect erroneous words [6]. Output of this ASR model was fed to a conditional random fields (CRF) model and an attention-based RNN NLU model.…”
Section: Related Workmentioning
confidence: 99%
“…More recently, to improve ASR error handling, acoustic word embeddings for ASR error detection were trained through a convolutional neural network (CNN) based ASR model to detect erroneous words. Output of this ASR model is fed to conditional random fields (CRF) and an attentionbased RNN NLU model [6]. The CRF outperformed the RNN approach and the concept error rate (CER) decreased by 1% integrating confidence measures.…”
Section: Related Workmentioning
confidence: 99%