ICASSP 2019 - 2019 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2019
DOI: 10.1109/icassp.2019.8683488
|View full text |Cite
|
Sign up to set email alerts
|

Bi-directional Lattice Recurrent Neural Networks for Confidence Estimation

Abstract: The standard approach to mitigate errors made by an automatic speech recognition system is to use confidence scores associated with each predicted word. In the simplest case, these scores are word posterior probabilities whilst more complex schemes utilise bi-directional recurrent neural network (BiRNN) models. A number of upstream and downstream applications, however, rely on confidence scores assigned not only to 1-best hypotheses but to all words found in confusion networks or lattices. These include but ar… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
2

Citation Types

0
28
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
6
1

Relationship

1
6

Authors

Journals

citations
Cited by 27 publications
(28 citation statements)
references
References 36 publications
0
28
0
Order By: Relevance
“…Those DAGs are highly flexible structures that can be additionally enriched with a wide range of features [21,22]. Recently there has been much interest in examining neural network extensions to DAGs and other general graph structures [23,18,24]. The key question that any such approach needs to answer is how information associated with multiple graph arcs or nodes is combined.…”
Section: Lattice Recurrent Neural Networkmentioning
confidence: 99%
See 2 more Smart Citations
“…Those DAGs are highly flexible structures that can be additionally enriched with a wide range of features [21,22]. Recently there has been much interest in examining neural network extensions to DAGs and other general graph structures [23,18,24]. The key question that any such approach needs to answer is how information associated with multiple graph arcs or nodes is combined.…”
Section: Lattice Recurrent Neural Networkmentioning
confidence: 99%
“…The confidence score prediction is then done using equation (2). The targets for lattice arc confidence scores are generated by extending the alignment algorithm for one-best sequences as described in [18].…”
Section: Lattice Recurrent Neural Networkmentioning
confidence: 99%
See 1 more Smart Citation
“…lattices or confusion networks [8,9]. Improved confidence estimation can be achieved by using model-based approaches, such as conditional random fields [10], recurrent neural networks [11,12] and graph neural networks [13], or leveraging more related information including phonetics, word/phone duration and language models [2,14].…”
Section: Introductionmentioning
confidence: 99%
“…We are particularly interested in "data-driven" confidence models [2] that are trained on recognition examples to learn systematic "mistakes" made by the speech recognizer and actively "correct" them. A major limitation in such confidence modeling methods in the literature is that they only look at equal-error rate (EER) [1] or normalized cross entropy (NCE) [3] and do not investigate their impact on speech recognizer accuracy in terms of word error rate (WER).…”
Section: Introductionmentioning
confidence: 99%