2016 IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP) 2016
DOI: 10.1109/icassp.2016.7472853
|View full text |Cite
|
Sign up to set email alerts
|

Groupwise learning for ASR k-best list reranking in spoken language translation

Abstract: This paper studies the enhancement of spoken language translation (SLT) with groupwise learning. Groupwise features were constructed by grouping pairs, triplets or M -plets of the ASR k-best outputs. Regression and classification models were learnt and a straightforward score combination strategy was used to capture the ranking relationship. Groupwise learning with pairwise regression models give the biggest gain over simple support vector regression models. Groupwise learning is robust to sentences with diffe… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1

Citation Types

0
3
0

Year Published

2016
2016
2020
2020

Publication Types

Select...
5
1

Relationship

1
5

Authors

Journals

citations
Cited by 7 publications
(3 citation statements)
references
References 24 publications
0
3
0
Order By: Relevance
“…Making complex ASR systems available was originally the intent of webASR, and, as such, ASR remains the main task in 3 newly developed systems covering 3 domains. All of them present a state-of-the-art speech transcription system, based on the latest research carried out at the University of Sheffield in topics such as Deep Neural Network (DNN) acoustic modelling [11,12,13,14], distant microphone recognition [15], adaptation to noisy environments [16,17,18], domain adaptation [19,20], Recurrent Neural Network (RNN) language modelling [21], Nbest re-ranking [22,23] and sentence-end detection [24,25].…”
Section: Transcription Systemsmentioning
confidence: 99%
“…Making complex ASR systems available was originally the intent of webASR, and, as such, ASR remains the main task in 3 newly developed systems covering 3 domains. All of them present a state-of-the-art speech transcription system, based on the latest research carried out at the University of Sheffield in topics such as Deep Neural Network (DNN) acoustic modelling [11,12,13,14], distant microphone recognition [15], adaptation to noisy environments [16,17,18], domain adaptation [19,20], Recurrent Neural Network (RNN) language modelling [21], Nbest re-ranking [22,23] and sentence-end detection [24,25].…”
Section: Transcription Systemsmentioning
confidence: 99%
“…In the cascade approach, an ASR system transcribes the input speech signal, and this is fed to a downstream MT system that carries out the translation. The provided input to the MT step can be the 1-best hypothesis, but also n-best lists (Ng et al, 2016) or even lattices (Matusov and Ney, 2011;Sperber et al, 2019). Additional techniques can also be used to improve the performance of the pipeline by better adapting the MT system to the expected input, such as training with transcribed text (Peitz et al, 2012) or chunking (Sperber et al, 2017).…”
Section: Introductionmentioning
confidence: 99%
“…The provided input to the MT step can be the 1-best hypothesis, but also n-best lists (Ng et al 2016) or even lattices (Matusov and Ney 2011;Sperber, Neubig, et al 2019). Additional techniques can also be used to improve the performance of the pipeline by better adapting the MT system to the expected input, such as training with transcribed text (Peitz et al 2012) or chunking (Sperber, Jan Niehues, and Waibel 2017).…”
Section: Discussionmentioning
confidence: 99%