Proceedings of the 59th Annual Meeting of the Association for Computational Linguistics and the 11th International Joint Confer 2021
DOI: 10.18653/v1/2021.acl-demo.7
|View full text |Cite
|
Sign up to set email alerts
|

NeurST: Neural Speech Translation Toolkit

Abstract: NeurST is an open-source toolkit for neural speech translation. The toolkit mainly focuses on end-to-end speech translation, which is easy to use, modify, and extend to advanced speech translation research and products. NeurST aims at facilitating the speech translation research for NLP researchers and building reliable benchmarks for this field. It provides step-by-step recipes for feature extraction, data preprocessing, distributed training, and evaluation. In this paper, we will introduce the framework desi… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

0
14
0

Year Published

2021
2021
2024
2024

Publication Types

Select...
5
1
1

Relationship

2
5

Authors

Journals

citations
Cited by 16 publications
(14 citation statements)
references
References 28 publications
0
14
0
Order By: Relevance
“…We conduct all our experiments using NeurST (Zhao et al, 2020) and report results for the submitted speech translation tasks in this section.…”
Section: Resultsmentioning
confidence: 99%
See 1 more Smart Citation
“…We conduct all our experiments using NeurST (Zhao et al, 2020) and report results for the submitted speech translation tasks in this section.…”
Section: Resultsmentioning
confidence: 99%
“…The results of our end-to-end solutions are presented in line 8-20, where line 8 is a benchmark model (Zhao et al, 2020) trained on the MuST-C dataset only. With the growth of model capacity (256d→768d) and data augmentation, we obtain 6.2 BLEU improvement on the tst-COMMON over the benchmark (line 8).…”
Section: Offline Speech Translationmentioning
confidence: 99%
“…In this section, we conduct ST experiments with speech transformer models and SSL-Transformer models. All models are implemented using NeurST [5].…”
Section: Methodsmentioning
confidence: 99%
“…In detail, we lowercase the English transcriptions, remove all punctuations and use SentencePiece 4 with a vocabulary of 15,000. For Chinese text, we first segment sentences by Jieba 5 , and then apply Byte-Pair Encoding (BPE) 6 [25] with 32,000 merge operations. German texts are first tokenized using Moses tokenizer, followed by BPE with 32,000 merge operations.…”
Section: Setupsmentioning
confidence: 99%
See 1 more Smart Citation