2021
DOI: 10.48550/arxiv.2107.02268
|View full text |Cite
Preprint
|
Sign up to set email alerts
|

Instant One-Shot Word-Learning for Context-Specific Neural Sequence-to-Sequence Speech Recognition

Abstract: Neural sequence-to-sequence systems deliver state-of-the-art performance for automatic speech recognition (ASR). When using appropriate modeling units, e.g., byte-pair encoded characters, these systems are in principal open vocabulary systems. In practice, however, they often fail to recognize words not seen during training, e.g., named entities, numbers or technical terms. To alleviate this problem we supplement an end-to-end ASR system with a word/phrase memory and a mechanism to access this memory to recogn… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1

Citation Types

0
1
0

Year Published

2021
2021
2021
2021

Publication Types

Select...
1

Relationship

0
1

Authors

Journals

citations
Cited by 1 publication
(1 citation statement)
references
References 20 publications
0
1
0
Order By: Relevance
“…In this way, the network output conditions not only on the speech signal and previous label sequence but also on the contextual phrase list. It was further extended into a two-step memory enhanced model in [190]. It is common that a contextual phrase list contains rare words, especially names which are not observed during training.…”
Section: C) Customizationmentioning
confidence: 99%
“…In this way, the network output conditions not only on the speech signal and previous label sequence but also on the contextual phrase list. It was further extended into a two-step memory enhanced model in [190]. It is common that a contextual phrase list contains rare words, especially names which are not observed during training.…”
Section: C) Customizationmentioning
confidence: 99%