2019 IEEE Automatic Speech Recognition and Understanding Workshop (ASRU) 2019
DOI: 10.1109/asru46091.2019.9003947
|View full text |Cite
|
Sign up to set email alerts
|

Spoken Language Identification Using Bidirectional LSTM Based LID Sequential Senones

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
2
2
2

Relationship

1
5

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 16 publications
0
7
0
Order By: Relevance
“…The architecture of the embedding extractor is shown in Fig. 2, which is motivated by the network in [15] and [16]. It contains two bidirectional long short-term memory (BLSTM) layers with 256 and 64 nodes respectively in first and second layer.…”
Section: Feature Extractor Block For Obtaining Fixed-length U-vectormentioning
confidence: 99%
See 2 more Smart Citations
“…The architecture of the embedding extractor is shown in Fig. 2, which is motivated by the network in [15] and [16]. It contains two bidirectional long short-term memory (BLSTM) layers with 256 and 64 nodes respectively in first and second layer.…”
Section: Feature Extractor Block For Obtaining Fixed-length U-vectormentioning
confidence: 99%
“…It contains two bidirectional long short-term memory (BLSTM) layers with 256 and 64 nodes respectively in first and second layer. These BLSTM layers analyze the input sequence of BNFs by dividing it into fixed-length chunks (with 50% overlap between successive chunks) to generate LID-seq-senones [15]. These LID-seq-senones are nothing but the activation obtained at the output of second BLSTM layer for each chunk of BNF vectors [15].…”
Section: Feature Extractor Block For Obtaining Fixed-length U-vectormentioning
confidence: 99%
See 1 more Smart Citation
“…The authors would like to thank Hugo Jair Escalante, Isabelle Guyon and Qiang Yang for guidance as advisors. The platform, automl.ai 6 , is built based on Codalab 7 , an web-based platform for machine learning competitions [26].…”
Section: Acknowledgementsmentioning
confidence: 99%
“…In the past few decades, machine learning, especially deep learning, has achieved remarkable breakthroughs in a wide range of speech tasks, e.g., speech recognition [1,2], speaker verification [3,4,5], language identification [6,7] and emotion classification [8,9]. Each speech task has its own specific techniques in achieving the state-of-the-art results [3,6,8,10,11,12], which require efforts of a large number of experts. Thus, it is very difficult to switch between different speech tasks without human efforts.…”
Section: Introductionmentioning
confidence: 99%