2018
DOI: 10.1007/978-3-030-04212-7_57
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Learning Based Multi-task Ensemble Model for Intent Detection and Slot Filling in Spoken Language Understanding

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

0
10
0

Year Published

2020
2020
2024
2024

Publication Types

Select...
4
2
1

Relationship

0
7

Authors

Journals

citations
Cited by 12 publications
(10 citation statements)
references
References 26 publications
0
10
0
Order By: Relevance
“…A special token is added to encapsulate the whole utterance for use in intent classification [Hakkani-Tür et al 2016]; • a Bi-LSTM encoder decoder but with separate losses for intent and slot prediction ( [Zheng et al 2017]; • rather than seq2seq [Kim et al 2017] perform a global slot prediction (learning the joint distribution) from a matrix of the hidden states to a matrix of slot tag probabilities for each word, intent is predicted from a sum of hidden states; • [Wen et al 2018] propose to use both a hierarchical (multilayer) and a contextual (BiLSTM or LSTM) approach, investigating various combinations and using differing layers for intent and slot prediction; • an ensemble using both BiLSTM and BiGRU fed to separate MLPs whose outputs are fused then projected and a softmax applied to predict intent and slots concurrently is proposed by [Firdaus et al 2018a].…”
Section: Recurrent Neural Networkmentioning
confidence: 99%
See 3 more Smart Citations
“…A special token is added to encapsulate the whole utterance for use in intent classification [Hakkani-Tür et al 2016]; • a Bi-LSTM encoder decoder but with separate losses for intent and slot prediction ( [Zheng et al 2017]; • rather than seq2seq [Kim et al 2017] perform a global slot prediction (learning the joint distribution) from a matrix of the hidden states to a matrix of slot tag probabilities for each word, intent is predicted from a sum of hidden states; • [Wen et al 2018] propose to use both a hierarchical (multilayer) and a contextual (BiLSTM or LSTM) approach, investigating various combinations and using differing layers for intent and slot prediction; • an ensemble using both BiLSTM and BiGRU fed to separate MLPs whose outputs are fused then projected and a softmax applied to predict intent and slots concurrently is proposed by [Firdaus et al 2018a].…”
Section: Recurrent Neural Networkmentioning
confidence: 99%
“…[Staliūnaitė and Iacobacci 2020] extended this work to a multi-task setting with extra mid-level capsules for NER and POS labels, with mixed results. [Wen et al 2018] Using hierarchy and context Two layer (Bi)LSTM [Wang et al 2018c] Capturing local semantic information CNN, BiLSTM encoder decoder [Firdaus et al 2018a] Domain dependence Ensemble model, GRU Slow training time Progressive multi-task model using user information [Li et al 2018a] Correlation of different tasks Multi-task model incl. POS tag [Li et al 2018b] Sharing semantic information Self-attention Tagging strategy Token tags include intent and slot [Zhang et al 2019a] Hierarchical structure Capsule network with rerouting (feedback) Spatial (context) and serial (order) information Encoder-decoder, CNN [Wang et al 2018b] slot2intent and intent2slot Bi-directional architecture [Siddhant et al 2019] Unsupervised learning ELMo on unused utterances, BiLSTM Use sequence labelling output for intent Cross attention, BiLSTM, CRF Hierarchical vector approach Learn vectors representing elements of frame [Jung et al 2018] Model relationship between text and its semantic frame…”
Section: Hierarchical Modelsmentioning
confidence: 99%
See 2 more Smart Citations
“…A classical example of slot filling with the sentence Show me the flights from Boston to New York today is shown in Table 1. While these tasks were previously treated separately, recent research have shown that joint models capable of answering to both tasks performed better [11,39,40].…”
Section: Intelligent Assistantsmentioning
confidence: 99%