Proceedings of the 2018 Conference of the North American Chapter Of the Association for Computational Linguistics: Hu 2018
DOI: 10.18653/v1/n18-2118
|View full text |Cite
|
Sign up to set email alerts
|

Slot-Gated Modeling for Joint Slot Filling and Intent Prediction

Abstract: Attention-based recurrent neural network models for joint intent detection and slot filling have achieved the state-of-the-art performance, while they have independent attention weights. Considering that slot and intent have the strong relationship, this paper proposes a slot gate that focuses on learning the relationship between intent and slot attention vectors in order to obtain better semantic frame results by the global optimization. The experiments show that our proposed model significantly improves sent… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

2
173
0
1

Year Published

2019
2019
2021
2021

Publication Types

Select...
4
3
1

Relationship

0
8

Authors

Journals

citations
Cited by 408 publications
(209 citation statements)
references
References 8 publications
2
173
0
1
Order By: Relevance
“…All our models were trained on a machine with 8 NVIDIA Tesla V100 GPUs, each with 16GB of memory. When using pretrained encoders, we leveraged gradual unfreezing to effectively tune the language Method Accuracy exact match intent Joint BiRNN [7] 80.70 92.60 Attention BiRNN [16] 78.90 91.10 Slot Gated Full Attention [5] 82.20 93.60 CapsuleNLU [36] 83 While training, the cross entropy loss function was modified with label smoothing with ϵ = 0.1. We used the Adam [12] optimizer with noam learning rate schedule [33], each adjusted differently for different datasets.…”
Section: Methodsmentioning
confidence: 99%
See 2 more Smart Citations
“…All our models were trained on a machine with 8 NVIDIA Tesla V100 GPUs, each with 16GB of memory. When using pretrained encoders, we leveraged gradual unfreezing to effectively tune the language Method Accuracy exact match intent Joint BiRNN [7] 80.70 92.60 Attention BiRNN [16] 78.90 91.10 Slot Gated Full Attention [5] 82.20 93.60 CapsuleNLU [36] 83 While training, the cross entropy loss function was modified with label smoothing with ϵ = 0.1. We used the Adam [12] optimizer with noam learning rate schedule [33], each adjusted differently for different datasets.…”
Section: Methodsmentioning
confidence: 99%
“…Source: play the song don't stop believin by journey Target: PlaySongIntent SongName( @pt r 3 @pt r 4 @pt r 5 )SongName ArtistName( @pt r 7 )ArtistName…”
Section: Query Formulationmentioning
confidence: 99%
See 1 more Smart Citation
“…End-to-end SLU architecture Train: (Utterances, Speakers) (115660, 77) Validation: (Utterances, Speakers) (3118, 10) Test: (Utterances, Speakers) (3793, 10) Unique Intents 31 Unique: (Actions, Objects, Locations)(6,14,4) …”
mentioning
confidence: 99%
“…Recently, joint intent classification and slot labeling SLU models based on recurrent neural networks has been shown to achieve state-of-the-art performance on benchmark datasets [Hakkani-Tür et al, 2016;Liu and Lane, 2016;Kim et al, 2017a;Goo et al, 2018;Wang et al, 2018]. However, these models often suffer from poor slot labeling accuracy when an utterance contain slots with large semantic variability, dissimilar to those encountered during training e.g.…”
Section: Introductionmentioning
confidence: 99%