Interspeech 2018 2018
DOI: 10.21437/interspeech.2018-1379
|View full text |Cite
|
Sign up to set email alerts
|

A Deep Reinforcement Learning Based Multimodal Coaching Model (DCM) for Slot Filling in Spoken Language Understanding(SLU)

Abstract: In this paper, a deep reinforcement learning(DRL) based multimodal coaching model (DCM) for slot filling task in SLU is proposed. The DCM takes advantage of a DRL based model as a coach of the system to learn the wrong labeled slots with/without user's feedback, hence may further improve the performance of an SLU system. This DCM model is an improved model of the deep reinforcement learning based augmented tagging model as introduced in [1], by using a better DRL model with different rewards and adding in a us… Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

0
7
0

Year Published

2019
2019
2022
2022

Publication Types

Select...
6
1

Relationship

3
4

Authors

Journals

citations
Cited by 10 publications
(7 citation statements)
references
References 26 publications
0
7
0
Order By: Relevance
“…The AC-NCR model is trained based on deep reinforcement learning (DRL) technique. Deep reinforcement learning has been widely used in a variety of NLP and machine learning tasks [5,6,7,8]. The input to the model is a state s t , and its output is an action a t .…”
Section: Ac-ncr Model Designmentioning
confidence: 99%
“…The AC-NCR model is trained based on deep reinforcement learning (DRL) technique. Deep reinforcement learning has been widely used in a variety of NLP and machine learning tasks [5,6,7,8]. The input to the model is a state s t , and its output is an action a t .…”
Section: Ac-ncr Model Designmentioning
confidence: 99%
“…This encoding was then used as an input for Slot Labeling, thereby capturing the global information from the input. Reference [17] proposed the use of a Deep Reinforcement Learning (DRL) based multimodal Coaching Model (DCM), where authors used a new reward mechanism for the RL based Slot Tagging. Even without any feedback from the user, this model was capable of understanding whatever has been wrongly labelled.…”
Section: B Slot Labellingmentioning
confidence: 99%
“…In previous years, several studies were carried out to improve and evaluate the performance of NLU tasks. Former works evaluated the impact of using different techniques to improve slot filling [ 4 , 5 , 6 ] and intent detection [ 7 , 8 ]. Interestingly, research in joint NLU (jointly learning both intent detection and slot filling) achieved better results in both tasks [ 9 , 10 ].…”
Section: Introductionmentioning
confidence: 99%