2018
DOI: 10.1007/978-3-030-04182-3_32
|View full text |Cite
|
Sign up to set email alerts
|

Reinforcement Learning Based Dialogue Management Strategy

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2

Citation Types

0
2
0

Year Published

2020
2020
2023
2023

Publication Types

Select...
4
1
1

Relationship

1
5

Authors

Journals

citations
Cited by 6 publications
(2 citation statements)
references
References 14 publications
0
2
0
Order By: Relevance
“…There is also a variant called DDQN with Prioritized Experience Replay (DDQN-PER) where it shows significant improvement with an average dialogue length of 1342.3, -13.51 average episodic reward, and only 52.56 hours of training time. Additionally, there is also a test where DDQN-PER is not paired with any algorithms and it has maintained a good balance between performance and training time [28].…”
Section: Algorithm Suited Use Case Commonly Used In Advantagesmentioning
confidence: 99%
See 1 more Smart Citation
“…There is also a variant called DDQN with Prioritized Experience Replay (DDQN-PER) where it shows significant improvement with an average dialogue length of 1342.3, -13.51 average episodic reward, and only 52.56 hours of training time. Additionally, there is also a test where DDQN-PER is not paired with any algorithms and it has maintained a good balance between performance and training time [28].…”
Section: Algorithm Suited Use Case Commonly Used In Advantagesmentioning
confidence: 99%
“…Standard Deep Q-Networks (DQN) also has showcased reasonably good performance with an average dialogue length of 673.45 and average episodic reward of -6.89. The training time needed for this variant is also relatively efficient clocking in only around 71.97 hours [28]. Nonetheless, the preferred variant can depend on the project scope and problems statements given.…”
Section: Algorithm Suited Use Case Commonly Used In Advantagesmentioning
confidence: 99%