2010
DOI: 10.1016/j.csl.2009.07.001
|View full text |Cite
|
Sign up to set email alerts
|

Evaluation of a hierarchical reinforcement learning spoken dialogue system

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
3
1
1

Citation Types

1
61
0

Year Published

2010
2010
2022
2022

Publication Types

Select...
6
2
1

Relationship

4
5

Authors

Journals

citations
Cited by 61 publications
(62 citation statements)
references
References 46 publications
1
61
0
Order By: Relevance
“…There has been some prior research on using machine learning techniques in human-robot interaction, though not for multi-user interactions. One such approach involved modelling the interaction as a Semi-Markov Decision Process (SMDP) and using Hierarchical Reinforcement Learning (HRL) for optimising decision-making [Cuayáhuitl and Dethlefs 2011;Cuayáhuitl et al 2010]. An alternative approach to using multiple policies within a dialogue manager has also incorporated POMDP models, but still focused on single-user interactions [Lison 2011].…”
Section: Discussionmentioning
confidence: 99%
“…There has been some prior research on using machine learning techniques in human-robot interaction, though not for multi-user interactions. One such approach involved modelling the interaction as a Semi-Markov Decision Process (SMDP) and using Hierarchical Reinforcement Learning (HRL) for optimising decision-making [Cuayáhuitl and Dethlefs 2011;Cuayáhuitl et al 2010]. An alternative approach to using multiple policies within a dialogue manager has also incorporated POMDP models, but still focused on single-user interactions [Lison 2011].…”
Section: Discussionmentioning
confidence: 99%
“…[15] used rule-based classifiers for predicting user intentions, which are executed using a Hierarchical Task Network incorporating expert knowledge. Trainable multi-domain dialogue systems using traditional reinforcement learning include [16], [17], [18], [19]. These systems use a modest amount of features, and in contrast to neural-based systems, they require manual feature engineering.…”
Section: Related Work and Discussionmentioning
confidence: 99%
“…It remains to be demonstrated how far one can go with such an approach. Future work includes to (a) compare different model architectures, training parameters and reward functions; (b) extend or improve the abilities of the proposed dialogue system; (c) train deep learning agents in other (larger scale) domains [7,8,9]; (d) evaluate end-to-end systems with real users; (e) compare or combine different types of neural nets [10]; and (e) perform fast learning based on parallel computing. Table 1 Example dialogue using the policy from Fig.2, where states are numerical representations of the last system and noisy user inputs, actions are dialogue acts, and user resposes are in brackets…”
Section: Discussionmentioning
confidence: 99%