Proceedings of the 17th Annual Meeting of the Special Interest Group on Discourse and Dialogue 2016
DOI: 10.18653/v1/w16-3643
|View full text |Cite
|
Sign up to set email alerts
|

Training an adaptive dialogue policy for interactive learning of visually grounded word meanings

Abstract: We present a multi-modal dialogue system for interactive learning of perceptually grounded word meanings from a human tutor. The system integrates an incremental, semantic parsing/generation framework -Dynamic Syntax and Type Theory with Records (DS-TTR) -with a set of visual classifiers that are learned throughout the interaction and which ground the meaning representations that it produces. We use this system in interaction with a simulated human tutor to study the effects of different dialogue policies and … Show more

Help me understand this report
View preprint versions

Search citation statements

Order By: Relevance

Paper Sections

Select...
1
1
1
1

Citation Types

1
33
0

Year Published

2016
2016
2021
2021

Publication Types

Select...
4
2
2

Relationship

5
3

Authors

Journals

citations
Cited by 17 publications
(34 citation statements)
references
References 28 publications
1
33
0
Order By: Relevance
“…Accuracy (%) KLD Utterance-level 77.98 0.2338 Act-level 84.96 0.188 In order to demonstrate how the BURCHAK corpus can be used, we train and evaluate a prototype interactive learning agent using Reinforcement Learning (RL) on the collected data. We follow previous task and experiment settings (see (Yu et al, 2016b;Yu et al, 2016c)) to compare the learned RL-based agent with a rule-based agent with the best performance from previous work. Instead of using hand-crafted dialogue examples as before, here we train the RL agent in interaction with the user simulation, itself trained from the BURCHAK data as above.…”
Section: Simulationmentioning
confidence: 99%
“…Accuracy (%) KLD Utterance-level 77.98 0.2338 Act-level 84.96 0.188 In order to demonstrate how the BURCHAK corpus can be used, we train and evaluate a prototype interactive learning agent using Reinforcement Learning (RL) on the collected data. We follow previous task and experiment settings (see (Yu et al, 2016b;Yu et al, 2016c)) to compare the learned RL-based agent with a rule-based agent with the best performance from previous work. Instead of using hand-crafted dialogue examples as before, here we train the RL agent in interaction with the user simulation, itself trained from the BURCHAK data as above.…”
Section: Simulationmentioning
confidence: 99%
“…Our work is similar in spirit to e.g. (Roy, 2002;Skocaj et al, 2011) but advances it in several aspects (Yu et al, 2016).…”
Section: Introductionmentioning
confidence: 78%
“…Following previous work (Yu et al, 2016), here we use a positive confidence threshold, which determines when the agent believes its own predictions. For instance, the learner can ask either polar or WH-questions about an attribute if its confidence score is higher than a certain threshold; otherwise, there should be no interaction about that attribute.…”
Section: When To Learn: Adaptive Confidence Thresholdmentioning
confidence: 99%
“…On the other hand, other models assume a much more explicit connection between symbols (either words or predicate symbols of some logical language) and perceptions (Kennington and Schlangen, 2015;Yu et al, 2016c;Skocaj et al, 2016;Dobnik et al, 2014;Matuszek et al, 2014). In this line of work, representations are both compositional and transparent, with their constituent atomic parts grounded individually in perceptual classifiers.…”
Section: Related Workmentioning
confidence: 99%