Proceedings of the 20th ACM International Conference on Intelligent Virtual Agents 2020
DOI: 10.1145/3383652.3423877
|View full text |Cite
|
Sign up to set email alerts
|

A framework to co-optimize task and social dialogue policies using Reinforcement Learning

Abstract: One of the main challenges for conversational agents is to select the optimal dialogue policy based on the state of the interaction. This challenge becomes even harder when the conversational agent not only has to achieve a specific task, but also aims at building rapport. Although some work already tried to tackle this challenge using a Reinforcement Learning (RL) approach, they tend to consider one single optimal policy for all the users, regardless of their conversational goals. In this work, we describe a … Show more

Help me understand this report

Search citation statements

Order By: Relevance

Paper Sections

Select...
2
1
1
1

Citation Types

1
9
0

Year Published

2022
2022
2023
2023

Publication Types

Select...
3
2

Relationship

3
2

Authors

Journals

citations
Cited by 7 publications
(10 citation statements)
references
References 24 publications
1
9
0
Order By: Relevance
“…Researchers have already started to investigate rapport-building conversational systems in different contexts, and study how to combine rapport-building and task-oriented strategies during an interaction ( Pecune et al, 2013 ; Zhao et al, 2018b ; Pecune and Marsella, 2020 ). With Rea the virtual Real Estate Agent, researchers investigated how small-talk influenced the price users were ready to invest in a new house ( Cassell and Bickmore, 2003 ).…”
Section: Related Workmentioning
confidence: 99%
“…Researchers have already started to investigate rapport-building conversational systems in different contexts, and study how to combine rapport-building and task-oriented strategies during an interaction ( Pecune et al, 2013 ; Zhao et al, 2018b ; Pecune and Marsella, 2020 ). With Rea the virtual Real Estate Agent, researchers investigated how small-talk influenced the price users were ready to invest in a new house ( Cassell and Bickmore, 2003 ).…”
Section: Related Workmentioning
confidence: 99%
“…To optimize the dialog policy of our SIA, we rely on Reinforcement Learning. As in Pecune and Marsella (2020), Shi andYu (2018), andYu et al (2017), we build a user simulator to approximate human users' behaviors and generate enough synthetic interactions to train the agent. The use of a simulated user allows to pretrain the model and avoid the "cold start problem" where the model behaves badly in the first interactions with human users.…”
Section: Our Approachmentioning
confidence: 99%
“…As shown by Pecune and Marsella (2020), people's conversational goals and preferences influence the perceived quality of an interaction. For instance, people only focusing on the task might quickly get disengaged if their interlocutor uses social conversational strategies.…”
Section: User's Preferences Conversational Preferencesmentioning
confidence: 99%
See 1 more Smart Citation
“…To simulate users interacting with an information-giving conversational agent, we built the following user simulator. Conversational preferences We rely on [8] to model SU's conversational preferences. Each SU has one of the following preferences: Action selection Each turn, the SU outputs a dialog act and a list of non-verbal behaviors.…”
Section: User Simulatormentioning
confidence: 99%