Learning Robot Speech Models to Predict Speech Acts in HRI
In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communica...
Main Authors: | , , , |
---|---|
Format: | Article |
Language: | English |
Published: |
De Gruyter
2016-08-01
|
Series: | Paladyn: Journal of Behavioral Robotics |
Subjects: | |
Online Access: | https://doi.org/10.1515/pjbr-2018-0015 |
id |
doaj-6fec45d8e45e42299331bcca3c6b2304 |
---|---|
record_format |
Article |
spelling |
doaj-6fec45d8e45e42299331bcca3c6b23042021-10-02T19:12:26ZengDe GruyterPaladyn: Journal of Behavioral Robotics2081-48362016-08-019129530610.1515/pjbr-2018-0015pjbr-2018-0015Learning Robot Speech Models to Predict Speech Acts in HRIArora Ankuj0Fiorino Humbert1Pellier Damien2Pesty Sylvie3Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceIn order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communicate with humans without the need of “wizard” intervention. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. It views the entire exchange in the Automated Planning (AP) paradigm, representing the dialog sequences (speech acts) in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal. We then exploit intra-action and inter-action dependencies, encoding them in the form of constraints. We attempt to satisfy these constraints using aweighted maximum satisfiability model known as MAX-SAT, and convert the solution into a speech model. This model could have many uses, such as planning of fresh dialogs. In this study, the learnt model is used to predict speech acts in the dialog sequences using the sequence labeling (predicting future acts based on previously seen ones) capabilities of the LSTM (Long Short Term Memory) class of recurrent neural networks. Encouraging empirical results demonstrate the utility of this learnt model and its long term potential to facilitate autonomous behavioral planning of robots, an aspect to be explored in future works.https://doi.org/10.1515/pjbr-2018-0015human robot interactionautomated planningsatlstm |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Arora Ankuj Fiorino Humbert Pellier Damien Pesty Sylvie |
spellingShingle |
Arora Ankuj Fiorino Humbert Pellier Damien Pesty Sylvie Learning Robot Speech Models to Predict Speech Acts in HRI Paladyn: Journal of Behavioral Robotics human robot interaction automated planning sat lstm |
author_facet |
Arora Ankuj Fiorino Humbert Pellier Damien Pesty Sylvie |
author_sort |
Arora Ankuj |
title |
Learning Robot Speech Models to Predict Speech Acts in HRI |
title_short |
Learning Robot Speech Models to Predict Speech Acts in HRI |
title_full |
Learning Robot Speech Models to Predict Speech Acts in HRI |
title_fullStr |
Learning Robot Speech Models to Predict Speech Acts in HRI |
title_full_unstemmed |
Learning Robot Speech Models to Predict Speech Acts in HRI |
title_sort |
learning robot speech models to predict speech acts in hri |
publisher |
De Gruyter |
series |
Paladyn: Journal of Behavioral Robotics |
issn |
2081-4836 |
publishDate |
2016-08-01 |
description |
In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communicate with humans without the need of “wizard” intervention. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. It views the entire exchange in the Automated Planning (AP) paradigm, representing the dialog sequences (speech acts) in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal. We then exploit intra-action and inter-action dependencies, encoding them in the form of constraints. We attempt to satisfy these constraints using aweighted maximum satisfiability model known as MAX-SAT, and convert the solution into a speech model. This model could have many uses, such as planning of fresh dialogs. In this study, the learnt model is used to predict speech acts in the dialog sequences using the sequence labeling (predicting future acts based on previously seen ones) capabilities of the LSTM (Long Short Term Memory) class of recurrent neural networks. Encouraging empirical results demonstrate the utility of this learnt model and its long term potential to facilitate autonomous behavioral planning of robots, an aspect to be explored in future works. |
topic |
human robot interaction automated planning sat lstm |
url |
https://doi.org/10.1515/pjbr-2018-0015 |
work_keys_str_mv |
AT aroraankuj learningrobotspeechmodelstopredictspeechactsinhri AT fiorinohumbert learningrobotspeechmodelstopredictspeechactsinhri AT pellierdamien learningrobotspeechmodelstopredictspeechactsinhri AT pestysylvie learningrobotspeechmodelstopredictspeechactsinhri |
_version_ |
1716847816079310848 |