Learning Robot Speech Models to Predict Speech Acts in HRI

In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communica...

Full description

Bibliographic Details
Main Authors: Arora Ankuj, Fiorino Humbert, Pellier Damien, Pesty Sylvie
Format: Article
Language:English
Published: De Gruyter 2016-08-01
Series:Paladyn: Journal of Behavioral Robotics
Subjects:
sat
Online Access:https://doi.org/10.1515/pjbr-2018-0015
id doaj-6fec45d8e45e42299331bcca3c6b2304
record_format Article
spelling doaj-6fec45d8e45e42299331bcca3c6b23042021-10-02T19:12:26ZengDe GruyterPaladyn: Journal of Behavioral Robotics2081-48362016-08-019129530610.1515/pjbr-2018-0015pjbr-2018-0015Learning Robot Speech Models to Predict Speech Acts in HRIArora Ankuj0Fiorino Humbert1Pellier Damien2Pesty Sylvie3Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceUniv. Grenoble Alpes, CNRS, Grenoble INP, LIG, 38000Grenoble, FranceIn order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communicate with humans without the need of “wizard” intervention. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. It views the entire exchange in the Automated Planning (AP) paradigm, representing the dialog sequences (speech acts) in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal. We then exploit intra-action and inter-action dependencies, encoding them in the form of constraints. We attempt to satisfy these constraints using aweighted maximum satisfiability model known as MAX-SAT, and convert the solution into a speech model. This model could have many uses, such as planning of fresh dialogs. In this study, the learnt model is used to predict speech acts in the dialog sequences using the sequence labeling (predicting future acts based on previously seen ones) capabilities of the LSTM (Long Short Term Memory) class of recurrent neural networks. Encouraging empirical results demonstrate the utility of this learnt model and its long term potential to facilitate autonomous behavioral planning of robots, an aspect to be explored in future works.https://doi.org/10.1515/pjbr-2018-0015human robot interactionautomated planningsatlstm
collection DOAJ
language English
format Article
sources DOAJ
author Arora Ankuj
Fiorino Humbert
Pellier Damien
Pesty Sylvie
spellingShingle Arora Ankuj
Fiorino Humbert
Pellier Damien
Pesty Sylvie
Learning Robot Speech Models to Predict Speech Acts in HRI
Paladyn: Journal of Behavioral Robotics
human robot interaction
automated planning
sat
lstm
author_facet Arora Ankuj
Fiorino Humbert
Pellier Damien
Pesty Sylvie
author_sort Arora Ankuj
title Learning Robot Speech Models to Predict Speech Acts in HRI
title_short Learning Robot Speech Models to Predict Speech Acts in HRI
title_full Learning Robot Speech Models to Predict Speech Acts in HRI
title_fullStr Learning Robot Speech Models to Predict Speech Acts in HRI
title_full_unstemmed Learning Robot Speech Models to Predict Speech Acts in HRI
title_sort learning robot speech models to predict speech acts in hri
publisher De Gruyter
series Paladyn: Journal of Behavioral Robotics
issn 2081-4836
publishDate 2016-08-01
description In order to be acceptable and able to “camouflage” into their physio-social context in the long run, robots need to be not just functional, but autonomously psycho-affective as well. This motivates a long term necessity of introducing behavioral autonomy in robots, so they can autonomously communicate with humans without the need of “wizard” intervention. This paper proposes a technique to learn robot speech models from human-robot dialog exchanges. It views the entire exchange in the Automated Planning (AP) paradigm, representing the dialog sequences (speech acts) in the form of action sequences that modify the state of the world upon execution, gradually propelling the state to a desired goal. We then exploit intra-action and inter-action dependencies, encoding them in the form of constraints. We attempt to satisfy these constraints using aweighted maximum satisfiability model known as MAX-SAT, and convert the solution into a speech model. This model could have many uses, such as planning of fresh dialogs. In this study, the learnt model is used to predict speech acts in the dialog sequences using the sequence labeling (predicting future acts based on previously seen ones) capabilities of the LSTM (Long Short Term Memory) class of recurrent neural networks. Encouraging empirical results demonstrate the utility of this learnt model and its long term potential to facilitate autonomous behavioral planning of robots, an aspect to be explored in future works.
topic human robot interaction
automated planning
sat
lstm
url https://doi.org/10.1515/pjbr-2018-0015
work_keys_str_mv AT aroraankuj learningrobotspeechmodelstopredictspeechactsinhri
AT fiorinohumbert learningrobotspeechmodelstopredictspeechactsinhri
AT pellierdamien learningrobotspeechmodelstopredictspeechactsinhri
AT pestysylvie learningrobotspeechmodelstopredictspeechactsinhri
_version_ 1716847816079310848