Active Vision for Robot Manipulators Using the Free Energy Principle

Occlusions, restricted field of view and limited resolution all constrain a robot's ability to sense its environment from a single observation. In these cases, the robot first needs to actively query multiple observations and accumulate information before it can complete a task. In this paper,...

Full description

Bibliographic Details
Main Authors: Toon Van de Maele, Tim Verbelen, Ozan Çatal, Cedric De Boom, Bart Dhoedt
Format: Article
Language:English
Published: Frontiers Media S.A. 2021-03-01
Series:Frontiers in Neurorobotics
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fnbot.2021.642780/full
id doaj-5f800eb7302e4f14b206d1988b39c01c
record_format Article
spelling doaj-5f800eb7302e4f14b206d1988b39c01c2021-03-05T06:01:35ZengFrontiers Media S.A.Frontiers in Neurorobotics1662-52182021-03-011510.3389/fnbot.2021.642780642780Active Vision for Robot Manipulators Using the Free Energy PrincipleToon Van de MaeleTim VerbelenOzan ÇatalCedric De BoomBart DhoedtOcclusions, restricted field of view and limited resolution all constrain a robot's ability to sense its environment from a single observation. In these cases, the robot first needs to actively query multiple observations and accumulate information before it can complete a task. In this paper, we cast this problem of active vision as active inference, which states that an intelligent agent maintains a generative model of its environment and acts in order to minimize its surprise, or expected free energy according to this model. We apply this to an object-reaching task for a 7-DOF robotic manipulator with an in-hand camera to scan the workspace. A novel generative model using deep neural networks is proposed that is able to fuse multiple views into an abstract representation and is trained from data by minimizing variational free energy. We validate our approach experimentally for a reaching task in simulation in which a robotic agent starts without any knowledge about its workspace. Each step, the next view pose is chosen by evaluating the expected free energy. We find that by minimizing the expected free energy, exploratory behavior emerges when the target object to reach is not in view, and the end effector is moved to the correct reach position once the target is located. Similar to an owl scavenging for prey, the robot naturally prefers higher ground for exploring, approaching its target once located.https://www.frontiersin.org/articles/10.3389/fnbot.2021.642780/fullactive visionactive inferencedeep learninggenerative modelingrobotics
collection DOAJ
language English
format Article
sources DOAJ
author Toon Van de Maele
Tim Verbelen
Ozan Çatal
Cedric De Boom
Bart Dhoedt
spellingShingle Toon Van de Maele
Tim Verbelen
Ozan Çatal
Cedric De Boom
Bart Dhoedt
Active Vision for Robot Manipulators Using the Free Energy Principle
Frontiers in Neurorobotics
active vision
active inference
deep learning
generative modeling
robotics
author_facet Toon Van de Maele
Tim Verbelen
Ozan Çatal
Cedric De Boom
Bart Dhoedt
author_sort Toon Van de Maele
title Active Vision for Robot Manipulators Using the Free Energy Principle
title_short Active Vision for Robot Manipulators Using the Free Energy Principle
title_full Active Vision for Robot Manipulators Using the Free Energy Principle
title_fullStr Active Vision for Robot Manipulators Using the Free Energy Principle
title_full_unstemmed Active Vision for Robot Manipulators Using the Free Energy Principle
title_sort active vision for robot manipulators using the free energy principle
publisher Frontiers Media S.A.
series Frontiers in Neurorobotics
issn 1662-5218
publishDate 2021-03-01
description Occlusions, restricted field of view and limited resolution all constrain a robot's ability to sense its environment from a single observation. In these cases, the robot first needs to actively query multiple observations and accumulate information before it can complete a task. In this paper, we cast this problem of active vision as active inference, which states that an intelligent agent maintains a generative model of its environment and acts in order to minimize its surprise, or expected free energy according to this model. We apply this to an object-reaching task for a 7-DOF robotic manipulator with an in-hand camera to scan the workspace. A novel generative model using deep neural networks is proposed that is able to fuse multiple views into an abstract representation and is trained from data by minimizing variational free energy. We validate our approach experimentally for a reaching task in simulation in which a robotic agent starts without any knowledge about its workspace. Each step, the next view pose is chosen by evaluating the expected free energy. We find that by minimizing the expected free energy, exploratory behavior emerges when the target object to reach is not in view, and the end effector is moved to the correct reach position once the target is located. Similar to an owl scavenging for prey, the robot naturally prefers higher ground for exploring, approaching its target once located.
topic active vision
active inference
deep learning
generative modeling
robotics
url https://www.frontiersin.org/articles/10.3389/fnbot.2021.642780/full
work_keys_str_mv AT toonvandemaele activevisionforrobotmanipulatorsusingthefreeenergyprinciple
AT timverbelen activevisionforrobotmanipulatorsusingthefreeenergyprinciple
AT ozancatal activevisionforrobotmanipulatorsusingthefreeenergyprinciple
AT cedricdeboom activevisionforrobotmanipulatorsusingthefreeenergyprinciple
AT bartdhoedt activevisionforrobotmanipulatorsusingthefreeenergyprinciple
_version_ 1724230789944573952