Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos

Motion capture of unrestrained moving animals is a major analytic tool in neuroethology and behavioral physiology. At present, several motion capture methodologies have been developed, all of which have particular limitations regarding experimental application. Whereas marker-based motion capture sy...

Full description

Bibliographic Details
Main Authors: Ilja Arent, Florian P. Schmidt, Mario Botsch, Volker Dürr
Format: Article
Language:English
Published: Frontiers Media S.A. 2021-04-01
Series:Frontiers in Behavioral Neuroscience
Subjects:
Online Access:https://www.frontiersin.org/articles/10.3389/fnbeh.2021.637806/full
id doaj-77c5659332e04e6896c96f95c322c030
record_format Article
spelling doaj-77c5659332e04e6896c96f95c322c0302021-04-22T06:21:20ZengFrontiers Media S.A.Frontiers in Behavioral Neuroscience1662-51532021-04-011510.3389/fnbeh.2021.637806637806Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic VideosIlja Arent0Florian P. Schmidt1Florian P. Schmidt2Mario Botsch3Mario Botsch4Volker Dürr5Volker Dürr6Biological Cybernetics, Faculty of Biology, Bielefeld University, Bielefeld, GermanyBiological Cybernetics, Faculty of Biology, Bielefeld University, Bielefeld, GermanyCenter for Cognitive Interaction Technology, Bielefeld University, Bielefeld, GermanyCenter for Cognitive Interaction Technology, Bielefeld University, Bielefeld, GermanyComputer Graphics, TU Dortmund University, Dortmund, GermanyBiological Cybernetics, Faculty of Biology, Bielefeld University, Bielefeld, GermanyCenter for Cognitive Interaction Technology, Bielefeld University, Bielefeld, GermanyMotion capture of unrestrained moving animals is a major analytic tool in neuroethology and behavioral physiology. At present, several motion capture methodologies have been developed, all of which have particular limitations regarding experimental application. Whereas marker-based motion capture systems are very robust and easily adjusted to suit different setups, tracked species, or body parts, they cannot be applied in experimental situations where markers obstruct the natural behavior (e.g., when tracking delicate, elastic, and/or sensitive body structures). On the other hand, marker-less motion capture systems typically require setup- and animal-specific adjustments, for example by means of tailored image processing, decision heuristics, and/or machine learning of specific sample data. Among the latter, deep-learning approaches have become very popular because of their applicability to virtually any sample of video data. Nevertheless, concise evaluation of their training requirements has rarely been done, particularly with regard to the transfer of trained networks from one application to another. To address this issue, the present study uses insect locomotion as a showcase example for systematic evaluation of variation and augmentation of the training data. For that, we use artificially generated video sequences with known combinations of observed, real animal postures and randomized body position, orientation, and size. Moreover, we evaluate the generalization ability of networks that have been pre-trained on synthetic videos to video recordings of real walking insects, and estimate the benefit in terms of reduced requirement for manual annotation. We show that tracking performance is affected only little by scaling factors ranging from 0.5 to 1.5. As expected from convolutional networks, the translation of the animal has no effect. On the other hand, we show that sufficient variation of rotation in the training data is essential for performance, and make concise suggestions about how much variation is required. Our results on transfer from synthetic to real videos show that pre-training reduces the amount of necessary manual annotation by about 50%.https://www.frontiersin.org/articles/10.3389/fnbeh.2021.637806/fullinsect locomotionmachine learningbehavioral analysismarker-less motion capturedeep neural networkmotion tracking
collection DOAJ
language English
format Article
sources DOAJ
author Ilja Arent
Florian P. Schmidt
Florian P. Schmidt
Mario Botsch
Mario Botsch
Volker Dürr
Volker Dürr
spellingShingle Ilja Arent
Florian P. Schmidt
Florian P. Schmidt
Mario Botsch
Mario Botsch
Volker Dürr
Volker Dürr
Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
Frontiers in Behavioral Neuroscience
insect locomotion
machine learning
behavioral analysis
marker-less motion capture
deep neural network
motion tracking
author_facet Ilja Arent
Florian P. Schmidt
Florian P. Schmidt
Mario Botsch
Mario Botsch
Volker Dürr
Volker Dürr
author_sort Ilja Arent
title Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
title_short Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
title_full Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
title_fullStr Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
title_full_unstemmed Marker-Less Motion Capture of Insect Locomotion With Deep Neural Networks Pre-trained on Synthetic Videos
title_sort marker-less motion capture of insect locomotion with deep neural networks pre-trained on synthetic videos
publisher Frontiers Media S.A.
series Frontiers in Behavioral Neuroscience
issn 1662-5153
publishDate 2021-04-01
description Motion capture of unrestrained moving animals is a major analytic tool in neuroethology and behavioral physiology. At present, several motion capture methodologies have been developed, all of which have particular limitations regarding experimental application. Whereas marker-based motion capture systems are very robust and easily adjusted to suit different setups, tracked species, or body parts, they cannot be applied in experimental situations where markers obstruct the natural behavior (e.g., when tracking delicate, elastic, and/or sensitive body structures). On the other hand, marker-less motion capture systems typically require setup- and animal-specific adjustments, for example by means of tailored image processing, decision heuristics, and/or machine learning of specific sample data. Among the latter, deep-learning approaches have become very popular because of their applicability to virtually any sample of video data. Nevertheless, concise evaluation of their training requirements has rarely been done, particularly with regard to the transfer of trained networks from one application to another. To address this issue, the present study uses insect locomotion as a showcase example for systematic evaluation of variation and augmentation of the training data. For that, we use artificially generated video sequences with known combinations of observed, real animal postures and randomized body position, orientation, and size. Moreover, we evaluate the generalization ability of networks that have been pre-trained on synthetic videos to video recordings of real walking insects, and estimate the benefit in terms of reduced requirement for manual annotation. We show that tracking performance is affected only little by scaling factors ranging from 0.5 to 1.5. As expected from convolutional networks, the translation of the animal has no effect. On the other hand, we show that sufficient variation of rotation in the training data is essential for performance, and make concise suggestions about how much variation is required. Our results on transfer from synthetic to real videos show that pre-training reduces the amount of necessary manual annotation by about 50%.
topic insect locomotion
machine learning
behavioral analysis
marker-less motion capture
deep neural network
motion tracking
url https://www.frontiersin.org/articles/10.3389/fnbeh.2021.637806/full
work_keys_str_mv AT iljaarent markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT florianpschmidt markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT florianpschmidt markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT mariobotsch markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT mariobotsch markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT volkerdurr markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
AT volkerdurr markerlessmotioncaptureofinsectlocomotionwithdeepneuralnetworkspretrainedonsyntheticvideos
_version_ 1721515008253231104