Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning

Machine-learning models of music often exist outside the worlds of musical performance practice and abstracted from the physical gestures of musicians. In this work, we consider how a recurrent neural network (RNN) model of simple music gestures may be integrated into a physical instrument so that p...

Full description

Bibliographic Details
Main Authors: Charles Patrick Martin, Kyrre Glette, Tønnes Frostad Nygaard, Jim Torresen
Format: Article
Language:English
Published: Frontiers Media S.A. 2020-03-01
Series:Frontiers in Artificial Intelligence
Subjects:
Online Access:https://www.frontiersin.org/article/10.3389/frai.2020.00006/full
id doaj-5f69ff51e94849369e032d59eb124f1a
record_format Article
spelling doaj-5f69ff51e94849369e032d59eb124f1a2020-11-25T02:31:42ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122020-03-01310.3389/frai.2020.00006508777Understanding Musical Predictions With an Embodied Interface for Musical Machine LearningCharles Patrick Martin0Charles Patrick Martin1Charles Patrick Martin2Kyrre Glette3Kyrre Glette4Tønnes Frostad Nygaard5Jim Torresen6Jim Torresen7Research School of Computer Science, Australian National University, Canberra, ACT, AustraliaDepartment of Informatics, University of Oslo, Oslo, NorwayRITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, NorwayDepartment of Informatics, University of Oslo, Oslo, NorwayRITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, NorwayDepartment of Informatics, University of Oslo, Oslo, NorwayDepartment of Informatics, University of Oslo, Oslo, NorwayRITMO Centre for Interdisciplinary Studies in Rhythm, Time and Motion, University of Oslo, Oslo, NorwayMachine-learning models of music often exist outside the worlds of musical performance practice and abstracted from the physical gestures of musicians. In this work, we consider how a recurrent neural network (RNN) model of simple music gestures may be integrated into a physical instrument so that predictions are sonically and physically entwined with the performer's actions. We introduce EMPI, an embodied musical prediction interface that simplifies musical interaction and prediction to just one dimension of continuous input and output. The predictive model is a mixture density RNN trained to estimate the performer's next physical input action and the time at which this will occur. Predictions are represented sonically through synthesized audio, and physically with a motorized output indicator. We use EMPI to investigate how performers understand and exploit different predictive models to make music through a controlled study of performances with different models and levels of physical feedback. We show that while performers often favor a model trained on human-sourced data, they find different musical affordances in models trained on synthetic, and even random, data. Physical representation of predictions seemed to affect the length of performances. This work contributes new understandings of how musicians use generative ML models in real-time performance backed up by experimental evidence. We argue that a constrained musical interface can expose the affordances of embodied predictive interactions.https://www.frontiersin.org/article/10.3389/frai.2020.00006/fullmusical performanceinterfacemixture density network (MDN)recurrent neural network (RNN)creativitypredictive interaction
collection DOAJ
language English
format Article
sources DOAJ
author Charles Patrick Martin
Charles Patrick Martin
Charles Patrick Martin
Kyrre Glette
Kyrre Glette
Tønnes Frostad Nygaard
Jim Torresen
Jim Torresen
spellingShingle Charles Patrick Martin
Charles Patrick Martin
Charles Patrick Martin
Kyrre Glette
Kyrre Glette
Tønnes Frostad Nygaard
Jim Torresen
Jim Torresen
Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
Frontiers in Artificial Intelligence
musical performance
interface
mixture density network (MDN)
recurrent neural network (RNN)
creativity
predictive interaction
author_facet Charles Patrick Martin
Charles Patrick Martin
Charles Patrick Martin
Kyrre Glette
Kyrre Glette
Tønnes Frostad Nygaard
Jim Torresen
Jim Torresen
author_sort Charles Patrick Martin
title Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
title_short Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
title_full Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
title_fullStr Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
title_full_unstemmed Understanding Musical Predictions With an Embodied Interface for Musical Machine Learning
title_sort understanding musical predictions with an embodied interface for musical machine learning
publisher Frontiers Media S.A.
series Frontiers in Artificial Intelligence
issn 2624-8212
publishDate 2020-03-01
description Machine-learning models of music often exist outside the worlds of musical performance practice and abstracted from the physical gestures of musicians. In this work, we consider how a recurrent neural network (RNN) model of simple music gestures may be integrated into a physical instrument so that predictions are sonically and physically entwined with the performer's actions. We introduce EMPI, an embodied musical prediction interface that simplifies musical interaction and prediction to just one dimension of continuous input and output. The predictive model is a mixture density RNN trained to estimate the performer's next physical input action and the time at which this will occur. Predictions are represented sonically through synthesized audio, and physically with a motorized output indicator. We use EMPI to investigate how performers understand and exploit different predictive models to make music through a controlled study of performances with different models and levels of physical feedback. We show that while performers often favor a model trained on human-sourced data, they find different musical affordances in models trained on synthetic, and even random, data. Physical representation of predictions seemed to affect the length of performances. This work contributes new understandings of how musicians use generative ML models in real-time performance backed up by experimental evidence. We argue that a constrained musical interface can expose the affordances of embodied predictive interactions.
topic musical performance
interface
mixture density network (MDN)
recurrent neural network (RNN)
creativity
predictive interaction
url https://www.frontiersin.org/article/10.3389/frai.2020.00006/full
work_keys_str_mv AT charlespatrickmartin understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT charlespatrickmartin understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT charlespatrickmartin understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT kyrreglette understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT kyrreglette understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT tønnesfrostadnygaard understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT jimtorresen understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
AT jimtorresen understandingmusicalpredictionswithanembodiedinterfaceformusicalmachinelearning
_version_ 1724822647608442880