Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy

By the early 2020s, emotional artificial intelligence (emotional AI) will become increasingly present in everyday objects and practices such as assistants, cars, games, mobile phones, wearables, toys, marketing, insurance, policing, education and border controls. There is also keen interest in using...

Full description

Bibliographic Details
Main Author: Andrew McStay
Format: Article
Language:English
Published: SAGE Publishing 2020-03-01
Series:Big Data & Society
Online Access:https://doi.org/10.1177/2053951720904386
id doaj-2011da6c161d4fd38eaff40fe111861b
record_format Article
spelling doaj-2011da6c161d4fd38eaff40fe111861b2020-11-25T03:35:04ZengSAGE PublishingBig Data & Society2053-95172020-03-01710.1177/2053951720904386Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacyAndrew McStayBy the early 2020s, emotional artificial intelligence (emotional AI) will become increasingly present in everyday objects and practices such as assistants, cars, games, mobile phones, wearables, toys, marketing, insurance, policing, education and border controls. There is also keen interest in using these technologies to regulate and optimize the emotional experiences of spaces, such as workplaces, hospitals, prisons, classrooms, travel infrastructures, restaurants, retail and chain stores. Developers frequently claim that their applications do not identify people. Taking the claim at face value, this paper asks, what are the privacy implications of emotional AI practices that do not identify individuals? To investigate privacy perspectives on soft non-identifying emotional AI, the paper draws upon the following: over 100 interviews with the emotion detection industry, legal community, policy-makers, regulators and NGOs interested in privacy; a workshop with stakeholders to design ethical codes for using data about emotions; a UK survey of 2068 citizens on feelings about emotion capture technologies. It finds a weak consensus among social stakeholders on the need for privacy, this driven by different interests and motivations. Given this weak consensus, it concludes that there exists a limited window of opportunity to societally agree principles of practice regarding privacy and the use of data about emotions.https://doi.org/10.1177/2053951720904386
collection DOAJ
language English
format Article
sources DOAJ
author Andrew McStay
spellingShingle Andrew McStay
Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
Big Data & Society
author_facet Andrew McStay
author_sort Andrew McStay
title Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
title_short Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
title_full Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
title_fullStr Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
title_full_unstemmed Emotional AI, soft biometrics and the surveillance of emotional life: An unusual consensus on privacy
title_sort emotional ai, soft biometrics and the surveillance of emotional life: an unusual consensus on privacy
publisher SAGE Publishing
series Big Data & Society
issn 2053-9517
publishDate 2020-03-01
description By the early 2020s, emotional artificial intelligence (emotional AI) will become increasingly present in everyday objects and practices such as assistants, cars, games, mobile phones, wearables, toys, marketing, insurance, policing, education and border controls. There is also keen interest in using these technologies to regulate and optimize the emotional experiences of spaces, such as workplaces, hospitals, prisons, classrooms, travel infrastructures, restaurants, retail and chain stores. Developers frequently claim that their applications do not identify people. Taking the claim at face value, this paper asks, what are the privacy implications of emotional AI practices that do not identify individuals? To investigate privacy perspectives on soft non-identifying emotional AI, the paper draws upon the following: over 100 interviews with the emotion detection industry, legal community, policy-makers, regulators and NGOs interested in privacy; a workshop with stakeholders to design ethical codes for using data about emotions; a UK survey of 2068 citizens on feelings about emotion capture technologies. It finds a weak consensus among social stakeholders on the need for privacy, this driven by different interests and motivations. Given this weak consensus, it concludes that there exists a limited window of opportunity to societally agree principles of practice regarding privacy and the use of data about emotions.
url https://doi.org/10.1177/2053951720904386
work_keys_str_mv AT andrewmcstay emotionalaisoftbiometricsandthesurveillanceofemotionallifeanunusualconsensusonprivacy
_version_ 1724555764768440320