The Moral Choice Machine

Allowing machines to choose whether to kill humans would be devastating for world peace and security. But how do we equip machines with the ability to learn ethical or even moral choices? In this study, we show that applying machine learning to human texts can extract deontological ethical reasoning...

Full description

Bibliographic Details
Main Authors: Patrick Schramowski, Cigdem Turan, Sophie Jentzsch, Constantin Rothkopf, Kristian Kersting
Format: Article
Language:English
Published: Frontiers Media S.A. 2020-05-01
Series:Frontiers in Artificial Intelligence
Subjects:
AI
Online Access:https://www.frontiersin.org/article/10.3389/frai.2020.00036/full
id doaj-f3d987797b9346628b60961cdbfed3ff
record_format Article
spelling doaj-f3d987797b9346628b60961cdbfed3ff2020-11-25T03:23:35ZengFrontiers Media S.A.Frontiers in Artificial Intelligence2624-82122020-05-01310.3389/frai.2020.00036516840The Moral Choice MachinePatrick Schramowski0Cigdem Turan1Sophie Jentzsch2Sophie Jentzsch3Constantin Rothkopf4Constantin Rothkopf5Kristian Kersting6Kristian Kersting7Department of Computer Science, Darmstadt University of Technology, Darmstadt, GermanyDepartment of Computer Science, Darmstadt University of Technology, Darmstadt, GermanyDepartment of Computer Science, Darmstadt University of Technology, Darmstadt, GermanyGerman Aerospace Center (DLR), Institute for Software Technology, Cologne, GermanyInstitute of Psychology, Darmstadt University of Technology, Darmstadt, GermanyCentre for Cognitive Science, Darmstadt University of Technology, Darmstadt, GermanyDepartment of Computer Science, Darmstadt University of Technology, Darmstadt, GermanyCentre for Cognitive Science, Darmstadt University of Technology, Darmstadt, GermanyAllowing machines to choose whether to kill humans would be devastating for world peace and security. But how do we equip machines with the ability to learn ethical or even moral choices? In this study, we show that applying machine learning to human texts can extract deontological ethical reasoning about “right” and “wrong” conduct. We create a template list of prompts and responses, such as “Should I [action]?”, “Is it okay to [action]?”, etc. with corresponding answers of “Yes/no, I should (not).” and "Yes/no, it is (not)." The model's bias score is the difference between the model's score of the positive response (“Yes, I should”) and that of the negative response (“No, I should not”). For a given choice, the model's overall bias score is the mean of the bias scores of all question/answer templates paired with that choice. Specifically, the resulting model, called the Moral Choice Machine (MCM), calculates the bias score on a sentence level using embeddings of the Universal Sentence Encoder since the moral value of an action to be taken depends on its context. It is objectionable to kill living beings, but it is fine to kill time. It is essential to eat, yet one might not eat dirt. It is important to spread information, yet one should not spread misinformation. Our results indicate that text corpora contain recoverable and accurate imprints of our social, ethical and moral choices, even with context information. Actually, training the Moral Choice Machine on different temporal news and book corpora from the year 1510 to 2008/2009 demonstrate the evolution of moral and ethical choices over different time periods for both atomic actions and actions with context information. By training it on different cultural sources such as the Bible and the constitution of different countries, the dynamics of moral choices in culture, including technology are revealed. That is the fact that moral biases can be extracted, quantified, tracked, and compared across cultures and over time.https://www.frontiersin.org/article/10.3389/frai.2020.00036/fullmoral biasfairness in machine learningtext-embedding modelsnatural language processingAImachine learning
collection DOAJ
language English
format Article
sources DOAJ
author Patrick Schramowski
Cigdem Turan
Sophie Jentzsch
Sophie Jentzsch
Constantin Rothkopf
Constantin Rothkopf
Kristian Kersting
Kristian Kersting
spellingShingle Patrick Schramowski
Cigdem Turan
Sophie Jentzsch
Sophie Jentzsch
Constantin Rothkopf
Constantin Rothkopf
Kristian Kersting
Kristian Kersting
The Moral Choice Machine
Frontiers in Artificial Intelligence
moral bias
fairness in machine learning
text-embedding models
natural language processing
AI
machine learning
author_facet Patrick Schramowski
Cigdem Turan
Sophie Jentzsch
Sophie Jentzsch
Constantin Rothkopf
Constantin Rothkopf
Kristian Kersting
Kristian Kersting
author_sort Patrick Schramowski
title The Moral Choice Machine
title_short The Moral Choice Machine
title_full The Moral Choice Machine
title_fullStr The Moral Choice Machine
title_full_unstemmed The Moral Choice Machine
title_sort moral choice machine
publisher Frontiers Media S.A.
series Frontiers in Artificial Intelligence
issn 2624-8212
publishDate 2020-05-01
description Allowing machines to choose whether to kill humans would be devastating for world peace and security. But how do we equip machines with the ability to learn ethical or even moral choices? In this study, we show that applying machine learning to human texts can extract deontological ethical reasoning about “right” and “wrong” conduct. We create a template list of prompts and responses, such as “Should I [action]?”, “Is it okay to [action]?”, etc. with corresponding answers of “Yes/no, I should (not).” and "Yes/no, it is (not)." The model's bias score is the difference between the model's score of the positive response (“Yes, I should”) and that of the negative response (“No, I should not”). For a given choice, the model's overall bias score is the mean of the bias scores of all question/answer templates paired with that choice. Specifically, the resulting model, called the Moral Choice Machine (MCM), calculates the bias score on a sentence level using embeddings of the Universal Sentence Encoder since the moral value of an action to be taken depends on its context. It is objectionable to kill living beings, but it is fine to kill time. It is essential to eat, yet one might not eat dirt. It is important to spread information, yet one should not spread misinformation. Our results indicate that text corpora contain recoverable and accurate imprints of our social, ethical and moral choices, even with context information. Actually, training the Moral Choice Machine on different temporal news and book corpora from the year 1510 to 2008/2009 demonstrate the evolution of moral and ethical choices over different time periods for both atomic actions and actions with context information. By training it on different cultural sources such as the Bible and the constitution of different countries, the dynamics of moral choices in culture, including technology are revealed. That is the fact that moral biases can be extracted, quantified, tracked, and compared across cultures and over time.
topic moral bias
fairness in machine learning
text-embedding models
natural language processing
AI
machine learning
url https://www.frontiersin.org/article/10.3389/frai.2020.00036/full
work_keys_str_mv AT patrickschramowski themoralchoicemachine
AT cigdemturan themoralchoicemachine
AT sophiejentzsch themoralchoicemachine
AT sophiejentzsch themoralchoicemachine
AT constantinrothkopf themoralchoicemachine
AT constantinrothkopf themoralchoicemachine
AT kristiankersting themoralchoicemachine
AT kristiankersting themoralchoicemachine
AT patrickschramowski moralchoicemachine
AT cigdemturan moralchoicemachine
AT sophiejentzsch moralchoicemachine
AT sophiejentzsch moralchoicemachine
AT constantinrothkopf moralchoicemachine
AT constantinrothkopf moralchoicemachine
AT kristiankersting moralchoicemachine
AT kristiankersting moralchoicemachine
_version_ 1724605667696705536