Maximum Relative Entropy Updating and the Value of Learning

We examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degre...

Full description

Bibliographic Details
Main Author: Patryk Dziurosz-Serafinowicz
Format: Article
Language:English
Published: MDPI AG 2015-03-01
Series:Entropy
Subjects:
Online Access:http://www.mdpi.com/1099-4300/17/3/1146
id doaj-780a441696c3472cb7fd84f60bc8d8c5
record_format Article
spelling doaj-780a441696c3472cb7fd84f60bc8d8c52020-11-24T22:24:28ZengMDPI AGEntropy1099-43002015-03-011731146116410.3390/e17031146e17031146Maximum Relative Entropy Updating and the Value of LearningPatryk Dziurosz-Serafinowicz0Faculty of Philosophy, University of Groningen, Oude Boteringestraat 52, Groningen, 9712 GL, The NetherlandsWe examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degrees of belief that are expected to be helpful and never harmful in making decisions. We call this requirement the value of learning. We consider the extent to which learning rules by MRE could satisfy this requirement and so could be a rational means for pursuing practical goals. First, by representing MRE updating as a conditioning model, we show that MRE satisfies the value of learning in cases where learning prompts a complete redistribution of one’s degrees of belief over a partition of propositions. Second, we show that the value of learning may not be generally satisfied by MRE updates in cases of updating on a change in one’s conditional degrees of belief. We explain that this is so because, contrary to what the value of learning requires, one’s prior degrees of belief might not be equal to the expectation of one’s posterior degrees of belief. This, in turn, points towards a more general moral: that the justification of MRE updating in terms of the value of learning may be sensitive to the context of a given learning experience. Moreover, this lends support to the idea that MRE is not a universal nor mechanical updating rule, but rather a rule whose application and justification may be context-sensitive.http://www.mdpi.com/1099-4300/17/3/1146maximum relative entropyprobabilistic updatingthe value of learning theoremdecision theorySkyrms’s condition Mthe Judy Benjamin problemcontext-sensitivity
collection DOAJ
language English
format Article
sources DOAJ
author Patryk Dziurosz-Serafinowicz
spellingShingle Patryk Dziurosz-Serafinowicz
Maximum Relative Entropy Updating and the Value of Learning
Entropy
maximum relative entropy
probabilistic updating
the value of learning theorem
decision theory
Skyrms’s condition M
the Judy Benjamin problem
context-sensitivity
author_facet Patryk Dziurosz-Serafinowicz
author_sort Patryk Dziurosz-Serafinowicz
title Maximum Relative Entropy Updating and the Value of Learning
title_short Maximum Relative Entropy Updating and the Value of Learning
title_full Maximum Relative Entropy Updating and the Value of Learning
title_fullStr Maximum Relative Entropy Updating and the Value of Learning
title_full_unstemmed Maximum Relative Entropy Updating and the Value of Learning
title_sort maximum relative entropy updating and the value of learning
publisher MDPI AG
series Entropy
issn 1099-4300
publishDate 2015-03-01
description We examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degrees of belief that are expected to be helpful and never harmful in making decisions. We call this requirement the value of learning. We consider the extent to which learning rules by MRE could satisfy this requirement and so could be a rational means for pursuing practical goals. First, by representing MRE updating as a conditioning model, we show that MRE satisfies the value of learning in cases where learning prompts a complete redistribution of one’s degrees of belief over a partition of propositions. Second, we show that the value of learning may not be generally satisfied by MRE updates in cases of updating on a change in one’s conditional degrees of belief. We explain that this is so because, contrary to what the value of learning requires, one’s prior degrees of belief might not be equal to the expectation of one’s posterior degrees of belief. This, in turn, points towards a more general moral: that the justification of MRE updating in terms of the value of learning may be sensitive to the context of a given learning experience. Moreover, this lends support to the idea that MRE is not a universal nor mechanical updating rule, but rather a rule whose application and justification may be context-sensitive.
topic maximum relative entropy
probabilistic updating
the value of learning theorem
decision theory
Skyrms’s condition M
the Judy Benjamin problem
context-sensitivity
url http://www.mdpi.com/1099-4300/17/3/1146
work_keys_str_mv AT patrykdziuroszserafinowicz maximumrelativeentropyupdatingandthevalueoflearning
_version_ 1725761184406175744