Maximum Relative Entropy Updating and the Value of Learning
We examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degre...
Main Author: | |
---|---|
Format: | Article |
Language: | English |
Published: |
MDPI AG
2015-03-01
|
Series: | Entropy |
Subjects: | |
Online Access: | http://www.mdpi.com/1099-4300/17/3/1146 |
id |
doaj-780a441696c3472cb7fd84f60bc8d8c5 |
---|---|
record_format |
Article |
spelling |
doaj-780a441696c3472cb7fd84f60bc8d8c52020-11-24T22:24:28ZengMDPI AGEntropy1099-43002015-03-011731146116410.3390/e17031146e17031146Maximum Relative Entropy Updating and the Value of LearningPatryk Dziurosz-Serafinowicz0Faculty of Philosophy, University of Groningen, Oude Boteringestraat 52, Groningen, 9712 GL, The NetherlandsWe examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degrees of belief that are expected to be helpful and never harmful in making decisions. We call this requirement the value of learning. We consider the extent to which learning rules by MRE could satisfy this requirement and so could be a rational means for pursuing practical goals. First, by representing MRE updating as a conditioning model, we show that MRE satisfies the value of learning in cases where learning prompts a complete redistribution of one’s degrees of belief over a partition of propositions. Second, we show that the value of learning may not be generally satisfied by MRE updates in cases of updating on a change in one’s conditional degrees of belief. We explain that this is so because, contrary to what the value of learning requires, one’s prior degrees of belief might not be equal to the expectation of one’s posterior degrees of belief. This, in turn, points towards a more general moral: that the justification of MRE updating in terms of the value of learning may be sensitive to the context of a given learning experience. Moreover, this lends support to the idea that MRE is not a universal nor mechanical updating rule, but rather a rule whose application and justification may be context-sensitive.http://www.mdpi.com/1099-4300/17/3/1146maximum relative entropyprobabilistic updatingthe value of learning theoremdecision theorySkyrms’s condition Mthe Judy Benjamin problemcontext-sensitivity |
collection |
DOAJ |
language |
English |
format |
Article |
sources |
DOAJ |
author |
Patryk Dziurosz-Serafinowicz |
spellingShingle |
Patryk Dziurosz-Serafinowicz Maximum Relative Entropy Updating and the Value of Learning Entropy maximum relative entropy probabilistic updating the value of learning theorem decision theory Skyrms’s condition M the Judy Benjamin problem context-sensitivity |
author_facet |
Patryk Dziurosz-Serafinowicz |
author_sort |
Patryk Dziurosz-Serafinowicz |
title |
Maximum Relative Entropy Updating and the Value of Learning |
title_short |
Maximum Relative Entropy Updating and the Value of Learning |
title_full |
Maximum Relative Entropy Updating and the Value of Learning |
title_fullStr |
Maximum Relative Entropy Updating and the Value of Learning |
title_full_unstemmed |
Maximum Relative Entropy Updating and the Value of Learning |
title_sort |
maximum relative entropy updating and the value of learning |
publisher |
MDPI AG |
series |
Entropy |
issn |
1099-4300 |
publishDate |
2015-03-01 |
description |
We examine the possibility of justifying the principle of maximum relative entropy (MRE) considered as an updating rule by looking at the value of learning theorem established in classical decision theory. This theorem captures an intuitive requirement for learning: learning should lead to new degrees of belief that are expected to be helpful and never harmful in making decisions. We call this requirement the value of learning. We consider the extent to which learning rules by MRE could satisfy this requirement and so could be a rational means for pursuing practical goals. First, by representing MRE updating as a conditioning model, we show that MRE satisfies the value of learning in cases where learning prompts a complete redistribution of one’s degrees of belief over a partition of propositions. Second, we show that the value of learning may not be generally satisfied by MRE updates in cases of updating on a change in one’s conditional degrees of belief. We explain that this is so because, contrary to what the value of learning requires, one’s prior degrees of belief might not be equal to the expectation of one’s posterior degrees of belief. This, in turn, points towards a more general moral: that the justification of MRE updating in terms of the value of learning may be sensitive to the context of a given learning experience. Moreover, this lends support to the idea that MRE is not a universal nor mechanical updating rule, but rather a rule whose application and justification may be context-sensitive. |
topic |
maximum relative entropy probabilistic updating the value of learning theorem decision theory Skyrms’s condition M the Judy Benjamin problem context-sensitivity |
url |
http://www.mdpi.com/1099-4300/17/3/1146 |
work_keys_str_mv |
AT patrykdziuroszserafinowicz maximumrelativeentropyupdatingandthevalueoflearning |
_version_ |
1725761184406175744 |