Warning: I worry there may be something wrong in the reasoning below.
Causal Decision Theory (CDT) and Epistemic Decision Theory (EDT) tend to disagree when the payoff of an option statistically depends on your propensity to go for that option. The most example of this phenomenon is Newcomb’s Problem (where money is literally put into a box or not depending on what your propensities are), and there is a large literature of other clever and mind-twisting examples. From the literature, one might get a feeling that these cases are all somehow weird, and normally there is no such dependence.
But here is a family of cases that happens literally almost all the time to us. Pretty much whenever we act we gain information relevant to facts about ourselves, and specifically to facts about our propensities to act. For instance, when you choose chocolate over vanilla ice cream you raise your credence for the hypothesis that you have a greater propensity to choose chocolate ice cream than to choose vanilla ice cream. But truth about oneself is valuable and falsehood about oneself is disvaluable. If in fact you have a greater propensity to choose chocolate ice cream, then by eating chocolate ice cream you gain credence in a truth, which is a good thing. If in fact your propensity for vanilla ice cream is at least as great as for chocolate ice cream, then by eating chocolate ice cream, you gain credence in a falsehood. The payoffs of your decision as to flavor of ice cream thus statistically depend on what your propensities actually are, and so this is exactly the kind of case where we would expect CDT and EDT to disagree.
Let’s be more precise. You have a choice between eating chocolate ice cream (C), eating vanilla ice cream (V) or not eating ice cream at all (N). Let H be the hypothesis that you have a greater propensity for eating chocolate ice cream than for eating vanilla ice cream. Then if you choose C, you will gain evidence for H. If you choose V, you will gain evidence for not-H. And if you choose N, you will (plausibly) gain no evidence for or against H. Your epistemic utility with respect to H is, let us suppose, measured by a single-proposition accuracy scoring rule, which we can think of as a pair of functions TH and FH, where TH(p) is the value of having credence p in H if in fact H is true and FH(p) is the value of having credence p in H if in fact H is false.
The expected evidential utilities of your three options are:
Ee(C) = P(H|C)TH(P(H|C)) + (1−P(H|C))FH(P(H|C))
Ee(V) = P(H|V)TH(P(H|V)) + (1−P(H|V))FH(P(H|V))
Ee(N) = P(H|N)TH(P(H|N)) + (1−P(H|N))FH(P(H|N)) = P(H)TH(P(H)) + (1−P(H))FH(P(H)).
The expected causal utilities are:
Ec(C) = P(H)TH(P(H|C)) + (1−P(H))FH(P(H|C))
Ec(V) = P(H)TH(P(H|V)) + (1−P(H))FH(P(H|V))
Ec(N) = P(H)TH(P(H|N)) + (1−P(H))FH(P(H|N)) = P(H)TH(P(H)) + (1−P(H))FH(P(H)).
We can make some quick observations in the case where the scoring rule is strictly proper, given that P(H|V) < P(H) < P(H|C):
Ec(C) < Ec(N)
Ec(V) < Ec(N)
At least one of Ee(C) > Ee(N) and Ee(V) > Ee(N) is true.
Observations 1 and 2 follow immediately from strict propriety and the formulas for Ec. Observation 3 follows from the fact that the expected accuracy score after Bayesian update on evidence is better (in non-trivial cases where the scoring rule is strictly proper) than before update, and the expected accuracy score after update on what you’ve chosen is:
- P(C)Ee(C) + P(V)Ee(V) + P(N)Ee(N)
while the expected accuracy score before update is equal to Ee(N). Since P(C) + P(V) + P(N) = 1, it follows from the superiority of the post-update expectation that at least one of Ee(C) and Ee(V) must be bigger than Ee(N).
The above results seem to be a black eye for CDT, which recommends that if what you care about is your epistemic utility with regard to your propensities regarding chocolate and vanilla ice cream, then you should always avoid eating ice cream!
(What about ratifiability? Some CDTers say that only ratifiable options should count. Is N ratifiable? Given that you’ve learned nothing about H from choosing N, I think N should be ratifiable. But I may be missing something. I find the epistemic utility case confusing.)
It also seems to me (I haven’t checked details) that on EDT there are cases where eating either flavor is good for you epistemically, but there are also cases where only one specific flavor is good for you.
No comments:
Post a Comment