Friday, May 31, 2024

Friday, May 24, 2024

Three or four ways to implement Bayesianism

We tend to imagine a Bayesian agent as starting with some credences, “the ur-priors”, and then updating the credences as the observations come in. It’s as if there was a book of credences in the mind, with credences constantly erased and re-written as the observations come in. When we ask the Bayesian agent for their credence in p, they search through the credence book for p and read off the number written beside it.

In this post, I will assume the ur-priors are “regular”: i.e., everything contingent has a credence strictly between zero and one. I will also assume that observations are always certain.

Still the above need not be the right model of how Bayesianism is actually implemented. Another way is to have a book of ur-priors in the mind, and an ever-growing mental book of observations. When you ask such a Bayesian agent what their credence in p, they on the spot look at their book of ur-priors and their book of observations, and then calculate the posterior for p.

The second way is not very efficient: you are constantly recalculating, and you need an ever-growing memory store for all the accumulated evidence. If you were making a Bayesian agent in software, the ever-changing credence book would be more efficient.

But here is an interesting way in which the second way would be better. Suppose you came to conclude that some of your ur-priors were stupid, through some kind of an epistemic conversion experience, say. Then you could simply change your ur-priors without rewriting anything else in your mind, and all your posteriors would automatically be computed correctly as needed.

In the first approach, if you had an epistemic conversion, you’d have to go back and reverse-engineer all your priors, and fix them up. Unfortunately, some priors will no longer be recoverable. From your posteriors after conditionalizing on E, you cannot recover your original priors for situations incompatible with E. And yet knowing what these priors were might be relevant to rewriting all your priors, including the ones compatible with E, in light of your conversion experience.

Here is a third way to implement Bayesianism that combines the best of the two approaches. You have a book of ur-priors and a book of current credences. You update the latter in ordinary updates. In case of an epistemic conversion experience, you rewrite your book of ur-priors, and conditionalize on the conjunction of all the propositions that you currently have credence one in, and replace the contents of your credence book with the result.

We’re not exactly Bayesian agents. Insofar as we approximate being Bayesian agents, I think we’re most like the agents of the first sort, the ones with one book which is ever rewritten. This makes epistemic conversions more difficult to conduct responsibly.

Perhaps we should try to make ourselves a bit more like Bayesian agents of the third sort by keeping track of our epistemic history—even if we cannot go all the way back to ur-priors. This could be done with a diary.

Thursday, May 23, 2024

A supertasked Sleeping Beauty

One of the unattractive ingredients of the Sleeping Beauty problem is that Beauty gets memory wipes. One might think that normal probabilistic reasoning presupposes no loss of evidence, and weird things happen when evidence is lost. In particular, thirding in Sleeping Beauty is supposed to be a counterexample to Van Fraassen’s reflection principle, that if you know for sure you will have a rational credence of p, you should already have one. But that principle only applies to rational credences, and it has been claimed that forgetting makes one not be rational.

Anyway, it occurred to me that a causal infinitist can manufacture something like a version of Sleeping Beauty with no loss of evidence.

Suppose that:

  • On heads, Beauty is woken up at 8 + 1/n hours for n = 2, 4, 6, ... (i.e., at 8.5 hours or 8:30, at 8.25 hours or 8:15, at 8.66… hours or 8:10, and so on).

  • On tails, Beauty is woken up at 8 + 1/n hours for n = 1, 2, 3, ... (i.e. at 9:00, 8:30, 8:20, 8:15, 8:10, …).

Each time Beauty is woken up, she remembers infinitely many wakeups. There is no forgetting. Intuitively she has twice as many wakeups on tails, which would suggest that the probability of heads is 1/3. If so, we have a counterexample to the reflection principle with no loss of memory.

Alas, though, the “twice as many” intuition is fishy, given that both infinities have the same cardinality. So we’ve traded the forgetting problem for an infinity problem.

Still, there may be a way of avoiding the infinity problem. Suppose a second independent fair coin is tossed. We then proceed as follows:

  • On heads+heads, Beauty is woken up at 8 + 1/n hours for n = 2, 4, 6, ...

  • On heads+tails, Beauty is woken up at 8 + 1/n hours for n = 1, 3, 5, ...

  • On tails+whatever, Beauty is woken up at 8 + 1/n hours for n = 1, 2, 3, ....

Then when Beauty wakes up, she can engage in standard Bayesian reasoning. She can stipulatively rigidly define t1 to be the current time. Then the probability of her waking up at t1 if the first coin is heads is 1/2, and the probability of her waking up at t1 if the first coin is tails is 1. And so by Bayes, it seems her credence in heads should be 1/3.

There is now neither forgetting nor fishy infinity stuff.

That said, one can specify that the reflection principle only applies if one can be sure ahead of time that one will at a specific time have a specific rational credence. I think one can do some further modifying of the above cases to handle that (e.g., one can maybe use time-dilation to set up a case where in one reference frame the wakeups for heads+heads are at different times from the wakeups for heads+tails, but in another frame they are the same).

All that said, the above stories all involve a supertask, so they require causal infinitism, which I reject.

Tuesday, May 21, 2024

A problem for probabilistic best systems accounts of laws

Suppose that we live in a Humean universe and the universe contains an extremely large collection of coins scattered on a flat surface. Statistical analysis of all the copper coins fits extremely well with the hypothesis that each coin was independently randomly placed with the chance of heads being 1/16 and that of tails being 15/16.

Additionally, there is a gold coin where you haven’t observed which side it’s on.

And there are no other coins.

On a Lewisian best systems account of laws of nature, if the number of coins is sufficeintly large, it will be a law of nature that all coins are independently randomly placed with the chance of heads being 1/16 and that of tails being 15/16. This is true regardless of whether the gold coin is heads or tails. If you know the information I just gave, and have done the requisite statistical analysis of the copper coins, you can be fully confident that this is indeed a law of nature.

If you are fully confident that it is a law of nature that the chance of tails is 15/16, then your credence for tails for the unobserved gold coin should also be 15/16 (I guess this is a case of the Principal Principle).

But that’s wrong. The fact that the coin is of a different material from the observed coins should affect your credence in its being tails. Inductive inferences are weakened by differences between the unobserved and the observed cases.

One might object that perhaps the Lewisian will say that instead of a law saying that the chance of tails on a coin is 15/16, there would be a law that the chance of tails on a copper coin is 15/16. But that’s mistaken. The latter law is not significantly more informative than the former (given that all but one coin is copper), but is significantly less brief. And laws are generated by balancing informativeness with brevity.

Friday, May 17, 2024

Yet another argument for thirding in Sleeping Beauty?

Suppose that a fair coin has been flipped in my absence. If it’s heads, there is an independent 50% chance that I will be irresistably brainwashed tonight after I go to bed in a way that permanently forces my credence in heads to zero. If it’s tails, there will be no brainwashing. When I wake up tomorrow, there will be a foul taste in my mouth of the brainwashing drugs if and only if I’ve been brainwashed.

So, I wake up tomorrow, find no taste of drugs in my mouth, and I wonder what I should to my credence of heads. The obvious Bayesian approach would be to conditionalize on not being brainwashed, and lower my credence in heads to 1/3.

Next let’s evaluate epistemic policies in terms of a strictly proper scoring accuracy rule (T,F) (i.e., T(p) and F(q) are the epistemic utilities of having credence p when the hypothesis is in fact true or false respectively). Let’s say that the policy is to assign credence p upon observing that I wasn’t brainwashed. My expected epistemic utility is then (1/4)T(p) + (1/4)T(0) + (1/2)F(p). Given any strictly proper scoring rule, this is optimized at p = 1/3. So we get the same advice as before.

So far so good. Now consider a variant where instead of a 50% chance of being brainwashed, I am put in a coma for the rest of my life. I think it shouldn’t matter whether I am brainwashed or put in a coma. Either way, I am no longer an active Bayesian agent with respect to the relevant proposition (namely, whether the coin was heads). So if I find myself awake, I should assign 1/3 to heads.

Next consider a variant where instead of a coma, I’m just kept asleep for all of tomorrow. Thus, on heads, I have a 50% chance of waking up tomorrow, and on tails I am certain to wake up tomorrow. It shouldn’t make a difference whether we’re dealing with a life-long coma or a day of sleep. Again, if I find myself away, I should assign 1/3 to heads.

Now suppose that for the next 1000 days, each day on heads I have a 50% chance of waking up, and on tails I am certain to wake up, and after each day my memory of that day is wiped. Each day is the same as the one day in the previous experiment, so each day I am awake I should assign 1/3 to heads.

But by the Law of Large Numbers, this is basically an extended version of Sleeping Beauty: on heads I will wake up on approximately 500 days and on tails on 1000 days. So I should assign 1/3 to heads in Sleeping Beauty.

Acting for the sake of rationality alone

Alice is confused about the nature of practical rationality and asks wrong philosopher about it. She is given this advice:

  1. For each of your options consider all the potential pleasures and pains for you that could result from the option. Quantify them on a single scale, multiply them by their probabilities, and add them up. Go for the option where the resulting number is biggest.

Some time later, Alice goes to a restaurant and follows the advice to the letter. After spending several hours pouring over the menu and performing back-of-the-envelope calculations she orders and eats the kale and salmon salad.

Traditional decision theory will try to explain Alice’s action in terms of ends and means. What is her end? The obvious guess is that it’s pleasure. But that need not be correct. Alice may not care at all about pleasure. She just cares about doing the action that maximizes the sum of pleasure quantities multiplied by their probabilities. She may not even know that this sum is an “expected value”. It’s just a formula, and she is simply relying on an expert’s opinion as to what formula to use. (If we want to, we could suppose the philosopher gives Alice a logically equivalent formula that was so complicated that she can’t tell that she is maximizing expected pleasure.)

I suppose the right end-means analysis of Alice’s action would be something like this:

  • End: Act rationally.

  • Means: Perform an action that maximizes the sum of products of pleasures and probabilities.

The means is constitutive rather than causal. In this case, there is no causal means that I can see. (Alice may have been misinformed by the same philosopher that there is no such thing as causation.)

The example thus shows that there can be cases of action where one’s aim is simply to act rationally, where one isn’t aiming at any other end. These may be defective cases, but they are nonetheless possible.

Wednesday, May 15, 2024

An interview

Every so often I get asked to do a video interview. I almost always turn down these requests. Recently, I gave in and agreed to do one, because I highly valued the work of the organization that asked me.

It was a terrible experience that has restored my judgment to avoid such things. After initially stumbling (not a big deal), I started talking at length and pretty fluently. But what I was saying was stuff that I hadn’t thought out. It sounded pretty good to me, but it just wasn’t backed up with arguments. Instead of a pattern where first I think and refine what I am about to say, and then I speak, I just spoke, and spoke in a manner that suggested more knowledge than I consciously had. Ugh!

For all I know, all that I said was true, and could be backed up by arguments. But maybe it wasn’t.

Very open-minded scoring rules

An accuracy scoring rule is open-minded provided that the expected value of the score after a Bayesian update on a prospective observation is always greater than or equal to the current expected value of the score.

Now consider a single-proposition accuracy scoring rule for a hypothesis H. This can be thought of as a pair of functions T and F where T(p) is the score for assigning credence p when H is true and F(p) is the score for assigning credence p when H is false. We say that the pair (T,F) is very open-minded provided that the conditional-on-H expected value of the T score after a Bayesian update on a prospective observation is greater than or equal to the current expected value of the T score and provided that the same is true for the F score with the expected value being conditional on not-H.

An example of a very open-minded scoring rule is the logarithmic rule where T(p) = log p and F(p) = log (1−p). The logarithmic rule has some nice philosophical properties which I discuss in this post, and it is easy to see that any very open-minded scoring rule has these properties. Basically, the idea is that if I measure epistemic utilities using a very open-minded scoring rule, then I will not be worried about Bayesian update on a prospective observation damaging other people’s epistemic utilities, as long as these other people agree with me on the likelihoods.

One might wonder if there are any other non-trivial proper and very open-minded scoring rules besides the logarithmic one. There are. Here’s a pretty easy to verify fact (see the Appendix):

  • A scoring rule (T,F) is very open-minded if and only if the functions xT(x) and (1−x)F(1−x) are both convex.

Here’s a cute scoring rule that is proper and very open-minded and proper:

  • T(x) =  − ((1−x)/x)1/2 and F(x) = T(1−x).

(For propriety, use Fact 1 here. For open-mindedness, note that the graph of xT(x) is the lower half of the semicircle with radius 1/2 and center at (1/2,0), and hence is convex.)

What’s cute about this rule? Well, it is symmetric (F(x) = T(1−x)) and it has the additional symmetry property that xT(x) = (1−x)T(1−x) = (1−x)F(x). Alas, though, T is not concave, and I think a good scoring rule should have T concave (i.e., there should be diminishing returns from getting closer to the truth).

Appendix:

Suppose that the prospective observation is as to which cell of the partition E1, ..., En we are in. The open-mindedness property with respect to T then requires:

  1. iP(Ei|H)T(P(H|Ei)) ≥ T(P(H)).

Now P(Ei|H) = P(H|Ei)P(Ei)/P(H). Thus what we need is:

  1. iP(Ei)P(H|Ei)T(P(H|Ei)) ≥ P(H)T(P(H)).

Given that P(H) = ∑iP(Ei)P(H|Ei), this follows immediately from the convexity of xT(x). The converse is easy, too.

Tuesday, May 14, 2024

An argument for purgatory

Here is a plausible argument for purgatory.

Start with these observations:

  1. Some people end up in heaven even though at the time of death they had not yet forgiven all wrongs done to them by other people who end up in heaven nor had they been performing such an act of forgiveness while dying.

  2. An act of forgiveness takes time, and at the beginning of the act one has not yet forgiven.

  3. It is impossible to be in heaven without having forgiven all wrongs done to one by other members of the heavenly community.

Premise 3 seems clearly true: the perfection of the heavenly community requires it.

Premise 1 is pretty plausible. It does not seem that a minor bit of unforgiveness would damn one to hell.

Premise 2 is what I am actually least confident of. It is pretty plausible in our present state. But I guess there is the possibility that we can forgive in the very first instant of our presence in heaven, so that the act is already completed in that very instant. Maybe, but it doesn’t seem very human.

It follows from 1-3 that some people who end up in heaven have to initiate the necessary act of forgiveness post-death. When they initiated the act of forgiveness, they were not in heaven. Nor were they in hell, since they ended up in heaven, and one cannot transfer between heaven and hell. Hence, they must have been in some intermediate state, which we may call purgatory.

Here is a difficulty, though. Suppose a person in heaven is wronged by someone on earth
who will end up in heaven. This surely happens: for instance, a parent is in heaven, and their child on earth fails to fulfill a promise they made to the parent. If an act of forgiveness takes time, isn’t there a short period of time before the person in heaven forgives?

I don’t think so. Perhaps a part of becoming the kind of person that ends up in heaven is one’s having engaged in a prospective forgiveness of all who might wrong one (or at least all who might wrong one and yet are going to be a part of the heavenly community, since the argument above only requires one to forgive such persons as a condition for heavenly beatitude). Some have engaged in it in this life, having transformed themselves into perfect forgivers who have always already forgiven, and others need purgatory.

An argument against strong universalism

  1. It is impossible end up in heaven without forgiving all evils done to one at least by other people who end up in heaven.

  2. Some people have had evils done to them by other people who end up in heaven.

  3. No one is necessitated to forgive evil done to them by other people.

  4. So, at least one person is not necessitated to end up in heaven.

This is an argument against a strong universalism on which God necessitates everyone to go to heaven. It is not an argument against a weaker universalism on which there is a possibility of eternal damnation but no one in fact chooses it. (For the record, alas, I think the Biblical evidence is that the weaker universalism is also false.)

Why do I think the premises are true?

Premise 1: Heavenly beatitude is that of a perfect community of love. Such a community of love is impossible if one has failed to forgive evils done to one by other members of the community.

Premise 2: St Paul did evil to a number of people before his conversion.

Premise 3: This is probably the most controversial of the premises. There are two ways of arguing for it. One is by saying that necessitating someone to forgive is unfitting, and so we have good reason to think God wouldn’t do that—and presumably nobody else but God would be capable of necessitating forgiveness. The second is to note that it is impossible to be forced to forgive. It’s just not forgiveness if it’s forced. One can be forced to stop resenting, one can be forced forget, but that’s not forgiveness. This is akin to promising: it is not possible to force someone to make a promise—the words just wouldn’t be binding.

It's worth noting that the argument also tells against Calvinism.

Monday, May 13, 2024

A feature of the logarithmic scoring rule

Accuracy scoring rules measure the epistemic utility of having some credence assignment. For simplicity, let’s assume that all credence assignments are probabilistically coherent. A strictly proper scoring rule has the property that always by one’s own lights, the expected value of one’s actual credence assignment is better than that of any other credence assignment.

A well-known fact is that a strictly proper scoring rules always makes it rational to update on non-trivial evidence. I.e., by one’s present lights, the expected epistemic utility after examining and updating on non-trivial evidence will be higher than the expected epistemic utility of ignoring that evidence. We might put this by saying that a strictly proper scoring rule is strictly open-minded.

The logarithmic scoring rule makes the score of assigning credence r be log r when the hypothesis is true and log (1−r) when the hypothesis is false. It is strictly proper and hence strictly open-minded.

The logarithmic scoring rule, however, satisfies a condition even stronger than strict open-mindedness. This condition is easiest to describe in a binary case where one is simply evaluating the score of one’s credence in a single hypothesis H. Assuming some non-triviality assumptions, it turns out that not only is the expected epistemic utility increased by examining evidence, but the expected epistemic utility conditional on H is increased by examining evidence. (This is a pretty easy calculation.)

So what?

Well, there are several reasons this matters. First, on my recent account of what it is to have a no-hedge commitment to a hypothesis H, if your epistemic utilities are measured by some scoring rules (e.g., Brier) and you have a no-hedge commitment to H but you do not have credence 1 in H, then you will sometimes have reason to refuse to look at evidence. But the above fact about the logarithmic scoring rule shows that this is not so for the logarithmic scoring rule. With the logarithmic scoring rule, it makes sense to look at the evidence even if you have a no-hedge commitment to H—i.e., even if all your betting behavior is “as if H”.

Second, let’s imagine that I run a funding agency and you come to me with an interest in doing some experiment relevant to a hypothesis H. Let’s suppose that the relevant epistemic community agrees on the relevant likelihoods with respect to the evidence obtainable from the experiment, and is perfectly rational, but differs with regard to the priors of H. I might then have this paternalistic worry about funding the experiment. Even though updating on the results of the experiment by my lights is expected to benefit me epistemically, if a strictly proper scoring rule is the appropriate measure of benefit, it may not be true that by my lights other members of the community will benefit epistemically from updating on the results of the experiment. I may, for instance, be close to certain of H, and think that some members of the community have credences that are sufficiently high that the benefit to them of getting a boost in credence in H from the experiment is outweighed by the risk of misleading evidence. If it is my job to watch out for the epistemic good of the community, this could give me reason to refuse funding.

But not so if I think the logarithmic rule is the right way to evaluate epistemic utility. If everyone shares likelihoods, and we differ only in priors for H, and everyone is rational, then when we measure epistemic utility with the logarithmic rule, I have a positive expectation of the epistemic utility effect of examining the experiment’s results on each member of the community. This is easily shown to follow from my above observation about the logarithmic scoring rule. (By my lights the expectation of a fellow community member’s epistemic utility after updating on the experimental results is a weighted sum of an expectation given H and an expectation given not-H. Each improves given the experiment.)

Saturday, May 11, 2024

What is it like not to be hedging?

Plausibly, a Christian commitment prohibits hedging. Thus in some sense even if one’s own credence in Christianity is less than 100%, one should act “as if it is 100%”, without hedging one’s bets. One shouldn’t have a backup plan if Christianity is false.

Understanding what this exactly means is difficult. Suppose Alice has Christian commitment, but her credence in Christianity is 97%. If someone asks Alice her credence in Christianity, she should not lie and say “100%”, even though that is literally acting “as if it is 100%”.

Here is a more controversial issue. Suppose Alice has a 97% credence in Christianity, but has the opportunity to examine a piece of evidence which will settle the question one way or the other—it will make her 100% certain Christianity is true or 100% certain it’s not. (Maybe she has an opportunity for a conversation with God.) If she were literally acting as if her credence were 100%, there would be no point to looking at any more evidence. But that seems the wrong answer. It seems to be a way of being scared that the evidence will refute Christianity, but that kind of a fear is opposed to the no-hedge attitude.

Here is a suggestion about how no-hedge decision-making should work. When I think about my credences, say in the context of decision-making, I can:

  1. think about the credences as psychological facts about me, or

  2. regulate my epistemic and practical behavior by the credences (use them to compute expected values, etc.).

The distinction between these two approaches to my credences is really clear from a third-person perspective. Bob, who is Alice’s therapist, thinks about Alice’s credences as psychological facts about her, but does not regulate his own behavior by these credences: Alice’s credences have a psychologically descriptive role for Bob but not a regulative role for Bob in his actions. In fact, they probably don’t even have a regulative role for Bob when he thinks about what actions are good for Alice. If Alice has a high credence in the danger of housecats, and Bob does not, Bob will not encourage Alice to avoid housecats—on the contrary, he may well try to change Alice’s credence, in order to get Alice to act more normally around them.

So, here is my suggestion about no-hedging commitments. When you have a no-hedging commitment to a set of claims, you regulate your behavior by them as if the claims had credence 100%, but when you take the credences into account as psychological facts about you, you give them the credence they actually have.

(I am neglecting here a subtle issue. Should we regulate our behavior by our credences or by our opinion about our credences? I suspect that it is by our credences—else a regress results. If that’s right, then there might be a very nice way to clarify the distinction between taking credences into account as psychological facts and taking them into account as regulative facts. When we take them into account as psychological facts, our behavior is regulated by our credences about the credences. When we take them into account regulatively, our behavior is directly regulated by the credences. If I am right about this, the whole story becomes neater.)

Thus, when Alice is asked what her credence in Christianity is, her decision of how to answer depends on the credence qua psychological fact. Hence, she answers “97%”. But when Alice decides whether or not to engage in Christian worship in a time of persecution, her decision on how to answer would normally depend on the credence qua regulative, and so she does not take into account the 3% probability of being wrong about Christianity—she just acts as if Christianity were certain.

Similarly, when Alice considers whether to look at a piece of evidence that might raise or lower her credence in Christianity, she does need to consider what her credence is as a psychological fact, because her interest is in what might happen to her actual psychological credence.

Let’s think about this in terms of epistemic utilities (or accuracy scoring rules). If Alice were proceeding “normally”, without any no-hedge commitment, when she evaluates the expected epistemic value of examining some piece of evidence—after all, it may be practically costly to examine it (it may involve digging in an archaeological site, or studying a new language)—she needs to take her credences into account in two different ways: psychologically when calculating the potential for epistemic gain from her credence getting closer to the truth and potential for epistemic loss from her credence getting further from the truth, and regulatively when calculating the expectations as well as when thinking about what is or is not true.

Now on to some fun technical stuff. Let ϕ(r,t) be the epistemic utility of having credence r in some fixed hypothesis of interest H when the truth value is t (which can be 0 or 1). Let’s suppose there is no as-if stuff going on, and I am evaluating the expected epistemic value of examining whether some piece of evidence E obtains. Then if P indicates my credences, the expected epistemic utility of examining the evidence is:

  1. VE = P(H)(P(E|H)ϕ(P(H|E),1)+P(∼E|H)ϕ(P(H|E),1)) + P(∼H)(P(E|∼H)ϕ(P(H|E),0)+P(∼E|∼H)ϕ(P(H|∼E),0)).

Basically, I am partitioning logical space based on whether H and E obtain.

Now, in the as-if case, basically the agent has two sets of credences: psychological credences and regulative credences, and they come apart. Let Ψ and R be the two. Then the formula above becomes:

  1. VE = R(H)(R(E|H)ϕ(Ψ(H|E),1)+R(∼E|H)ϕ(Ψ(H|∼E),1)) + R(∼H)(R(E|∼H)ϕ(Ψ(H|E),0)+R(∼E|∼H)ϕ(Ψ(H|∼E),0)).

The no-hedging case that interests us makes R(H) = 1: we regulatively ignore the possibility that the hypothesis is false. Our expected value of examining whether E obtains is then:

  1. VE = R(E|H)ϕ(Ψ(H|E),1) + R(∼E|H)ϕ(Ψ(H|∼E),1).

Let’s make a simplifying assumption that the doctrines that we are as-if committed to do not affect the likelihoods P(E|H) and P(E|∣H) (granted the latter may be a bit fishy if P(H) = 1, but let’s suppose we have Popper functions or something like that to take care of that), so that R(E|H) = Ψ(E|H) and R(E|∣H) = Ψ(E|∣H).

We then have:

  1. Ψ(H|E) = Ψ(H)R(E|H)/(R(E|H)Ψ(H)+R(E|∼H)Ψ(∼H)).

  2. Ψ(H|∼E) = Ψ(H)R(∼E|H)/(R(∼E|H)Ψ(H)+R(∼E|∼H)Ψ(∼H)).

Assuming Alice has a preferred scoring rule, we now have a formula that can guide Alice what evidence to look at: she can just check whether VE is bigger than ϕ(Ψ(H),1), which is her current score regulatively evaluated, i.e., evaluated in the as-if H is true way. If VE is bigger, it’s worth checking whether E is true.

One might hope for something really nice, like that if the scoring rule ϕ is strictly proper, then it’s always worth looking at the evidence. Not so, alas.

It’s easy to see that VE beats the current epistemic utility when E is perfectly correlated with H, assuming ϕ(x,1) is strictly monotonic increasing in x.

Surprisingly and sadly, numerical calculations with the Brier score ϕ(x,t) =  − (xt)2 show that if Alice’s credence is 0.97, then unless the Bayes’ factor is very far from 1, current epistemic utility beats VE, and so no-hedging Alice should not look at the evidence, except in rare cases where the evidence is extreme. Interestingly, though, if Alice’s current credence were 0.5, then Alice should always look at the evidence. I suppose the reason is that if Alice is at 0.97, there is not much room for her Brier score to go up assuming the hypothesis is correct, but there is a lot of room for her score to go down. If we took seriously the possibility that the hypothesis could be false, it would be worth examining the evidence just in case the hypothesis is false. But that would be a form of hedging.

Wednesday, May 8, 2024

Forgiving the forgiven

Suppose that Alice wronged Bob, repented, and God forgave Alice for it. Bob, however, withholds his forgiveness. First, it is interesting to ask the conceptual question: What is it that Bob withholds? On my account of objective guilt, when Alice wronged Bob, she gained a normative burden of guilt (minimally, she came to owe it to Bob that she think of herself as guilty), and forgiveness is the removal of that normative burden.

Now in forgiveness, God removed Alice’s normative burden not just to himself, but to Bob. For if God did not remove Alice’s normative burden owed to Bob, then it would be in principle possible that Alice is in heaven—having been forgiven by God—and yet still carries the burden of having wronged Bob. But no one in heaven has a burden.

But if Alice’s normative burden owed to Bob has also been removed by God, and forgiveness is the removal of the burden, then what is it that Bob is withholding?

I think the answer is that there are two parts of forgiveness: there is the removal of the burden of objective guilt and the acknowledgment of the removal of that burden. When God has removed the burden of objective guilt from Alice, all that’s left for Bob to do is to acknowledge this removal.

Note, too, that it would be rather bad for Bob to fail to acknowledge the removal of Alice’s burden, because we should acknowledge what is real and good, and this removal is real and good.

One might think this problem is entirely generated by the idea that God can forgive not just sins against God but also sins against other people. Not so. There seems to be a secular variant of this problem, too. For there seems to be a way in which one’s normative burden of objective guilt of wrongs against fellow humans can be removed without God’s involvement: one can repent of the wrong and suffer an adequate punishment. (Of course, any wrong against neighbor is also a sin against God, and this only removes the guilt with respect to neighbor, unless the punishment is adequate to sin against God, too.) In that case, the burden is presumably removed, but the victim should still acknowledge this removal.

This points to a view of forgiveness on which we ought to forgive those whose normative burden has been removed. If we think that God always forgives the repentant, then this implies that we should always forgive the repentant.

This is close to Aquinas’s view (in his Catechetical Instructions) that we are all required to forgive all those who seek our forgiveness, but it is even better (“perfect” is his phrase) if we forgive even those who do not.

Tuesday, May 7, 2024

Mushrooms

Some people have the intuition that there is something fishy about doing standard Bayesian update on evidence E when one couldn’t have observed the absence of E. A standard case here is where the evidence E is being alive, as in firing squad or fine-tuning cases. In such cases, the intuition goes, you should just ignore the evidence.

I had a great conversation with a student who found this line of thought compelling, and came up with this pretty convincing (and probably fairly standard) case that you shouldn’t ignore evidence E like that. You’re stranded on a desert island, and the only food is mushrooms. They come in a variety of easily distinguishable species. You know that half of the species have a 99% chance of instantly killing you, and otherwise having no effect on you other than nourishment, and the other half have a 1% chance of instantly killing you, again otherwise having no effect on you other than nourishment. You don’t know which are which.

To survive until rescue, you need to eat one mushroom a day. Consider two strategies:

  1. Eat a mushroom from a random species the first day. If you survive, conclude that this species is likely good, and keep on eating mushrooms of the same species.

  2. Eat a mushroom from a random species every day.

The second strategy makes just as much sense as the first if your survival does not count as evidence. But we all know what will happen if you follow the second strategy: you’ll be very likely dead after a few days, as your chance of surviving n mushrooms is (1/2)n. On the other hand, if you follow the first strategy, your chance of surviving n mushrooms is slightly bigger than (1/2)(0.99)n. And the first strategy is precisely what is favored by updating on your survival: you take your survival to be evidence that the mushroom you ate was one of the safer ones, so you keep on eating mushrooms from the same species. If you want to live until rescue, the first strategy is your best bet.

Suppose you’re not yet convinced. Here’s a variant. You have a phone. You call your mom on the first day, and describe your predicament. She comforts you and tells you that rescue will come in a week. And then she tells you that she was once stuck for a week on this very island, and ate the pink lacy mushrooms. Then your battery dies. You rejoice: you will eat the pink lacy mushrooms and thus survive! But then suddenly you get worried. You don’t know when your mom was stuck on the island. If she was stuck on the island before you were conceived, then had she not survived the mushrooms, you wouldn’t have been around to hear it. And in that case, you think her evidence is worthless, because you wouldn’t have any evidence had she not survived. So now it becomes oddly epistemically relevant to you whether your mom was on the island before or after you were conceived. But it seems largely epistemically irrelevant when your mom’s visit to the island was.

Socrates' harm thesis

Socrates famously held that a wrongdoer harms themselves more than they harm their victim.

This is a correct rule of thumb, but I doubt that it is true in general.

First, Socrates was probably thinking of the harm to self resulting from becoming a vicious person. But one can imagine cases where a wrongdoer does not become any more vicious, because they have already maxed out on the vice. I don’t know if such cases are real, though.

But here is a more realistic kind of case. It is said that often abusers were themselves abused. Thus it seems that by abusing another one may cause them to become an abuser. Suppose Alice physically abuses Bob and thereby causes Bob to become an abuser. Then Alice has produced three primary harms:

  1. Bob’s physical suffering

  2. Bob’s being an abuser, and

  3. Alice’s being an abuser.

It seems, then, that Alice has harmed Bob worse than she has harmed herself. For she has harmed herself by turning herself into an abuser. But she has harmed Bob by both turning Bob into an abuser and making him suffer physically.

Objection 1: If Bob becomes an abuser because he was abused, then his responsibility for being an abuser is somewhat mitigated, and hence the moral harm to Bob is less than the moral harm to Alice.

Response: Maybe. But this objection fails if we further suppose that Alice herself was the victim of similar abuse, which mitigated her responsibility to exactly the same degree as Alice’s abuse of Bob mitigates Bob’s responsibility.

Objection 2: One does not cause another to become vicious: one at worst provides an occasion for them to choose to become vicious.

Response: Whether one causes another to become vicious or not is beside the point. One harms the other by putting them in circumstances where they are likely to be vicious. This is why corrupting the youth is so wicked, and why Jesus talks of millstones in connection with those who make others trip up.

From the normative burden of wrongdoing to the existence of God

In recent posts I’ve been exploring the idea that wrongdoing imposes on us a debt of a normative burden.

This yields this argument:

  1. Whenever one does wrong, one comes to have a debt of a normative burden to one who has been wronged.

  2. A debt can only be owed to a person.

  3. One cannot owe a debt to oneself.

  4. Therefore, every wrongdoing includes a wrong to a person.

This has some interesting consequences.

First, it is possible to do wrong to future generations, but one cannot owe anything to the nonexistent. So either eternalism is true, and future generations exist simpliciter, or God exists and we owe a normative burden to God when wrong future generations, or both. So we get the disjunction of eternalism and God’s existence.

Second, we simply get the existence of God. For it is wrong to engage in cruelty to animals even if no human is wronged, other than perhaps oneself. But one cannot be in debt to a non-person or to oneself (debts are the sort of thing one can be released from by the one to whom one owes them; this makes no sense if the creditor is oneself, and impossible if the creditor is a non-person). So the only explanation of whom one can owe the normative burden to is that it’s God, who creates and loves the animals.

If one thinks that it is possible to owe a debt to animals, or one is unconvinced that cruelty to animals is wrong, there is yet another argument for the existence of God. Suppose Alice is the only finite conscious thing in the universe. However, Alice comes across misleading evidence that there are many other finite persons, and that there is a button that, when pressed, will result in excruciating pain to these persons. She then maliciously presses the button. Alice has done wrong, but the only finite conscious thing she can be counted as wronging is herself. She doesn’t owe a normative debt to herself. So she must owe it to something other than a finite conscious being. One cannot owe a debt to anything but a conscious being. So there must be an infinite conscious being, i.e., God.

A perhaps underemphasized aspect of Christ's atonement

Usually, Christ’s sacrifice of the Cross is thought of as atonement for our sins before God. This leads to old theological question: Why can’t God simply forgive our sins, without the need for any atoning sacrifice? Aquinas’s answer is: God could, but it’s more fitting that the debt be paid. I want to explore a different answer.

Suppose that when you do a wrong to someone, you come to owe it to them to be punished. But now instead of thinking of God as the aggrieved party, think of all the times when we have done wrong to other human beings. Some of them have released or will release us from our debt through forgiveness. But, probably, not everyone. But what, now, if we think of Christ’s sacrifice as atomenent for our sins before the unforgiving. We don’t need to pay to other unforgiving humans the debt of being punished, because Christ has paid it on our behalf.

This neatly answers the question of why God’s can’t simply forgive us our sins: God can simply release us from our debt to God, but it is either impossible or at least significantly unfitting for God to simply release us from our debt to fellow human beings.

Here is a consequence of the story. If we fail to forgive our fellow human beings, that is yet another way in which we become shamefully co-responsible for Christ’s sufferings, since now Christ is atoning for these fellow human beings before us. We should then be ashamed of ourselves, especially given that Christ is also suffering for us.

The story isn’t complete. Christ’s atonement applies not just to my sins against my neighbor, but also to my sins against God alone and my sins against myself. But once we have seen that some atoning sacrifice is needed on our behalf, the idea of a total atoning sacrifice, capable of atoning for everyone’s debts to everyone, including to God, looks even more fitting.

Monday, May 6, 2024

Forgiveness

If I have done you a serious wrong, I bear a burden. I can be relieved of that burder by forgiveness. What is the burden and what is the relief?

The burden need not consist of anything emotional or dispositional on your side, such as your harboring resentment or being disposed not to interact with me in as amicable a way as before or pursuing my punishment. For, first, if I secretly betrayed you in such a way that you never found out you were wronged, my burden is still there. And, second, if you die without forgiving me, then the burden feels intact—unless perhaps I believe in life after death or divine forgiveness.

The burden need not consist of something emotional or dispositional on my side, either. For if it had to, I could be relieved of it by therapy. But therapy might make it easier to bear the burden, or (if badly done) may make me think the burden is gone, but the burden will still be there.

People often talk about forgiveness as healing a damaged relationship. But that’s not quite right, either. Suppose I have done many grave wrongs to you over the years that have completely ruptured the relationship. You have finally, generously, brought yourself to forgive me some but not all of them. (A perhaps psychologically odd story: you are working backwards through your life, forgiving all who have wronged you, year by year. So far you’ve forgivenes the wrongs in the last three years of your life. But my earlier wrongs remain.) The remaining ones may be sufficient to make our relationship remain completely ruptured.

The burden is fundamentally a normative feature of reality, as is hinted at by the use of “debt” language in the Lord’s Prayer (“Forgive us our debts as we forgive those indebted to us”). By wronging someone, we make a move in normative space: we burden ourself with an objective, and not merely emotional, guilt. In forgiveness, the burden is removed, but the feeling of burden can remain—one can still feel guilty, just as one’s back can continue hurt when a load is removed from one’s back.

Insofar as there is a healing of a relationship, it is primarily a normative healing. There need not be any great psychological change, as can be seen from the case where you have forgiven me some but not all wrongs. Moreover, psychological change can be slow: forgiveness can be fast, but healing the effects of the wrongdoing can take long.

So far we have identified the type of thing that forgiveness is: it is a move in normative space that relieves something that the wrongdoer owes to a victim. But we are still not clear on what it is that the wrongdoer owes to the victim. And I don’t really know the answer here.

One possibility it is that it has something to do with punishment: I owe it to you to be punished. If so, then there are two ways for the burden to be cleared: one is by being punished and the other is by being forgiven. I can think of one objection to the punishment account: even after being adequately punished, you still can choose whether to forgive me. But if punishment clears the burden, what does your forgiveness do? Maybe it is at this point that the psychological components of forgiveness can enter: it’s up to you whether you stop resenting, whether you accept the clearing of the burden? Plus, in practice, it may be that the punishment is not actually sufficient to clear the burden—a lifetime in jail is not enough for some crimes.

Another possibility is that there is something normative and emotional. I owe it to you to feel guilty, and you can clear that debt and make it no longer obligatory for me to feel that way. That, too, doesn’t seem quite right. One problem is circularity: objective guilt consists in me owing you a feeling of guilt, but a feeling of guilt is a feeling that I am objectively guilty. Maybe the owed feeling has some other description? I don’t know!

But whatever the answer is, I am convinced now that the crucial move in forgiveness is normative.

Thursday, May 2, 2024

The essentiality of dignity

Start with this:

  1. Dignity is an essential property of anything that has it.

  2. Necessarily, something has dignity if and only if it is a person.

  3. Therefore, personhood is an essential property of anything that has it.

Now, suppose the standard philosophical pro-choice view that

  1. Personhood consists in developed sophisticated cognitive faculties of the sort that fetuses and newborns lack but typical toddlers have.

Consider a newborn, Alice. By (4) Alice is not a person, but if she grows up into a typical toddler, that toddler will be a person. By (3), however, we cannot say that Alice will have become that person, since personhood is an essential property, and one cannot gain essential properties—either you necessarily have them or you necessarily lack them.

Call the toddler person “Alicia”. Then Alice is a different individual from Alicia.

So, what happens to Alice once we get to Alicia? Either Alice perishes or where Alicia is, there is Alice co-located with her.

Let’s suppose first the co-location option. We then have two conscious beings, Alice and Alicia, feeling the same things with the same brain, one (Alice) older than the other. We have standard and well-known problems with this absurd position (e.g., how does Alicia know that she is a person rather than just being an ex-fetus?).

But the option that Alice perishes when Alicia comes on the scene is also very strange. For even though Alice is not a person, it is obviously appropriate that Alice’s parents love for and care for her deeply. But if they love for and care for her deeply, they will have significant moral reason to prevent her from perishing. Therefore, they will have significant moral reason to give Alice drugs to arrest her intellectual development at a pre-personhood stage, to ensure that Alice does not perish. But this is a truly abhorrent conclusion!

Thus, we get absurdities from (3) and (4). This means that the pro-choice thinker who accepts (4) will have to reject (3). And they generally do so. This in turn requires them to reject (1) or (2). If they reject (2) but keep (1), then Alice the newborn must have dignity, since otherwise we have to say that Alice is a different entity from the later dignified Alicia, and both the theory that Alice perishes and the theory that Alice doesn’t perish is unacceptable. But if Alice the newborn has dignity, then the pro-choice argument from the lack of developed sophisticated cognitive abilities fails, because Alice the newborn lacks these abilities and so dignity comes apart from these abilities. But if dignity comes apart from these abilities, then the pro-choice argument based on personhood and these cognitive abilities is irrelevant. For it dignity is sufficient to ground a right to life, even absent personhood.

So, I think the pro-choice thinker who focuses on cognitive abilities will in the end need to deny that dignity is an essential property. I suspect most do deny that dignity is an essential property.

But I think the essentiality of dignity is pretty plausible. Dignity doesn’t seem to be something that can come and go. It seems no more alienable than the inalienable rights it grounds. It’s not an achievement, but is at the foundation of what we are.

From fetal pain to the impermissibility of abortion

At some point in pregnancy it is widely acknowledged that fetuses start to feel pain. Estimates of this point vary from around seven to thirty weeks of gestation.

We cannot directly conclude from the fact that some fetus can feel pain that killing that fetus is impermissible. For it seems permissible, given good reason, to humanely kill a conscious non-human animal. But perhaps there is an indirect argument. I want to try out one.

It has been argued that if the fetus is the same individual as the adult person that the fetus would grow into, then it is wrong to kill the fetus for the same reason that it is wrong to kill the adult: the victim is the same, and no more deserving of death, while the harm of death is greater (the fetus is deprived of a greater chunk of life).

But if a fetus can feel pain, then this offers significant support for the hypothesis that the fetus is the same individual as the resultant adult. Imagine the fetus has a constant minor chronic pain, is carried to term, and grows into an adult, without ever any relief to the pain. The adult will then feel the pain. If the fetus is not the same individual as the adult, there are two possibilities at the time of adulthood:

  1. There are two beings feeling pain: the adult and the grown-up fetus.

  2. At some point the grown-up fetus had perished and was replaced by a new individual feeling pain.

Option (1) seems crazy: if I have a headache while sitting alone on the sofa, there is only one entity in pain on the sofa, namely me, rather than me and some grown-up fetus. Option (2) is also rather implausible. On our hypothesis we have the continuous presence of a brain state correlated with pain, and yet allegedly at some point the individual with the pain perishes and a new individual inherits the brain with the pain. That doesn’t seem right.

If we reject both (1) and (2), we have to conclude that the fetus in pain is the same individual as the adult that it grows up into. And thus we conclude that at least once fetuses are capable of pain, abortion is wrong.

This argument doesn’t say anything about what happens prior to the possibility of fetal pain. I think that is still the same individual, but that requires another argument.