Wednesday, April 22, 2026

A nuanced compatibilism and the problem of heavenly freedom

The problem of heavenly freedom is the apparent tension between these two claims:

  1. The blessed in heaven are free

  2. The blessed in heaven cannot sin.

One solution is compatibilism, but as Pawl and Timpe note, this undercuts the Free Will Defense.

But there is another move. One can be a compatibilist and say that while one can have freedom without the ability to do otherwise, nonetheless freedom with the ability to do otherwise is better. If one accepts this version of compatibilism, one can affirm (1) and (2) while yet offering a Free Will Defense.

This, however, leads to an obvious riposte: If freedom with the ability to do otherwise is better, why don’t we have that kind of freedom in heaven? Isn’t heaven supposed to be the best state for us?

One can, however, add another nuance. There are some activities that it is good to have done at some point, but repetition significantly diminishes the value. It is of some value to have read The Murder of Roger Ackroyd. To re-read it, not so much. Or for a religious example, think of the Hajj. Suppose freedom with the ability to do otherwise is like that. Perhaps, then, it is valuable to have made the choice for God with the ability to do otherwise. But a repeat of that choice is of rather lesser value. So much lesser, that if on earth one has made the choice for God with the ability to do otherwise, in heaven the value of doing so again is outweighed by the value of making guaranteed righteous choices.

This is not too different from Pawl and Timpe’s preferred solution of allowing for derivative freedom in heaven. But there may be an advantage to the above solution. Pawl and Timpe’s solution doesn’t solve the problem of infants who go to heaven without being able to make a free choice in this life—they don’t seem to have derivative freedom. (One of my undergraduate students has ably pressed this problem.) The nuanced compatibilism I have suggested can help with that: the infants in heaven genuinely have freedom. Granted, their death has denied them one of the goods proper to earthly life—the good of choosing righteousness with the ability to do otherwise. But that they have lost something by their untimely death is indeed rather intuitive.

We might ask: But why wouldn’t God then give them a chance to make a decision with ability to choose otherwise after death? Wouldn’t that be better? In one respect, it would indeed be better: in the respect of choosing with the freedom to choose otherwise. But in another respect, it would be less good: in the respect of having the risk of choosing wrongly. These are incommensurable considerations, and God can reasonably follow either one.

Granted, this move weakens the force of the Free Will Defense. We can no longer say that it’s better all things considered for God to give us the kind of freedom that allows us to reject him. For while that’s a better kind of freedom, it comes with an incommensurable cost—the risk that we will reject him. However, we can still say that God can rightly choose to follow either of the incommensurable considerations. In our case, he has opted to give us the better freedom despite the risk; in the case where he has taken some infants to himself, he opted for the guarantee of freedom being rightly used.

I don’t endorse the above solution. But I think it’s possible.

Extending Good's Theorem to experiments and not just observations

Good’s Theorem basically says that a utility-maximizing agent can expect to make decisions that are at least as good if they get more information. (And under some additional conditions, one can expect the decisions to be better.)

Now consider this case:

  1. You will be offered a chance to make a bet at certain odds on the result of a coin toss, where as far as you can tell it’s equally likely that the coin is fair and that it is double-headed. Someone offers to tell you how the previous toss of the coin went.

Good’s Theorem says your decision whether to make the bet will be at least as good given the information about the previous three tosses as without that information. Hence, if the information is being announced, you don’t need to cover your ears. This is, of course, very intuitive. But now consider a slightly different case:

  1. Things are set up just as in (1), except now instead of information about the previous toss, you are offered a chance to have the following experiment get performed before your decision: the coin will be tossed an extra time and the result will be announced to you.

The difference is that in (2) you are not simply being offered additional information about how things are. For whether you go for the experiment or not, either way, you have full information about the experiment and its results. If you don’t go for the experiment, that full information is that the coin was not tossed an extra time (and hence did not land either heads or tails). If you do go for the experiment, the full information is that the coin was tossed and it landed heads, or else that it was tossed and it landed tails. In (2), you are not just finding out information by going for the deal: you are making something happen—an extra toss—and then finding out something about that.

So you can’t apply Good’s Theorem directly to (2). It would be nice to have a formulation of Good’s Theorem that works in cases where instead of merely finding out information, you perform an experiment.

I initially thought this would be easy. Maybe it is, but I don’t see it. There are, after all, cases where performing a cost-free experiment is not a good idea. Suppose, for instance, that you will be allowed to bet tomorrow that a certain car has more than 10 gallons of gasoline. The experiment is to start up the car and look at the gas gauge. But starting the car reduces the amount of gasoline in it, and one can easily rig the case so that benefits from the information gain are outweighed by the fact that you have made that bet less favorable.

So, we want to rule out cases where there is dependence between whether you perform the experiment and the payoffs of the wagers. If F is the event of performing the experiment, it may seems initially we should assume something like:

  1. E(U|WiF) = E(U|WiFc) for all i,

where Wi is your choosing wager i and U is the utility random variable. In other words, the expected utility of each wager is unaffected by whether the experiment has been performed. But no! Suppose a coin has been tossed, and you are choosing between W1 where you get a dollar on heads and W2 where you get a dollar on tails. But let F be the experiment of looking at the coin. (This is a case for the original Good’s Theorem.) Then E(U|WiFc) = 0.50, while E(U|WiF) is very close to 1.00 for the reason that when you find out what the coin is like, you are close to certain to bet on what you see, and hence you are close to certain to win your bet.

If F1 is heads and F2 is tails, we solve the problem by replacing (3) with:

  1. E(U|WiFjF) = E(U|WiFjFc) for i and j.

Namely, the expected utility of wager Wi given information Fj is independent of whether you performed the experiment F. But that only works because it makes sense to ask what the coin is showing if you aren’t looking: it makes sense to conditionalize on Fj ∩ Fc. But in the cases that interest me, there is no fact of the matter as to the result of the experiment when the experiment is not performed, since Molinism is false and we live in an indeterministic world. And in these cases, Fj ∩ Fc is the empty set: the Fj represent the possible results of the experiment but the experiment has no result when it is not performed.

I can get something by supposing a two-step procedure. You perform the experiment, event F, and you learn the result, event L. Then we can assume:

  1. E(U|WiFLc) = E(U|WiFc) for all i

  2. E(U|WiFjFL) = E(U|WiFjFLc) for all i and j

  3. P(Fj|FL) = P(Fj|FLc).

Assumption (5) says that it makes no difference to the expected utility of a wager whether (3) the experiment is performed but its result is not learned or (b) the experiment is not performed at all. In other words, the experiment itself doesn’t affect things. Assumption (6) says that given a specific experimental result, learning the result makes no difference to the expected utility of each wager–result pair. Assumption (7) says that the results of the experiment are unaffected by whether you learn the result of the experiment.

Without (6) or (7), we wouldn’t expect to get the result we want. If we don’t have (6), it might be that utilities are wildly affected by whether you learn the result. (The simplest case is that the wagers all have a big negative payoff on L.) If we don’t have (7), then learning the result might have some evidential or retrocausal impact on what the result is, and then again we shouldn’t expect that learning the result is a good thing.

Given (5)–(7), I think we can now reason as follows. You are choosing between:

  1. performing the experiment and learning the results

and

  1. not performing the experiment and (hence) not learning the results.

By (5), a rational agent will decide the same way in (ii) as in:

  1. performing the experiment and not learning the results,

and the expected utilities of (ii) and (iii) will be the same for this rational agent.

We now apply Good’s Theorem to the choice between (i) and (iii) (we will use (6) and (7) here, and assume the case is non-Newcombian and hence allows the use of Evidential Decision Theory) and get the result that (i) is at least as good as (iii). Since we have indifference between (ii) and (iii), it follows that (i) is at least as good as (ii). (We can also analyze the cases of a strict expected utility inequality.)

This is roundabout, but that’s not main main worry.

What I am really worried about is one technicality. To run the above argument, I had to assume that there is a way of performing the experiment without learning the result, namely that F ∩ Lc is non-empty. In general, however, we cannot assume this. Suppose, for instance, that we have a world with a quantum mechanics where observation causes collapse. Then the experiment of collapsing a wavefunction by means of observation cannot be done without observing the result of the experiment. In such scenarios, I cannot simply introduce a third option of performing the experiment and not learning the results, since that third option may not be consistent with the laws of physics. (And, of course, the utilities for breaking the laws of physics could be wild.)

But without introducing that third option, namely F ∩ Lc, I don’t know how to formulate the independence assumptions that are needed. I also don’t know if the problem is “merely technical” or “deep”. If I had to bet at even odds, I would bet on its being merely technical. But it might be deep.

Monday, April 20, 2026

Consciousness, fine-tuning and skepticism

Models of the emergence of consciousness from a material substrate (whether weak or strong emergence—it won’t matter for this post) differ on how easy it is for consciousness to emerge. Functionalist or computationalist models make it relatively easy: as long as there is a functional isomorphism between a thing and a conscious thing, the former is conscious as well. Biological models, on the other hand, make it harder, by putting constraints on what kind of biological realization of a functional structure gives rise to consciousness.

It’s interesting to note that there the more permissive a model of consciousness is, the easier it is to tune the universe to get consciousness, and hence the better the response that can be given to fine-tuning arguments for theism or a multiverse. On the other hand, the more permissive a model, the greater the danger of skepticism from the fact that the buzzing atoms in a random rock have some sort of isomorphism to a human brain, and hence it is not clear that we have good reason to think we’re not rocks.

On the other hand, the more restrictive a model of consciousness is, the harder it is to tune the universe to get consciousness. On one extreme, you need brains to be conscious. But brains are a specific type of physical organ in DNA-based life forms, so you need life-forms rather like us to have consciousness, and the fine-tuning needed becomes more stringent. On the other hand, the more human-like that conscious things have to be the less skepticism we have to worry about.

Is there some kind of a Goldilocks zone in the range of theories of consciousness where the fine-tuning is not too onerous and skepticism is not an issue? I don’t know.

Lifetime epistemic value

Suppose I discover some fact that I never end up using for anything, or even occurrently thinking about after the discovery. Now, knowledge is good. If I learn the fact earlier in life, then I will have had the knowledge for a longer period of time. So is it better for me to have learned the fact earlier in life?

I doubt it. Consider two scenarios. On the first, I learn what the capital of Zambia is just before I enter a ten-year coma. On the second, I learn it right right after I exit the coma. Learning it before the coma gives me ten more years of knowing it. But that seems a worthless gain. I conclude that in the case of non-occurrent knowing, it doesn’t matter much how long I know.

What about for occurrent knowledge? Other things being equal, if I learn some fact earlier in life, I will occurrently know the fact more times. Is that valuable?

I am less sure. But consider a daily ritual where every morning after waking up, before I am capable of any serious intellectual activity, I think to myself: Sheep have four legs. Thereby, I greatly increase the number of instances in which that piece of knowledge is being occurrently known. Again, this doesn’t seem to be worth the bother.

So it seems that neither for non-occurrent or occurrent knowledge is there non-instrumental value in knowing the thing for a longer period of time. Of course, there typically is instrumental value in knowing something for a longer period of time, both instrumental epistemic value—you can use it in your intellectual investigations of more things—and often instrumental pragmatic value.

This suggests the following. If an agent never loses knowledge, then the lifetime non-instrumental value of their knowledge depends on what they have come to know, not on when they have come to know it. The analogous thesis for perfect Bayesian agents and scoring rules is that their lifetime epistemic utility is the epistemic accuracy score at the latest point in their lives. (If we apply this to Sleeping Beauty, we are apt to get halving. But we shouldn’t apply this to Sleeping Beauty, as she forgets about her first wakeup.)

Things are more complicated in the case of agents who do lose knowledge, whether to memory loss, irrationality or misleading evidence. If we count such an agent’s lifetime non-instrumental epistemic value based on all that they have ever known, that means that if they lost knowledge of p, there is no gain to them from getting it back. But obviously they are better off epistemically if they do get it back. Things get messy and complicated now. A short-period loss in old age doesn’t seem as bad as a case where you found out something early in life and then didn’t have it for the rest of your life.

This is getting messy.

The epistemic value of experiments

You perform an experiment and are going to rationally update on its results. It seems that you should expect this to be good for your epistemic utility as compared to non-performance of the experiment.

Not always! Silly case: Your boss has tasked you with performing a boring chemistry experiment. If you do the experiment, you will find out very little. But if you don’t do it, you will find out a lot about the range of swear words that your boss knows.

What makes this case silly is that you should really think of it as a choice of which experiment to perform, one in chemistry or one in psychology, and in this case the psychology experiment is the more interesting one.

So if we want to say that an experiment can be expected to improve your epistemic utility, we need to be a bit more careful. We need to ensure that non-performance of the experiment doesn’t itself generate information.

But it always does. At the very least, non-performance of the experiment generates the information that the experiment has not been performed by you. You find out something about yourself, and that might far outweigh the value of anything you find out from the experiment. Granted, you also find out something about yourself by performance of the experiment, but it is easy to imagine cases where what you find out by non-performance is more significant. For instance, it could be that your refusal to perform the experiment shows that you have a very specific and rare personality type, while your performance of the experiment gives you nothing so specific.

Suppose, for instance, that you score your epistemic utility by bits of information. The experiment consists in bending down to see which side an unusual coin lying on the ground is facing—that’s one bit of information. Your prior probability that you will look at the coin is 3/4: you are the sort of person who tends to look. So by looking at the coin, you will gain 1 − log2(3/4) = 1.4 bits, mostly regarding the coin but also a little bit about yourself. By not looking at the coin, you will gain 0 − log2(1/4) = 4 bits, all about yourself. Better not to look!

Of course, there are Newcomb-like issues here.

Lesson: The principle that performing a non-trivial experiment should be expected to improve epistemic utility is going to be difficult to formulate.

Epistemic possibility and the Liar

Here’s a fun Liar paradox involving epistemic possibility. Say that a proposition p is epistemically possible if it is consistent with all you know.

Construct a sentence G such that:

  1. G is true if and only if G is not epistemically possible.

E.g., “The proposition expressed by the first sentence in this post found in quotation marks is not epistemically possible.”

Now, you only know truths, and truth is consistent with truth. Thus:

  1. If G is true, then it is consistent with everything you know.

But G is true if and only if it is not epistemically possible. So:

  1. If G is true, then it is not consistent with everything you know.

Hence:

  1. G is not true.

But now that you’ve seen this argument, you surely are in a position to know G not to be true. Suppose you exploit this and indeed come to know G not to be true. But then we have a contradiction. For if you know G not to be true, then G is not epistemically possible, and hence by (0), it must be that G is true.

A piece of Wordle prehistory

A couple of years ago I helped make a variant on Wordle (same rules, copyright-free vocabulary) for the Nintendo Gameboy (you can play it online here), and I would play the official version. Since December, my hobby project has been reverse-engineering the computer built into my early 1990s HP 1653B logic analyzer/oscilloscope, and creating an SDK for programming it. Yesterday, I ported Davison's EhBASIC to it, and was trying out various games in Ahl's BASIC games book from the 1970s (1974 DEC version here), based on the EhBASIC ports here

One of the games I tried last night was Word, credited in the 1974 version of Ahl's book to Charles Reid of Lexington High School. It turns out to have rules very similar to Wordle. It hides a 5-letter puzzle word (there are only 12 in its puzzle vocabulary) and asks you to guess a 5-letter word. Then it shows you which of your letters are correct and in the right position and gives you a list of all the letters that match regardless of position. Basically the same as Wordle. There is no limit on the number of guesses. Here it is running on my oscilloscope. The keyboard is a Mac Quadra keyboard connected via a home-made adapter to the scope's serial port.

Interestingly Word leaks information that Wordle does not. It generates the list of position-independent matches in the array P by the following nested loop where S is the correct solution and L is the user's input word.


The outer loop goes over the letters in the solution S, in order from left-to-right, and adds the position-independent matches to P. Because P is then later printed as is, this means that you know the order in which the position-independent matches appear in the solution, which leaks information (e.g., if you were to put all the right letters in but in a different order, it would actually print the solution). 

Furthermore, if the solution has n repeats of a letter and your guess has m repeats of the same letter, then it will print that letter nm times, and you thus know exactly how many times the letter appears in the solution. Whether this is a bug or just an interesting mechanic depends presumably on what Mr. Charles Reid was thinking half a century ago. (Moreover, if nm>7, the program will crash, because only 7 slots were allocated in the S array. But I think there is no combination of words in game's 12-word vocabulary and English five-letter word that will result in more than 7 slots being occupied.) 

UPDATE: I've been scooped. And you can play the original game in your browser.

Thursday, April 16, 2026

A method for living forever

Maybe you have a cancer that would kill you in three months.

So, get a powerful rocket.

Accelerate close to the speed of light, and make a one light-year round-trip journey that from your reference frame takes about a month, but takes slightly over a year from the point of view of the earth. If your speed during the first journey was v1, now repeat the same trip with a speed of v2 = (3c2+v12)1/2/2. Then repeat with a speed of v3 = (3c2+v22)1/2/2. And so on, forever.

Fact: Each journey will take a bit more than a year of earth-time but only half of the you-time of the previous. So the total you-time of your journeying will be 1 + 1/2 + 1/4 + 1/8 + ... = 2 months. You’ll never die. At every future time, you will be alive.

But this is pointless. You might as well stay on earth, and then you’ll have three months of you-time. Three months of you-time followed by death is better than two months of you-time with no death.

A Christian argument against eternalism, with some remarks on "finite" and "infinite"

  1. We have an infinite future.

  2. If eternalism is true, then anything that has an infinite future is infinite.

  3. We are finite.

  4. So, eternalism is not true.

The crucial premise is 2. One thought behind 2 is that our best version of eternalism holds that we four-dimensional, and if we have an infinite future, that makes us infinite in the fourth dimension.

But I think we can do better than that. Plausibly, part of what we mean by “We have an infinite future” is that we will have infinitely many token future mental states (if not, add that to the premises). On eternalism, all these mental states exist. And they are clearly all ours. So if we have an infinite future, we have an infinite mental life, and that is a way of being infinite.

I am an eternalist, and I want to affirm 1 and 3. What can I do? One move is this. The relevant sense of “finite” in 3 is not a mathematical sense, but something more “metaphysical” like limited. Now, to be limited is to have one or more limits. This is quite compatible with there being respects in which we lack a limit. Thus, the charged infinite rod that sometimes figures in physics homework has limits: not limits of length, but limits of width and height (and others). In the metaphysical sense, then, the rod is finite. Likewise, then, even if we are temporally infinite or infinite in the number of mental states, we are still limited in other ways.

If we go for this move, we have to make a choice what to mean by “infinite”. We could say that something is infinite provided there is some respect in which it is unlimited. If we did that, then one thing could be finite and infinite—as long as it is limited in one way and unlimited in another. The “infinite rod” would then be both finite and infinite. And, if eternalism is true and there is an eternal afterlife, we are finite and infinite. On this take, the argument is invalid, because it is missing the assumption that nothing is both finite and infinite.

A second otion is to make “infinite” mean unlimited in all respects. In that case, we are finite and not infinite. Indeed, only God is infinite then. A set with what the mathematician calls “infinite cardinality” is limited by not having a greater cardinality than the one it has.

A third option would be to take “finite” to mean limited in every way, “infinite” to mean unlimited in all respects, and then allow for the possibility of things that are neither finite or infinite—perhaps us.

Wednesday, April 15, 2026

Anti-Lucretian preferences

Lucretius famously argued that non-existence at the end of one’s life is no more to be feared than non-existence before the beginning of one’s life. Nagel famously argued that there is an asymmetry. One could exist later than one will but one couldn’t have existed earlier than one did. I think he’s barking up the wrong tree. Death wouldn’t be less scary if it turned out to be metaphysically inevitable.

But in any case, I think there is a way to prescind from the metaphysics questions. You’ve just woken up after an operation. You have amnesia. You expect the amnesia to wear off—somehow you have knowledge of how such things go. But for now you have it. You look through some files a careless actuary left lying about. You expect one of these files is about you. The files describe these cases:

  • 35:20. Thirty-five-year-old expected to live twenty years more.

  • 30:20. Thirty-year-old expected to live twenty years more.

  • 20:30. Twenty-year-old expected to live thirty years more.

  • 30:30. Thirty-year-old expected to live thirty years more.

  • 20:20. Twenty-year-old expected to live twenty years more.

You can’t, of course, choose which of these is you, but you can have hopes and preferences. And suppose you think there is no afterlife.

My own preferences would be:

  • 30 : 30 > 20 : 30 > 35 : 20 > 30 : 20 > 20 : 20.

I consistently have a preference for a longer future other things being equal, and a longer past other things being equal, but I tend to prefer a longer future to a longer past even if that results in a somewhat shorter overall life.

But only to a point. Suppose another file is:

  • 50:28.

I definitely would greatly prefer that over 20:30, and not insignificantly over 30:30. The reason is that it seems quite a lot better to live 78 years than 50 or 60, even at the cost of two years of future life.

In any case, as regards my own preferences, Lucretius is just wrong. I would want more of a past life. Though to some degree my intuitions are distorted by the thought that in a longer life I am more likely to have more meaningful achievements.

What worries me philosophically about all this is whether I can reconcile my preferences with my belief in the B-theory of time. I think I can. It makes sense to me that the preferences I have at t should have a relationship to where t is located in my life.

Fear of death is not exactly fear of death or being dead

You don’t believe in the afterlife. Your doctor tells you that you will die in a week. You are terrified. A couple of minutes later, the doctor comes back, herself looking terrified. She tells you that she has both good news and bad news. The good news is that she had misdiagnosed you—you are just fine. However, the bad news is that her sister who is a cosmologist has just discovered that the everything—the universe, space and time—is coming to an end in a week. (She begs you not to tell anyone, because that will cause a panic.)

Out of nerdy curiosity, you ask the doctor whether there will be a last moment of time. She says that the same question occurred to her, and there won’t be. The interval of time is open on the upper end: for every time t, there is a later time t′. It’s just that time is literally running out, and all the remaining times are less than about a week from now.

With grim amusement you note that you won’t die. For at every time in the future you will be alive, and there won’t even be a last time which one might want to identify as the “time of death”.

You reflect. It’s a bit of a plus that none of your friends will suffer from your death, but a big minus that they all have only a week left. In any case, there is no relief from fear of death.

I think this case shows that it’s not death or being dead that we fear when we don’t believe in an afterlife. We fear the fact that our future is finite. If this is right, then people like Lucretius who thought that we somehow confusedly imagined ourselves as existing after the end of our existence and that this was what explained the fear of death are likely mistaken.

A nearly equivalent version of the above thought experiment would be one where you find out that you’re going to live for an infinite amount of time, but your life will exponentially slow down. In the next week of life, you will experience half a subjective week. In the week after that, you will only experience a quarter of a subjective week, and then an eighth and so on. Your subjective future will be a week. But you will never die. That’s just as bad as permanently dying.

Tuesday, April 14, 2026

A problem with perfectly rational agents and decision theory

Suppose I am perfectly rational in the decision theoretic sense. A coin is about to be tossed, and I will get five dollars on heads (H) and one dollar on tails (T). I have a choice whether to leave the coin fair (F) or load it (L) in favor of tails so that the probability of tails is 3/4.

It is obvious what I do. I calculate the expected utilities of my options F and L as follows.

  • EU(F) = P(H|F) ⋅ $5 + P(T|F) ⋅ $1 = (1/2) ⋅ $5 + (1/2) ⋅ $1 = $3

and

  • EU(L) = P(H|L) ⋅ $5 + P(T|L) ⋅ $1 = (1/4) ⋅ $5 + (3/4) ⋅ $1 = $2.

And then I choose F.

Except it’s not so simple. For I am perfectly rational. But since, as we just saw, the perfectly rational agent has to choose F, it follows that P(L) = 0, and so P(H|L) and P(T|L) are undefined. So I can’t decide! So now there is no guarantee how I will act, and P(H|L) and P(T|L) once again make sense. And then again they don’t. Oops!

What can be done? Causal decision theorists will note that I reasoned like an evidential decision theorist above. But this makes no difference in this case. The causalist’s story will be a bit more complicated but will end up with the same problem.

We might want to introduce primitive conditional probabilities like Popper functions that let you conditionalize on events with zero probability, and then have P(H|L) = 1/4 and P(T|L) = 3/4, even though P(L) = 0. But that is introducing a lot of complications. Primitive conditional probabilities are not unproblematic.

What should we do? Maybe we should suppose something like primitive suppositional decision theory, where what we are primitively given are the suppositional probabilities PF and PL, without them being defined in terms of conditional and unconditional credences as in evidential and causal decision theories. But this seems problematic. Do we have to suppose that in addition to conditional and unconditional credences, we have suppositional credences? Maybe.

Or perhaps decision theory only applies to agents that have non-zero credences of going for all the options.

Monday, April 13, 2026

An argument from wonderful people

  1. Person x appears like they are in the image of God.

  2. So, probably, x is in the image of God.

  3. If God doesn’t exist, no one is in the image of God.

  4. So, God exists.

From the imago Dei to God

  1. It is not inappropriate to have a level of respect for human beings at least as great as what would be fitting for beings in the image and likeness of God.

  2. If naturalism is true about humans and God does not exist, then it is inappropriate to have a level of respect for human being that is at least as great as what would be fitting for beings in the image and likeness of God.

  3. So, either naturalism is false about humans or God exists (or both).

Regarding premise 1, think about how problematic it would be to say that someone like Mother Teresa had too much respect for her fellow human beings.

Regarding premise 2, naturalism tells us that innately we’re just an arrangement of atoms, and if we add to that that God doesn’t exist, then this arrangement of atoms doesn’t have a special God-directed significance, so it seems inappropriate to bestow on us the level of respect that a being in the image and likeness of God would have.

I think one can strengthen the argument to provide additional evidence for the existence of God. If God doesn’t exist, then the only plausible way that humans could deserve the imago Dei level of respect is if human beings have a deep and very valuable reality going far beyond the neural networks in our brains, a reality that intrinsically calls for that very high level of respect. (If there is a God, then we don’t need quite as much intrinsic value for us to be worthy of that kind of respect, because we could derive value from our relation to the infinite God.) This is much more than ordinary non-naturalisms about consciousness give.

We thus learn from considerations of respect that if there is no God, humans need to be very non-natural in a god-like way. And beings like that are very hard to explain apart from God. So if we are beings like, this provides significant evidence for theism.

A double lottery and non-normalized probabilities

Suppose a positive integer N is generated by a fair lottery.

Then, a random integer K is chosen between 1 and N (inclusive).

What information does this give you about N?

Obviously you now know that N ≥ K. Anything else?

Consider some specific pair of numbers n ≥ k, and suppose we’ve found out that K = k. What’s the probability that N = n? Of course P(N=n|K=k) = 0/0. But what if we do this as a limiting procedure. Suppose first that N is randomly chosen between 1 and M where M ≥ n, and let PM be the probabilities for this case. Then

  • PM(N=n|K=k) = (1/M)(1/n)/[(1/M)Σj=kMj−1] = (1/n)/Σj=kMj−1.

Take the limit as M goes to infinity. Since Σn=kj−1 = ∞, the limit is zero, so we don’t have a meaningful distribution for N.

On the other hand, what if we independently choose two random integers K1 and K2 between 1 and N? Suppose n ≥ ki for i = 1, 2. Let k* = max (k1,k2). Then:

  • PM(N=n|K1=k1,K2=k2) = (1/M)(1/n2)/[(1/M)Σj=k*Mj−2] = (1/n2)/Σj=k*Mj−2.

Take the limit as M → ∞ and call that P(N=n|K1=k1,K2=k2). The limit behaves like ck*/n2, for a constant c > 0, and generates a well-defined probability for N = n.

With zero samples, we don’t have a well-defined probability for N. With one sample, we still don’t. But with two samples (or more), now we do. This is a rummy thing: how is it that sampling turns probabilistic nonsense into sense?

This is making me more friendly to using non-normalized probabilities. After all, the fair lottery for N is easily modeled by the constant probability p0(n) = 1. With one sample N = k, we have p1(n) = 1/n for n ≥ k and p1(n) = 0 for n < k. With two samples k1, k2, we have p2(n) = 1/n2 for n ≥ max (k1,k2) and p2(n) otherwise. All this makes perfect sense. And there is a lovely mathematical feature of non-normalized probabilities: conditionalization is conjunction. The conditional probability of an event A on event B is just the probability of A ∩ B.

Non-normalized probabilities aren’t going to solve all problems with infinite fair lotteries. For instance, I toss a fair coin and generate a number N with the following rule. On heads, I choose N with my fair lottery on the positive integers. On tails, I choose N such that the probability of N = n is 2n (e.g., I toss an independent fair coin and let N be the number of the first toss that gives heads). What’s my non-normalized probability p(x,n), where x is heads or tails and n is a positive integer? We surely want np(H,n) = ∑np(T,n): the total probability of the heads options equals the total probability of the tails options. But clearly p(T,n) has to exponentially decrease so np(T,n) is finite and non-zero. On the other hand, p(H,n) is constant, so np(H,n) is zero or infinity. So they can’t be equal.

But I wonder if one could say something like this: Non-normalized probabilities make sense in certain cases, and in those cases it’s reasonable to use them?

Thursday, April 9, 2026

Predictability and epistemic utility

You’re thinking whether to become an assembly-line worker or an artist. Then you reflect on the value of knowledge. And you become a factory worker, on the grounds that if you become an assembly-line worker, you will know what you’ll be doing every working day of your future, but if you’re an artist, your activities will be unpredictable.

Some remarks. First, there is something perverse about using the value of knowledge in this way. The normal way to pursue the value of knowledge is to find out things that are independent of your pursuit. But here you are pursuing knowledge by making there be less to know about the world (or your world). Yet, paradoxically, it sure seems like the line of thought above makes sense.

Second, the my initial story depends on Molinism being false. For if there are comprehensive subjective conditionals of free will, then by becoming an artist you get to know the conditionals about what you would do in the various artistic situations you’re in. But on the assembly line story, you don’t get to know these. So the Molinist doesn’t have the paradox. I suppose that’s a bit of evidence for Molinism.

Life-time epistemic utility

I’ve been thinking about the diachronic aspects of epistemic utility. In the case of non-epistemic utility, we can get a decent first approximation to life-time utility by adding up (or, if time is continuous, integrating) momentary utility. But I think this works less well for the epistemic case. For many things of purely epistemic importance, figuring them out is much more important than when one figures them out. (Granted, figuring them out earlier is instrumentally epistemically valuable, because it gives one more time to use leverage the knowledge to figure other things.)

Here’s an extreme version of encoding the value of “figuring something out”. Assuming one does not suffer from mental decline, the epistemic value of one’s life is the epistemic value of the very last moment of it. It’s interesting to note that this won’t work. For imagine that no matter what other credences you had at a given time, you always set the credence of “This is the last moment of my life” to one, while being careful to (inconsistently) make no use of this credence in update. If only the last moment counts, this modification to your credences would be a good idea: it makes sure that when the last moment comes, you get the epistemic utility credit for it.

I suspect that other weightings that favor later over earlier beliefs will suffer from a similar problem—they make it a good idea to err on the side of pessimism about how close death is.

But at the same time, I think some sort of favoring of later beliefs over earlier ones seems appropriate. I don’t know how to resolve this difficulty.

Gender/sex minimization and Christian sexual ethics

Here is a way to live a life: Generally strive to minimize the number of cases where one’s having a particular sex or gender functions as a reason for one’s actions or emotional attitudes.

An extreme version of this is not compatible with traditional Christian sexual ethics unless one is planning on celibacy. It is also not compatible with American law which requires one to correctly fill out various forms, such as census forms, that ask what one’s sex is. And it is not compatible with common-sense morality which requires one to respect things like sex- or gender-segregated bathrooms.

A moderate version of such a gender-minimizing practice, however, could be perhaps sustainable within the bounds of traditional Christian practice, American law and common-sense morality.

One might think that given the heterosexualism of traditional Christian practice, it is hard in romantic contexts to avoid basing decisions on reasons like “I’m a a man and she’s a woman.” In this post I want to explore the idea that one could instead base one’s romantic actions and attitudes on: “We are an opposite-sex pair.”

One might object: “That’s cheating. The reason why the two are an opposite-sex pair is that one is a man and the other is a woman.” But this is “reason” in a different sense of “reason” from that of reasons for actions and attitudes. That one is a man and the other is a woman is a metaphysical ground of the two being an opposite-sex pair, and it may well be one’s epistemic reason for thinking the two are an opposite-sex pair. But it need not be one’s reason for, say, asking the other out on a date—the reason for asking the other out on a date could just be “we are an opposite-sex pair”, and of course the delights of the other’s person, even if the evidence and metaphysical ground for “we are an opposite-sex pair” is the more fine-grained fact that one is a man and the other is a woman.

One could do the same thing when discerning a vocation to the priesthood. Instead of thinking “I’m a man, so I should consider the priesthood”, one might think “I am of the opposite sex to the symbolic sex of the Church (which in turn is the opposite sex to the sex of the incarnate Word), so I should consider the priesthood.”

Would formulating one’s reasons for action in such an unusual way have any benefits? I think so. “We are an opposite-sex pair” focuses one on a relation between the persons. On Trinitarian grounds, there is reason to think that relationality is central to personhood. Half of “I am a man and she is a woman” is self-focused. Better to use “we” than “I” in romantic thinking.

Moreover, we perhaps shouldn’t focus on what is morally irrelevant to a decision. Suppose two people are a good romantic match in terms of character traits, interests, etc., and one is male and the other is female. Supposing (perhaps per impossibile) that their sexes were swapped, but their character traits stayed the same, plausibly they would still be a good romantic match. That they are of the opposite sex is relevant romantically, assuming traditional Christian sexual ethics. But perhaps which one is a man and which one is a woman is not very relevant—what’s relevant is that the couple has one of each sex.

The picture in the above exploration is that the significance of men and women is largely relational: there is a relationship possible to a man and a woman that is not possible to two men or to two women. This is presumably because a man and a woman are an opposite-sex pair, or a potential mating pair, or something like that. This fact about a man and a woman is a relational fact. Granted, this relational fact is metaphysically grounded in certain biological features of the man and the woman, which features may not be themselves relational (say, the existence of body parts with a certain shape, or at least of activated genetic coding for them; though even there the teleology of the parts is relational). But even if the features are not themselves metaphysically relational, their ethical significance could still be largely relational.

Of course, in the end, the physical consummation of love in marital union will require each party to pay attention to the sexed nature of their own and the other’s bodies. That’s unavoidable. But perhaps that’s just a detail? I doubt it’s just a detail myself, but I could very well be wrong.

Another objection to the above story is that in love we focus on the specific features of the other, and in romantic love this includes sex-linked physical features. So when Juliet loves Romeo, she does not love him just as a “someone of the opposite-sex with character traits T1, ..., Tn”, but also as someone with a rich set of lovely physical features, for which it is important that he is male, as many of them would be aesthetically and biologically unfitting in a woman. Agreed! But that doesn’t mean that Juliet’s own femaleness needs to be a part of her reasons for loving Romeo. Instead, she can love him as “someone of the opposite-sex to me with character traits T1, ..., Tn, and with physical features Φ1, ..., Φm which are splendidly fitted to his maleness.” Of course that Romeo is of the opposite-sex to Juliet and that Romeo is male implies that Juliet is female, but even though it implies this, it need not be a part of her reasons for love. I do like the thought that we should minimize focus on self in other-love.

Perhaps the above story isn’t right. I don’t endorse it. It’s entirely hypothetical. I find some features of the story attractive, but my credence in the story is well below 50%. There may well be ethically significant non-relational features of being and being female. Perhaps swapping the sexes of a well-matched romantic couple might in fact change whether they are a good match.

For instance, maybe there are specifically male virtues and specifically female virtues, and then swapping sexes while keeping character the same produces someone who lacks virtues that they should have. Maybe, but I am inclined to be skeptical of this suggestion. A more moderate view would be C. S. Lewis’s, that although men and women should have the same virtues, the lack of certain virtues in a man is worse than their lack in a woman and vice versa. That has more of a chance of being right. I could imagine future scientific research telling us that the typical hormonal make-up of men tends to make some virtues easier for them and the typical hormonal make-up of women tends to make other virtues easier for them. Lacking an easier virtue seems worse than lacking a harder virtue, other things being equal. If so, then if you swapped sexes while keeping characters unchanegd, the moral evaluation could change. Perhaps Alice and Bob are respectively a decent woman and a decent man, but Alice would make a terrible woman and Bob would make a terrible man. Maybe. But if they both got worse symmetrically in this way, while keeping the same actual virtues, maybe it wouldn’t make a big difference to their romantic relationships. They’d still have the same virtues between the two of them, the same vices between the two of them, it’s just that the evaluation of these virtues and vices would be a bit different. And in any case on a story like the hormonal one, it’s not clear that the specific sexes matter except causally, as tending to produce the hormones.

My thinking on this was sparked by two things. First, for a while I’ve been exploring imagining what life would be like if we were isogamous heterothallic organisms, which of course we are not. Second, I recently attended the defense of a very interesting dissertation arguing among other things that it is compatible with Catholic orthodoxy to hold that gender (though not biological sex) distinctions are a major result of original sin.

Epistemic utilities and death

In the previous post, I proved that we get a proper scoring rule if we compute epistemic utilities as follows. We start with our current credence assignment, consider what credence assignment we will have in the future after we update on some further evidence, and then score that. I then suggested that one could get a lifetime epistemic utility by adding up the epistemic utilities over all the moments of life, and as long as death wasn’t random—as long as the lifespan was fixed—this would generate a proper scoring rule. I then said that if death is random (as it is) it might be the case that you don’t get a proper scoring rule.

My conjecture was wrong. You still get a proper scoring rule despite random death. It’s easy to see this. The basic idea is this. Suppose that you might die the next moment. This partitions the probability space into two subsets, D and L, for death and life. Your current credence is p. Next moment, on D, your credence either doesn’t exist (because you don’t exist or you exist in some supernatural state where you don’t have credences) or doesn’t count (because I am after lifetime credences). Thus, the appropriate way to do a forward-looking scoring of your credence p is to score it s(pL) on L and 0 on D, where pL is the result of conditionalizing your credence on the evidence L (after all, if you are alive, you will conditionalize on being alive), and s is some proper scoring rule. In other words, your forward looking score is sL(p) = 1L ⋅ s(pL).

Is this score proper? Yes! For by propriety of s we have:

  • EpL(s(pL)) ≥ EpL(s(qL)).

But this is the same as:

  • (p(L))−1Ep(1Ls(pL)) ≥ (p(L))−1Ep(1Ls(qL)).

Multiplying both sides by p(L) (I am assuming a non-zero probability of survival), we get:

  • Ep(sL(p)) ≥ Ep(sL(q)).

We can combine this with a more complex set of future investigations as in the previous post, and things will still work.

It is crucial to the above argument that when you’re alive, you can tell you’re alive. I suppose that’s not always true. When you’re asleep, you are alive, but can’t tell you’re alive. So to generalize beyond the above toy example, replace death with unconsciousness or something like that.

Forward-looking scoring rules

An accuracy scoring rule assigns a score to a probability function representing an agent’s credences, ostensibly measuring how close that probability function is to the truth. The score s(p) of a probability function p is a random variable, because the value of the score depends on what is actually true, i.e., on where we are in the probability space.

A proper scoring rule (on probabilistic credences) satisfies the propriety inequality

  1. Eps(p) ≥ Eps(q)

which says that the expected score of your current lights—your current credences p—by your current lights is optimal: you won’t improve your expected score (by your current lights) by switching to a different credence q.

You can think of a proper scoring rule as representing the epistemic utility of having a credence p.

But now let’s think about things dynamically. In the future, you will receive additional evidence. As a good Bayesian agent, you will update on this evidence by conditionalization. Perhaps instead of thinking about maximizing your current score, you should think about maximizing your future score. Maybe your true epistemic utility is the score you will end up with after all the future evidence is in.

A simple model of this is as follows. There is some finite partition I = (I1,...,In) of your probability space Ω with each cell Ii of the partition representing a possibility for what you might learn given future evidence. Your current credence function is p, and p(Ii) > 0 for all i. There is then a random credence function pI where pI(ω) is the credence function you will have once the evidene is in if you are at ω ∈ Ω. In other words, pI(ω)(A) = p(AIi) where Ii is the member of the partition that contains ω. (Technically, the function that maps ω to pI(ω)(A) is equal to the conditional probability p(AG) where G is the algebra generated by I.)

Now, given a proper scoring rule s, define a new scoring rule sI as follows:

  1. sI(p)(ω) = s(pI(ω))(ω).

Your sI-score for p at ω then represents the score you will have at ω once you learn which cell of the partition I you are in.

Theorem: The scoring rule sI is proper if s is proper.

Note that sI won’t be strictly proper (i.e., (1) won’t always have strict inequality when p and q are distinct) if I has two or more cells, because pI and qI are going to be the same if p and q assign different probabilities to the cells, but have the same conditional probabilities on each cell. But it might still be the case that sI is strictly proper with respect to some relevant subfield of Ω—that needs some further investigation.

Suppose now you are a Bayesian agent who is guaranteed to consciously live for n moments. In each moment, new information comes in. Thus, we have a sequence J0, ..., Jn of finer and finer partitions, with J0 being the trivial partition, and with pJk representing the credence you will have at time k. Your overall epistemic lifetime score is then:

  1. sΣ(p) = ∑ksJk(p).

It follows from the Theorem that sΣ is a proper scoring rule if s is. And if J0 is the trivial partition, then sJ0 = s, and so if s is strictly proper, then the lifetime score sΣ is strictly proper, since the sum of a strictly proper rule and a proper rule is strictly proper. So, lifetime scores are strictly proper if they are constructed from an instantaneous score—in the above toy model.

Alas, the toy model is not fully adequate, because it is random when we will die, and so our lifespan doesn’t have a fixed sequence of moments. Once we take into account the randomness of when we will die, the overall epistemic lifetime score might stop being proper: this needs further investigation.

Proof of Theorem: By the Greaves and Wallace Theorem, an optimal method of updating credences with respect to expected proper score is by Bayesian conditionalization. Apply the Greaves and Wallace Theorem to the scoring rule s and the starting credence p with the following two strategies:

A. Bayesian conditionalization on the true cell of I.

B. Switch your credence from p to q, then apply Bayesian conditionalization on the true cell of I.

Saying that (A) is at least as good as (B) is equivalent to the the proper scoring rule inequality (1) for sI.

Monday, April 6, 2026

Aristotle on sexual reproduction

Infamously, Aristotle thought that in animal reproduction (including the human case) the male contributed form and the female the matter.

I think it would have made for a neater metaphysics if he thought it was the other way around. For his story involves the complication that there are three ways that a form of an animal exists:

  1. as informing a piece of matter and making it an animal of the relevant type

  2. as being transmitted by semen, without turning the semen into an animal of the relevant type

  3. as being found abstracted in an intellect thinking about the animal.

But if he supposed instead that the form was contributed by the female, he could have said that the semen contributes a small packet of initial matter for the organism and the female’s form works upon that packet in utero. At no point in this process is there a form that is separated out from an organism with that form (such a case will still have to happen when an intellect abstracts that form from matter), and so case (b) disappears, along with all the consequent metaphysical complications to the system.

It’s not clear to me how exactly Aristotle handled this. In Generation of Animals, he talks of the “power of the male residing in the semen”. This sounds like a kind of virtual presence of the form of the animal, a presence of the form as implicit in a causal power. It would be nice to be able to skip this complication. (Granted, Aristotle would still have a similar problem with regard to his view of delayed ensoulment, on which the offspring of a human is first a kind of plant, then gains an animal form, and then gains a rational form, for then the plant has a power to make an animal and the animal a human.)

I guess one might try to find Aristotle a way of reducing (b) to (a) while making the male still be the contributor of form as follows. One might think the semen is a living part of the male’s body, metaphysically like other parts, but separated spatially from the rest of the body (given what we know about how we are made of molecules, it is clear that spatial separation cannot be a bar to bodily unity). Thus, the semen would still be informed by the father’s form, in exactly the same way that the father’s ear or heart is. But this would have a downside. Suppose the father apparently died while the semen continued living, as does happen. Then it seems that the father would not have actually died—instead, his body would be reduced to a packet of semen, and he will have literally gone to seed. This seems absurd.

Why did Aristotle not go for the female as source of form solution, which is so neat metaphysically? Maybe for two reasons. First, the semen obviously could only contribute very little matter. (But then why not suppose that the initial organism is very small?) Second, sexism.

Thursday, April 2, 2026

A really big choice

A messy Irenaean world which leaves much to be desired and then grows in value due to our contribution has a great value.

A perfectionistic world where everyone is already perfectly virtuous and everything is good and we just maintain that level of good also has great value.

Quite plausibly the two worlds have incommensurable value.

God could just choose between these incommensurable values. But what God does is something else: he leaves it to the first humans to choose. This seems an especially great gift: we get to choose the basic shape of the value that our world takes.

We sin, and we get the Irenaean world.

Little evils

From time to time I wonder why it is that we focus on big evils in discussing the problem of evil. After all, first consider this argument:

  1. As far as we can tell, Alice’s cancer is a gratuitous evil.

  2. So, probably, there is a gratuitous evil.

  3. If there is a God, there is no gratuitious evil.

  4. So, probably, there is no God.

And then this one:

  1. As far as we can tell, Bob’s stubbed toe is a gratuitous evil.

  2. So, probably, there is a gratuitous evil.

  3. If there is a God, there is no gratuitious evil.

  4. So, probably, there is no God.

The same kind of backing can be given to (1) and (5). We’ve thought long and hard about Alice’s cancer and its connections to the rest of the world, and can’t find a theodicy. We’ve thought long and hard about BOb’s stubbed toe and its connections to the rest of the world, and can’t find a theodicy. And the rest of the two arguments seems to flow the same way. But as far as I can tell (!), nobody’s become an atheist because of arguments like (5)-(8) while many have become atheists because of arguments like (1)-(4).

What’s going on?

Here is one answer, which at times I’ve felt some pull towards: we shouldn’t take (1)-(4) as seriously as we do.

Here is another answer: we should take (5)-(8) more seriously than we do.

But there may be a good conservative answer, one that maintains our intuition that (1)-(4) should be taken seriously while (5)-(8) shouldn’t. There is a difference in the two arguments, and it lies in the difference between the inference from (1) to (2) and the inference from (5) to (6). In the cancer case, a good that would justify God’s permission of the evil would have to be a very large good. In the stubbed toe case, a justifying good could be pretty small. But it is much easier to overlook a small good than a very large good. Thus, (5) confers much less probability on the existence of gratuitous evil than (1) does.

When I first thought of this conservative answer, I had the image of an elephant: it’s hard to overlook an elephant in the room, but a smaller thing is easier to overlook.

Is that a good image? Are bigger goods easier to see?

Sometimes. A tiny increment in pleasure may be hard to see just as it’s hard to tell the difference in color between two nearby shades of green. If the stubbed toe led to Bob having a slightly greater appreciation of his health the next day, he might not notice.

On the other hand, some really big goods are things we may take for granted, and they need to be pointed out first by some wise person for us to expressly notice them. For instance, the value of human dignity. However, goods that we take for granted are likely irrelevant here—for goods that we take for granted are unlikely to be goods that need Bob’s cancer to exist.

Furthermore, it may not be all about smaller things being harder to see. Consider Rover the dog. How likely is it that there is a cat bigger than Rover? Well, it is more likely when Rover is small than when Rover is big. Similarly, if an evil is smaller, then it is more likely that it is connected to good big enough to justfy it than if the evil is bigger.

So, I think, one can make a case that the inference from a cancer to gratuitous evil is better than the inference from a stubbed toe to a gratuitious evil. But is it much, much better? That I am not sure about. Yet we take (1)-(4) very seriously and we take (5)-(8) not seriously at all. Even if we are right to take (1)-(4) more seriously than (5)-(8), perhaps we make too big a difference between them. Perhaps we shouldn’t take (1)-(4) quite as seriously as we do, and perhaps we should take (5)-(8) more seriously, and perhaps both!