Tuesday, February 18, 2025

An example of a value-driven epistemological approach to metaphysics

  1. Everything that exists is intrinsically valuable.

  2. Shadows and holes are not intrinsically values.

  3. So, neither shadows nor holes exist.

Monday, February 17, 2025

Incompleteness

For years in my logic classes I’ve been giving a rough but fairly accessible sketch of the fact that there are unprovable arithmetical truths (a special case of Tarski’s indefinability of truth), using an explicit Goedel sentence using concatenation of strings of symbols rather than Goedel encoding and the diagonal lemma.

I’ve finally revised the sketch to give the full First Incompleteness theorem, using Rosser’s trick. Here is a draft.

Friday, February 14, 2025

What numbers could be

Benacerraf famously argued that no set theoretic reduction can capture the natural numbers. While one might conclude from this that the natural numbers are some kind of sui generis entities, Benacerraf instead opts for a structuralist view on which different things can play the role of different numbers.

The argument that no set theoretic reduction captures the negative numbers is based on thinking about two common reductions. On both, 0 is the empty set . But then the two accounts differ in how the successor sn of a number n is formed:

  1. sn = n ∪ {n}

  2. sn = {n}.

On the first account, the number 5 is equal to the set {0, 1, 2, 3, 4}. On the second account, the number 5 is equal to the singleton {{{{{⌀}}}}}. Benacerraf thinks that we couldn’t imagine a good argument for preferring one account over another, and hence (I don’t know how this is supposed to follow) there can’t be a fact of the matter about why one account—or any other set-theoretic reductive account—is correct.

But I think there is a way to adjudicate different set-theoretic reductions of numbers. Plausibly, there is reference magnetism to simpler referrents of our terminology. Consider an as consisting of a set of natural numbers, a relation <, and two operations + and ⋅, satisfying some axioms. We might then say that our ordinary language arithmetic is attracted to the abstract entities that are most simply defined in terms of the fundamental relations. If the only relevant fundamental relation is set membership , then we can ask which of the two accounts (a) and (b) more simply defines <, + and .

If simplicity is brevity of expression in first order logic, then this can be made a well-defined mathematical question. For instance, on (a), we can define a < b as a ∈ b. One provably cannot get briefer than that. (Any definition of a < b will need to contain a, b and .) On the other hand, on (b), there is no way to define a < b as simply. Now it could turn out that + or can be defined more simply on (b), in a way that offsets (a)’s victory with <, but it seems unlikely to me. So I conjecture that on the above account, (a) beats (b), and so there is a way to decide between the two reductions of numbers—(b) is the wrong one, while (a) at least has a chance of being right, unless there is a third that gives a simpler reduction.

In any case, on this picture, there is a way forward in the debate, which undercuts Benacerraf’s claim that there is no way forward.

I am not endorsing this. I worry about the multiplicity of first-order languages (e.g., infix-notation FOL vs. Polish-notation FOL).

Tuesday, February 11, 2025

Theistic Humeanism?

Here’s an option that is underexplored: theistic Humeanism. There are two paths to it.

The path from orthodoxy: Start with a standard theistic concurrentism: whenever we have a creaturely cause C with effect E, E only eventuates because God concurs, i.e., God cooperates with the creaturely causal relation. Now add to this a story about what creaturely causation is. This will be a Humean story—the best I know is the David Lewis one that reduces causation to laws and laws to arrangements of stuff. Keep all the deep theistic metaphysics of divine causation.

The path from heterodoxy: Start with the metaphysics of occasionalism. Don’t change any of the metaphysics. But now add a Humean analysis of creaturely causation in terms of regularities. Since the metaphysics of occasionalism affirms regularities in the world, we haven’t changed the metaphysics of occasionalism, but have redescribed it as actually involving creaturely causation.

The two paths meet in a single view, a theistic Humeanism with the metaphysics of occasionalism and the language of concurrentism, and with creaturely causation described in a Humean way.

This theistic Humeanism is more complex than standard non-theistic Humeanism, but overcomes the central problem with non-theistic Humeanism: the difficulty of finding explanation in nature. If the fact that heat causes boiling is just a statement of regularity, it does not seem that heat explains boiling. But on theistic Humeanism, we have a genuine explanatory link: God makes the water boil because God is aware of the heat.

There is one special objection to theistic Humeanism. It has two causal relations, a divine one and a creaturely one. But the two are very different—they don’t both seem to be kinds of causation. However, on some orthodox concurrentisms, such as Aquinas’s, there isn’t a single kind of thing that divine and creaturely causation are species of. Instead, the two stand in an analogical relationship. Couldn’t the theistic Humean say the same thing? Maybe, though one might also object that Humean creaturely causation is too different from divine causation for the two to count as analogous.

I suppose the main objection to theistic Humeanism is that it feels like a cheat. The creaturely causation seems fake. The metaphysics is that of occasionalism, and there is no creaturely causation there. But if theistic Humeanism is a cheat, then standard non-theistic Humeanism is as well, since they share the same metaphysics of creaturely causation. If non-theistic Humeanism really does have causation, then our theistic Humeanism really does have creaturely causation. If one has fake causation, so does the other. I think both have fake causation. :-)

Monday, February 10, 2025

Autonomy and relativism

Individual relativism may initially seem to do justice to the idea of our autonomy: our moral rules are set by ourselves. But this attractiveness of relativism disappears as soon as we realize that our beliefs are largely not up to us—that, as the saying goes, we catch them like we catch the flu. This seems especially true of our moral beliefs, most of which are inherited from our surrounding culture. Thus, what individual relativism gives to us in terms of autonomy is largely taken away by reflection on our beliefs.

Tuesday, February 4, 2025

Asymmetry between moral and physical excellence

We can use a Mahatma Ghandi or a Mother Teresa as a moral exemplar to figure out what our virtues should be. But we cannot use an Usain Bolt or a Serena Williams as a physical exemplar to figure out what our physical capabilities should be. Why this disanalogy between moral and physical excellence?

It’s our intuition that Bolt and Williams exceed the physical norms for humans to a significant degree. But although Ghandi and Mother Teresa did many supererogatory things, I do not think they overall exceed the moral norms for human character to a significant degree. We should be like them, and our falling short is largely our fault.

My LaTeX "ide"

I haven’t found a LaTeX IDE that I am happy with (texmaker comes close, but I don’t like the fact that it doesn’t properly underline the trigger letter in menus, even if Windows is set to do that), and so I ended up defaulting to just editing my book and papers with notepad++ and running pdflatex manually. But it’s a bit of a nuisance to get the preview: ctrl-s to save, alt-tab to command-line, up-arrow and enter to re-run pdflatex, alt-tab to pdf viewer. So I wrote a little python script that watches my current directory and if any .tex file changes in it, it re-runs pdflatex. So now it’s just ctrl-s, alt-tab to get the preview. I guess it’s only four keystrokes saved, but it feels more seamless and straightforward. The script also launches notepad++ and my pdf viewer at the start of the session to save me some typing.

Thursday, January 30, 2025

Teleology and the normal/abnormal distinction

Believers in teleology also tend to believe in a distinction between the normal and the abnormal. I think teleology can be prised apart from a normal/abnormal distinction, however, if we do something that I think we should do for independent reasons: recognize teleological directedness without a telos-to-be-attained, a target to be hit. An example of such teleological directedness is an athlete trying to run as fast as possible. There isn’t a target telos: for any speed the athlete reaches, a higher speed would fit even better with the athlete’s aims. But there is a directional telos, an aim tlos: the athlete aims in the direction of higher speed.

One might then say the human body in producing eyes has a directional telos: to see as well as possible. Whether one has 20/20 or 20/15 or 20/10 vision, more acuity would fulfill that directional telos better. On this view, there is no target telos, just a direction towards better acuity. If there were a target telos, say a specific level of acuity, we could identify non-attainment with abnormalcy and attainment with normalcy. But we need not. We could just say that this is all a matter of degree, with continuous variation between 20/0 (not humanly available) and 20/∞ (alas humanly available, i.e., total blindness).

I am not endorsing the view that there is no normal/abnormal in humans. I think there is (e.g., an immoral action is abnormal; a moral action is normal). But perhaps the distinction is less often applicable than friends of teleology think.

Wednesday, January 29, 2025

More on experiments

We all perform experiments very often. When I hear a noise and deliberately turn my head, I perform an experiment to find out what I will see if I turn my head. If I ask a question not knowing what answer I will hear, I am engaging in (human!) experimentation. Roughly, experiments are actions done in order to generate observations as evidence.

There are typically differences in rigor between the experiments we perform in daily life and the experiments scientists perform in the lab, but only typically so. Sometimes we are rigorous in ordinary life and sometimes scientists are sloppy.

The epistemic value to one of an experiment depends on multiple factors in a Bayesian framework.

  1. The set of questions towards answers to which the experiment’s results are expected to contribute.

  2. Specifications of the value of different levels of credence regarding the answers to the questions in Factor 1.

  3. One’s prior levels of credence for the answers.

  4. The likelihoods of different experimental outcomes given different answers.

It is easiest to think of Factor 2 in practical terms. If I am thinking of going for a recreational swim but I am not sure whether my swim goggles have sprung a leak, it may be that if the probability of the goggles being sound is at least 50%, it’s worth going to the trouble of heading out for the pool, but otherwise it’s not. So an experiment that could only yield a 45% confidence in the goggles is useless to my decision whether to go to the pool, and there is no difference in value between an experiment that yields a 55% confidence and one that yields a 95% confidence. On the other hand, if I am an astronaut and am considering performing a non-essential extravehicular task, but I am worried that the only available spacesuit might have sprung a leak, an experiment that can only yield 95% confidence in the soundness of the spacesuit is pointless—if my credence in the spacesuit’s soundness is only 95%, I won’t use the spacesuit.

Factor 3 is relevant in combination with Factor 4, because these two factors tell us how likely I am to end up with different posterior probabilities for the answers to the Factor 1 questions after the experiment. For instance, if I saw that one of my goggles is missing its gasket, my prior credence in the goggle’s soundness is so low that even a positive experimental result (say, no water in my eye after submerging my head in the sink) would not give me 50% credence that the goggle is fine, and so the experiment is pointless.

In a series of posts over the last couple of days, I explored the idea of a somewhat interest-independent comparison between the values of experiments, where one still fixes a set of questions (Factor 1), but says that one experiment is at least as good as another provided that it has at least as good an expected epistemic utility as the other for every proper scoring rule (Factor 2). This comparison criterion is equivalent to one that goes back to the 1950s. This is somewhat interest-independent, because it is still relativized to a set of questions.

A somewhat interesting question that occurred to me yesterday is what effect Factor 3 has on this somewhat interest-independent comparison of experiments. If experiment E2 is at least as good as experiment E1 for every scoring rule on the question algebra, is this true regardless of which consistent and regular priors one has on the question algebra?

A bit of thought showed me a somewhat interesting fact. If there is only one binary (yes/no) question under Factor 1, then it turns out that the somewhat interest-independent comparison of experiments does not depend on the prior probability for the answer to this question (assuming it’s regular, i.e., neither 0 nor 1). But if the question algebra is any larger, this is no longer true. Now, whether an experiment is at least as good as another in this somewhat interest-independent way depends on the choice of priors in Factor 3.

We might now ask: Under what circumstances is an experiment at least as good as another for every proper scoring rule and every consistent and regular assignment of priors on the answers, assuming the question algebra has more than two non-trivial members? I suspect this is a non-trivial question.

Tuesday, January 28, 2025

And one more post on comparing experiments

In my last couple of posts, starting here, I’ve been thinking about comparing the epistemic quality of experiments for a set of questions. I gave a complete geometric characterization for the case where the experiments are binary—each experiment has only two possible outcomes.

Now I want to finally note that there is a literature for the relevant concepts, and it gives a characterization of the comparison of the epistemic quality of experiments, at least in the case of a finite probability space (and in some infinite cases).

Suppose that Ω is our probability space with a finite number of points, and that FQ is the algebra of subsets of Ω corresponding to the set of questions Q (a question partitions Ω into subsets and asks which partition we live in; the algebra FQ is generated by all these partitions). Let X be the space of all probability measures on FQ. This can be identified with an (n−1)-dimensional subset of Euclidean Rn consisting of the points with non-negative coordinates summing to one, where n is the number of atoms in FQ. An experiment E also corresponds to a partition of Ω—it answers the question where in that partition we live. The experiment has some finite number of possible outcomes A1, ..., Am, and in each outcome Ai our Bayesian agent will have a different posterior PAi = P(⋅∣Ai). The posteriors are members of X. The experiment defines an atomic measure μE on X where μE(ν) is the probability that E will generate an outcome whose posterior matches ν on FQ. Thus:

  • μE(ν) = P(⋃{Ai:PAi|FQ=ν}).

Given the correspondence between convex functions and proper scoring rules, we can see that experiment E2 is at least as good as E1 for Q just in case for every convex function c on X we have:

  • XcdμE2 ≥ ∫XcdμE1.

There is an accepted name for this relation: μE2 convexly dominates μE1. Thus, we have it that experiment E2 is at least as good as experiment E1 for Q provided that there is a convex domination relation between the distributions the experiments induce on the possible posteriors for the questions in Q. And it turns out that there is a known mathematical characterization of when this happens, and it includes some infinite cases as well.

In fact, the work on this epistemic comparison of experiments turns out to go back to a 1953 paper by Blackwell. The only difference is that Blackwell (following 1950 work by Bohnenblust, Karlin and Sherman) uses non-epistemic utility while my focus is on scoring rules and epistemic utility. But the mathematics is the same, given that non-epistemic decision problems correspond to proper scoring rules and vice versa.

Comparing binary experiments for non-binary questions

In my last two posts (here and here), I introduced the notion of an experiment being epistemically at least as good as another for a set of questions. I then announced a characterization of when this happens in the special case where the set of questions consists of a single binary (yes/no) question and the experiments are themselves binary.

The characterization was as follows. A binary experiment will result in one of two posterior probabilities for the hypothesis that our yes/no question concerns, and we can form the “posterior interval” between them. It turns out that one experiment is at least as good as another provided that the first one’s posterior interval contains the second one’s.

I then noted that I didn’t know what to say for non-binary questions (e.g., “How many mountains are there on Mars?”) but still binary experiments. Well, with a bit of thought, I think I now have it, and it’s almost exactly the same. A binary experiment now defines a “posterior line segment” in the space of probabilities, joining the two possible credence outcomes. (In the case of a probability space with a finite number n of points, the space of probabilities can be identified as the set of points in n-dimensional Euclidean space all of whose coordinates are non-negative and add up to 1.) A bit of thought about convex functions makes it pretty obvious that E2 is at least as good as E1 if and only if E2’s posterior line segment contains E1’s posterior line segment. (The necessity of this geometric condition is easy to see: consider a convex function that is zero everywhere on E2’s posterior line segment but non-zero on one of E1’s two possible posteriors, and use that convex function to generate the scoring rule.)

This is a pretty hard to satisfy condition. The two experiments have to be pretty carefully gerrymandered to make their posterior line segments be parallel, much less to make one a subset of the other. I conclude that when one’s interest is in more than just one binary question, one binary experiment will not be overall better than another except in very special cases.

Recall that my notion of “better” quantified over all proper scoring rules. I guess the upshot of this is that interesting comparisons of scoring rules are not only relative to a set of questions but to a specific proper scoring rule.

Monday, January 27, 2025

Comparing binary experiments for binary questions

In my previous post I introduced the notion of an experiment being better than another experiment for a set of questions, and gave a definition in terms of strictly proper (or strictly open-minded, which yields the same definition) scoring rules. I gave a sufficient condition for E2 to be at least as good as E1: E2’s associated partition is essentially at least as fine as that of E1.

I then ended with an open question as to what the necessary and sufficient conditions for a binary (yes/no) experiment to be at least as good as another binary one for a binary question.

I think I now have an answer. For a binary experiment E and a hypothesis H, say that E’s posterior interval for H is the closed interval joining P(HE) with P(H∣∼E). Then, I think:

  • Given the binary question whether a hypothesis H is true, and binary experiments E1 and E2, experiment E2 is at least as good as E1 if and only if its posterior interval for H contains the E1’s posterior interval for H.

Let’s imagine that you want to be confident of H, because H is nice. Then the above condition says that an experiment that’s better than another will have at least as big potential benefit (i.e., confidence in H) and at least as big potential risk (i.e., confidence in  ∼ H). No benefits without risks in the epistemic game!

The proof (which I only have a sketch of) follows from expressing the expected score after an experiment using formula (4) here, and using convexity considerations.

The above answer doesn’t work for non-binary experiments. The natural analogue to the posterior interval is the convex hull of the set of possible posteriors. But now imagine two experiments to determine whether a coin is fair or double-headed. The first experiment just tosses the coin and looks at the answer. The second experiment tosses an auxiliary independent and fair coin, and if that one comes out heads, then the coin that we are interested in is tossed. The second experiment is worse, because there is probability 1/2 that the auxiliary coin is tails in which case we get no information. But the posterior interval is the same for both experiments.

I don’t know what to say about binary experiments and non-binary questions. A necessary condition is containment of posterior intervals for all possible answers to the question. I don’t know if that’s sufficient.

Comparing experiments

When you’re investigating reality as a scientist (and often as an ordinary person) you perform experiments. Epistemologists and philosophers of science have spent a lot of time thinking about how to evaluate what you should do with the results of the experiments—how they should affect your beliefs or credences—but relatively little on the important question of which experiments you should perform epistemologically speaking. (Of course, ethicists have spent a good deal of time thinking about which experiments you should not perform morally speaking.) Here I understand “experiment” in a broad sense that includes such things as pulling out a telescope and looking in a particular direction.

One might think there is not much to say. After all, it all depends on messy questions of research priorities and costs of time and material. But we can at least abstract from the costs and quantify over epistemically reasonable research priorities, and define:

  1. E2 is epistemically at least as good an experiment as E1 provided that for every epistemically reasonable research priority, E2 would serve the priority at least as well as E1 would.

That’s not quite right, however. For we don’t know how well an experiment would serve a research priority unless we know the result of the experiment. So a better version is:

  1. E2 is epistemically at least as good an experiment as E1 provided that for every epistemically reasonable research priority, the expected degree to which E2 would serve the priority is at least as high as the expected degree to which E1 would.

Now we have a question we can address formally.

Let’s try.

  1. A reasonable epistemic research priority is a strictly proper scoring rule or epistemic utility, and the expected degree to which an experiment would serve that priority is equal to the expected value of the score after Bayesian update on the result of the experiment.

(Since we’re only interested in expected values of scores, we can replace “strictly proper” with “strictly open-minded”.)

And we can identify an experiment with a partition of the probability space: the experiment tells us where we are in that partition. (E.g., if you are measuring some quantity to some number of significant digits, the cells of the partition are equivalence classes under equality of the quantity up to those many significant digits.) The following is then easy to prove:

Proposition 1: On definitions (2) and (3), an experiment E2 is epistemically at least as good as experiment E1 if and only if the partition associated with E2 is essentially at least as fine as the partition associated with E1.

A partition R2 is essentially at least as fine as a partition R1 provided that for every event A in R1 there is an event B in R2 such that with probability one B happens if and only if A happens. The definition is relative to the current credences which are assumed to be probabilistic. If the current credences are regular—all non-empty events have non-zero probability—then “essentially” can be dropped.

However, Proposition 1 suggests that our choice of definitions isn’t that helpful. Consider two experiments. On E1, all the faculty members from your Geology Department have their weight measured to the nearest hundred kilograms. On E2, a thousand randomly chosen individiduals around the world have their weight measured to the nearest kilogram. Intuitively, E1 is better. But Proposition 1 shows that in the above sense neither experiment is better than the other, since they generate partitions neither of which is essentially finer than the other (the event of there being a member of the Geology Department with weight at least 150 kilograms is in the partition of E2 but nothing coinciding with that event up to probability zero is in the partition of E1). And this is to be expected. For suppose that our research priority is to know whether any members of your Geology Department are at least than 150 kilograms in weight, because we need to know if for a departmental cave exploring trip the current selection of harnesses all of which are rated for users under 150 kilograms are sufficient. Then E1 is better. On the other hand, if our research priority is to know the average weight of a human being to the nearest ten kilograms, then E2 is better.

The problem with our definitions is that the range of possible research priorities is just too broad. Here is one interesting way to narrow it down. When we are talking about an experiment’s epistemic value, we mean the value of the experiment towards a set of questions. If the set of questions is a scientifically typical set of questions about human population weight distribution, then E1 seems better than E2. But if it is an atypical set of questions about the Geology Department members’ weight distribution, then E2 might be better. We can formalize this, too. We can identify a set Q of questions with a partition of probability space representing the possible answers. This partition then generates an algebra FQ on the probability space, which we can call the “question algebra”. Now we can relativize our definitions to a set of questions.

  1. E2 is epistemically at least as good an experiment as E1 for a set of questions Q provided that for every epistemically reasonable research priority on Q, the expected degree to which E2 would serve the priority is at least as high as the expected degree to which E1 would.

  2. A reasonable epistemic research priority on a set of questions Q is a strictly proper scoring rule or epistemic utility on FQ, and the expected degree to which an experiment would serve Q is equal to the expected value of the score after Bayesian update on the result of the experiment.

We recover the old definitions by being omnicurious, namely letting Q be all possible questions.

What about Proposition 1? Well, one direction remains: if E2’s partition is essentially at least as fine as E1’s, then E2 is better with regard any set of questions, an in particular better with regard to Q. But what about the other direction? Now the answer is negative. Suppose the question is what the average weight of the six members of the Geology Department is up to the nearest 100 kg. Consider two experiments: on the first, the members are ordered alphabetically by first name, and a fair die is rolled to choose one (if you roll 1, you choose the first, etc.), and their height is measured. On the second, the same is done but with the ordering being by last name. Assuming the two orderings are different, neither experiment’s partition is essentially at least as fine as the other’s, but the expected contributions of both experiments towards our question is equal.

Is there a nice characterization in terms of partitions of when E2 is at least as good as E1 with regard to a set of questions Q? I don’t know. It wouldn’t surprise me if there was something in the literature. A nice start would be to see if we can answer the question in the special case where Q is a single binary question and where E1 and E2 are binary experiments. But I need to go for a dental appointment now.

Thursday, January 23, 2025

Recollection and two types of "Aha!" experiences

On some argumentatively central occasions, Plato refers to an intellectual “aha!” experience of seeing some point (say, something philosophical or mathematical). This is supposed to be evidence for the theory of recollection, because the experience is similar to remembering a nearly forgotten thing.

After insightful comments from students in my philosophy of mathematics seminar today, I think “aha!” experiencess come in two varieties. We might express paradigm instances of the two varieties like this:

  1. Aha! I’ve always thought this, but never quite put it into words!

  2. Aha! Now that I think about this, I see it’s got to be true!

An example of the first variety might be someone who hears about the Golden Rule, and realizes that whenever they were at their best, they were acting in accordance with it. I had a case of the second variety when I was introduced to the distributive law in arithmetic in grade three: I had never thought about whether a ⋅ (b+c) = a ⋅ b + a ⋅ c, but as soon as the question came up, with some sort of an illustrating mental picture, it was clear that it was true.

The two experiences are phenomenologically quite distinct. Type (i) experiences fit better with the Platonic picture of innate knowledge, since type (ii) experiences feel like a new acquisition rather than the recovery of something one already had. Another difference between type (i) and type (ii) experiences is that in type (ii) experiences, we not only take ourselves to have evidence for the thing being true, but the thing becomes quite unmysterious: we see how it has to be true. But type (i) experiences need not have this explanatory feature. When I have the vision of the truth of the distributive law of arithmetic, I see why it’s got to be true though I may not be able to put it into words. Not so with the Golden Rule. I can continue to be mystified by the incumbent obligations, but cannot deny them.

Literal remembering of a forgotten thing seems less like (ii) than like (i). When I remember a forgotten phone number by some prompt, I don’t have an experience of seeing why it’s got to be that.

Plato’s theory of recollection does not account for the phenomenology of type (ii) experiences. And perhaps Plato would admit that. In the Republic, he talks of “the eye of the soul”. The context there is the abilities of this life, rather than recollection. Perhaps type (ii) experiences fit more with the activity of the eye of the soul than with recollection.

At the same time, while (i) is a bit more like remembering, it’s not exactly like it, either. Remembering need not have any “I’ve thought this all along” aspect to it, which type (i) experiences tend to have. So I think neither of our “Aha!” experiences is quite like the theory of recollection leads us to. Is there a third “Aha!” experience that does? I doubt it, but maybe.

Tuesday, January 21, 2025

Competent language use without knowledge

I can competently use a word without knowing what the word means. Just imagine some Gettier case, such as that my English teacher tried to teach me a falsehood about what “lynx” means, but due to themselves misremembering what the word means, they taught me the correct meaning. Justified true belief is clearly enough for competent use.

But if I then use “lynx”, even though I don’t know what the word means, I do know what I mean by it. Could one manufacture a case where I competently use a word but don’t even know what I mean by it?

Maybe. Suppose I am a student and a philosopher professor convinces me that I am so confused that don’t know what I mean when I use the word “supervenience” in a paper. I stop using the word. But then someone comments on an old online post of mine from the same period as the paper, in which post I used “supervenience”. The commenter praises how insightfully I have grasped the essence of the concept. This someone uses a false name, that of an eminent philosopher. I come to believe on the supposed authority of this person that I meant by “supervenience” what I in fact did mean by it, and I resume using it. But the authority is false. It seems that now I am using the word without knowing what I mean by it. And I could be entirely competent.