How to not do decision theory backwards
By Anthony DiGiovanni đ¸ @ 2026-03-17T07:22 (+35)
(Subtitle: âAnd ethics, and epistemology, andâŚâ. Cross-posted from my Substack.)
We want to make decisions for good reasons. But I worry some common approaches to decision theory stray from this purpose. They start with a bottom-line verdict, âI should choose this actionâ, then use this verdict to justify claims about how to make decisions. While I feel sympathetic to this move at times, I ultimately think itâs backwards. And embedding such a move within a more nuanced methodology, like reflective equilibrium, doesnât make it less backwards.
To see what I mean, take this exaggerated example.
Imagine you ask your friend what they did over the weekend, and they say they went to a casino to play a dice game. You raise an eyebrow. Arenât they really smart? You ask why they went to a casino.
âIt just seemed super intuitive that I should play this game,â they say.
âOh, huh.â Again, your friend is really smart, so you charitably ask, âSo youâve played it a lot before and know itâs profitable?â
âNah, it was a new game.â
You blink. âUm. I guess this is your trollish way of saying: It seemed super intuitive that you should put equal probability on each possible outcome? So, somehow, the game was positive EV?â
Now they blink. âWhat? No, the rules were too complicated for me to figure out the odds. It just seemed super intuitive that I should play the game. And thatâs why it makes sense to put higher probability on the outcomes where I win, and maximize EV given those probabilities.â
This is, as I say, an exaggeration. Decision theory is supposed to help us with problems much more confusing than games of chance. Nonetheless, in those confusing problems, we might have bottom-line intuitions as powerful as âObviously I shouldnât pay in Pascalâs muggingâ, or âObviously I should cooperate in a Prisonerâs Dilemma with my copyâ. These intuitions can lead us to make the same kind of mistake as the friend above.
In brief, Iâll argue:[1]
- A bottom-line verdict about what to choose, whether in a particular case or in general, should be justified by reasons: considerations about the problem that make the chosen option worth choosing. For example, the friendâs choice whether to play the game should be justified by its expected value.
- A brute intuition in favor of such a verdict, a âverdict-level intuitionâ, isnât a reason. Roughly, this is because a verdict that I should choose A is a claim that I have reasons for A â so those reasons themselves provide the justification, if any.
- Instead of taking our verdict-level intuitions as a direct source of justification, we should do decision theory by:
- using verdict-level intuitions only to help us discover possible reasons, and
- reflecting on how intuitively compelling these reasons are on their own merits.
- The main counterargument: Our verdict-level intuitions might give us evidence of good reasons that we canât explicitly articulate. But, in contexts with poor feedback on how well these intuitions track good reasons, this evidence seems weak. And these are the contexts where verdict-level intuitions would do work beyond my methodology in (3).
These are claims about our fundamental methodology in decision theory, and in other kinds of normativity.[2] So, our views on these claims will shape our answers to many action-relevant questions. Should we take Pascalian bets? Should we act as if we control the decisions of agents in other universes? Should consequentialists only act on the near-term consequences of their actions?
Before we get our hands dirty, a few quick points of background.
Background: Intuitions as predictors vs. intuitions as normative expressions
First, it will be important to disentangle two ways we might appeal to âintuitionsâ in decision theory (both of which can be valid depending on the intuitionâs content):
- Intuitions as predictors: We might appeal to an intuition because we think itâs reliably correlated with something we want to predict. E.g., âMy very strong intuition that I should do X is evidence that, after thinking more, Iâd find good reasons for X.â
- Intuitions as normative expressions: We might report an intuition simply to express how normatively forceful or plausible we find something. E.g., âIt just seems clear that I shouldnât follow a dominated strategy.â Or, âIt seems clear that when Iâm aiming for reflective equilibrium, âcausal decision theory is correctâ should be one of my starting points.â
(If this isnât clear, see the short post âWhen do intuitions need to be reliable?â.)
Unless I say otherwise, Iâm talking about intuitions as normative expressions. But intuitions as predictors will be important later. To spoil things, Iâll argue that our intuitions about bottom-line verdicts donât do much work, either as normative expressions or predictors. However, this isnât an argument against intuitions in general. Itâs specific to intuitions that, by their nature, seem to rely on other intuitions for justification. Iâll unpack what that means soon.
Background: âWinning isnât enoughâ
Second, in this post Iâll assume the following framing of the purpose of decision theory (summarizing âWinning isnât enoughâ, by Jesse Clifton and myself):
You canât evaluate which beliefs and decision theory to endorse just by asking âwhich ones perform the best?â Because the whole question is what it means to systematically perform better, under uncertainty.
Now, you could agree with this, yet still think that âwhat it means to systematically perform betterâ includes taking pre-theoretically intuitive actions. For instance, SMK endorses this framing,[3] but takes it as an axiom that a good decision theory should recommend cooperation in the Twin Prisonerâs Dilemma. Whatâs wrong with that?
Against verdict-level curve-fitting (and reflective equilibrium)
Letâs say I reflect deeply about a bunch of decision problems, trying my hardest to correct for biases as I do so. Afterwards, for each problem or set of problems, I have a strong intuition favoring some bottom-line verdict about what to choose â so strong, perhaps, that denying it feels downright crazy. Call this my verdict-level intuition about the problem(s).
When I do decision theory, is my goal to find the best model of my verdict-level intuitions? That is, do I start with these problems and their âsolutionsâ, then look for an algorithm that returns these solutions (and generalizes well to new problems)?[4]
I donât think so. I donât take my verdict-level intuitions as normative ground truth to fit a curve to. Not even if I admit these intuitions are fallible, or curve-fit other intuitions along with verdict-level ones.
Instead, I think my goal in decision theory is to weigh up the reasons for the acts Iâm choosing between. Reasons are considerations about the decision situation that make an act worth choosing. But a verdict-level intuition isnât a reason. It doesnât give direct justification â over and above the justification that other reasons provide â for choosing the act it favors. Why? Because a verdict that I should choose A is, by its nature, nothing more than a claim that I have decisive reasons for A. So if the verdict-level intuition can justify anything, that justification has to come from those reasons. But then itâs not direct justification.
By contrast, thereâs nothing inherent to the concept of âreasonâ that claims that deeper reasons exist. So reasons can provide direct justification, if theyâre intrinsically compelling. For example, take the reason âIf A would certainly cause better consequences than B, you should prefer Aâ. Seems really intuitive! And its intuitiveness doesnât seem to clearly depend on any further intuitions.
Under my preferred methodology, then, verdict-level intuitions are still practically important. Yet their role is only to help me discover candidate reasons. For example, reflecting on Pascalâs mugging points me to the possibility that I shouldnât maximize the expectation of an unbounded utility function. From there, I should ask: Are there independently intuitive reasons not to endorse âI should maximize unbounded expected utilityâ? More generally, I figure out which reasons are independently intuitive by reflecting on questions like (more in âSo what should we do?â):
- What do I believe about the possible consequences of this act, or the duties it might violate, or the kind of character it implies?
- Which ways of making tradeoffs between uncertain consequences seem reasonable in their own right?
- What notion of âexpected consequencesâ has the most reasonable conceptual arguments behind it?
A couple key clarifications:
Iâm deliberately not calling reasons âprinciplesâ, for a couple (ahem) reasons. See footnote for details.[5] Briefly, what matters isnât how formal or general a reason is, but that a reason identifies why some action is worth choosing.
- You might think, âOf course verdict-level intuitions in themselves arenât reasons. Thatâs not the point. Theyâre still useful as reliable predictors of independently compelling reasons we havenât discovered yet.â Iâll address that view later. For now, Iâm only critiquing verdict-level intuitions as normative expressions.
My concerns apply just as well to the widely accepted methodology of reflective equilibrium. I see two interpretations of this methodology. On one view (the dominant one, as I understand it[6]), both verdict-level intuitions and independently plausible principles defeasibly and directly justify each other. This approach at least doesnât treat verdict-level intuitions as bedrock, or as infallible. But this doesnât avoid the core problem. Weâre still saying these intuitions provide direct justification, so my critique above applies, even on a coherentist account of justification.
On another view of reflective equilibrium, we only go back and forth between verdict-level intuitions and principles as a practical heuristic, to discover/predict reasons. That is, the verdict-level intuitions arenât taken as direct justification, but as a source of evidence. I have no problem with this approach as far as it goes (which, Iâll argue, isnât very far). I just think it would be clearer not to use the term âreflective equilibriumâ for this idea, since it can be conflated with the first interpretation.
In the rest of the post, Iâll give examples of verdict-level curve-fitting, and of the alternative approach, then respond to a few objections. The examples are mostly independent, so feel free to skip any that donât seem relevant.
Example: Pascalâs mugging
Consider this version of Pascalâs mugging, designed to leave out various messy confounding factors like âThe mugger is just pulling a probability out of their assâ. Someone on the street approaches you, gives you incontrovertible proof that everything they say is accurate, and says:
I can see youâre a bit of a do-gooder. I know that if you walk away, youâll make a $5000 donation to the Against Malaria Foundation. (No need to have an existential crisis about free will or anything, since Iâll wipe your memory after this.) If you give me that $5000 instead, Iâll roll a trillion-sided die. And if I roll a seven, Iâll turn the whole accessible universe into a utopia! Deal?
Iâll bet your verdict-level intuition is to not pay. Verdict-level curve-fitting says, âBased on this intuition, it seems like you just know you shouldnât pay. This suggests you shouldnât make decisions in a way that implies paying, e.g., maximizing expected utility with an unbounded utility function.â
Letâs take the most charitable interpretation of this view. When I imagine paying the mugger, and feeling like this is a mistake, Iâm not merely imagining feeling silly. Rather, it feels like I have a strong reason not to pay. From here, I have a couple options.
On one hand, I could say, âHereâs a way to explain my strong intuition against paying: I have a bounded utility function! No need to arbitrarily round tiny probabilities down to zero, give up expected utility maximization, or baldly insist on not paying.â But this doesnât seem to be enough. I havenât looked at the bounded utility function itself, to check if it seems like a strong reason. Instead, Iâve said that my bottom-line verdict about Pascalâs mugging justifies bounding my utility function. This seems backwards. Isnât my utility function supposed to represent how much I value different outcomes? And arenât these values supposed to be my reasons for my verdict?
On the other hand, I could examine the possible reason that a bounded utility function is meant to formalize, namely âTurning the universe into a utopia isnât a trillion times better than saving a lifeâ. And I could consider what makes utopias, or human lives, more or less valuable.[7] That is, I could assess this possible reason not to pay on its own merits.
Perhaps, in the end, I donât find this way of valuing outcomes independently compelling, nor any other candidate reasons Iâve thought of. And suppose I also donât expect my verdict-level intuition to reliably predict good reasons Iâm missing. If so, all Iâm left with is the brute sense that I shouldnât pay. It might be a very strong sense. But by hypothesis, Iâve already ruled out the other roles this intuition might play. After that, I struggle to see how this intuition could have any force as a normative expression.[8]
Example: Smoking Lesion and Twin Prisonerâs Dilemma
(This example assumes background knowledge of the CDT vs. EDT vs. FDT (etc.) debate.)
Like most people, I have the strong verdict-level intuition that you should smoke in Smoking Lesion. What does this tell me?
Well, the underlying reason behind the intuition seems to be âDecision-making isnât about âmanaging the newsâ, i.e., choosing acts in order to gain evidence of good outcomesâ. This is a reason because it tells me whatâs wrong with not-smoking: not-smoking doesnât seem to lead to good outcomes, per se, but to (at best) merely give evidence of them.
However, I also have a strong verdict-level intuition that I should cooperate in the Twin Prisonerâs Dilemma. The reason behind that intuition seems to be âItâs logically guaranteed that the other guy cooperates if and only if I do, even if I canât physically cause him to cooperateâ. But that also sounds like managing the news. So then I ask myself, can I make sense of how this âlogical guaranteeâ can justify my choice, without appealing to news management? If I canât do that, and the independent motivations for EDT are strong enough, perhaps I end up embracing news management after all.
Enter FDT, which says, âAs a decision-maker, âyouâ are all copies of your decision algorithm. So you actually can cause your copy to cooperate.â Okay ⌠am I an algorithm? It seems like Iâm just, well, a human. Maybe there are good independent arguments for the algorithmic ontology. But if I retrofit my ontology of myself as a decision-maker to my intuition that I should cooperate, as many FDT proponents seem to do,[9] I think that gets things backwards. My conception of what a âdecision-makerâ is should inform my verdicts about what to choose, not vice versa. Thus, Iâm wary of endorsing FDT based on my verdict-level intuition about Twin PD.
Rather, I should look at the candidate reasons my verdict-level intuition is pointing at: a) âThe relevant expected values are evidential rather than causalâ, or b) ââIâ as a decision-maker can in some sense cause outcomes without physically causing themâ. I can then ask myself how much I endorse (a) and (b) in themselves. The same goes for Smoking Lesion and âDecision-making isnât about managing the newsâ.
To its credit, my verdict-level intuition was quite useful for identifying these candidate reasons! Beyond that, though, what work is there left for this intuition to do?
Example: Cluelessness
If this acausal decision theory stuff is too speculative for you, letâs come back to Earth. Imagine you walk away from the mugger to go make your donation. Well, Mogensen (2021) argues that weâre clueless about whether donating to the Against Malaria Foundation (AMF) is better than donating to the Make-a-Wish Foundation (MAWF). The details of that argument are out of scope. Roughly, though, the idea is that the long-term consequences of these donations canât be weighed up without making arbitrary calls.
Despite his argument, Mogensen writes:
Iâm most inclined to think this is one of those cases where weâve got a philosophical argument we donât immediately know how to refute for a conclusion that we should nonetheless reject, and so we ought to infer that one of the premises must be false.
That is, he puts a lot of weight on his verdict-level intuition that donating to AMF is better. I share this intuition, even when âbetterâ is meant in an impartial consequentialist sense! But what might be the reason behind it? As far as I can tell, the candidates are:
- âThe expected total consequences of AMF, across the whole cosmos, are better than those of MAWF.â (Perhaps supported by an appeal to heuristics.)
- âThe expected total of some subset of the consequences of AMF is better than those of MAWF. And, even though Iâm clueless about the other consequences, this shouldnât override my non-cluelessness about the subset. (As per âbracketingâ.)â
Mogensen himself shows why (1) is implausible (see also this series of posts). I have some sympathy for (2). Yet my sense is that many impartial consequentialists want to say things like âAMF has higher expected value than MAWFâ, as in (1), and they donât follow a particularly coherent theory of (2).[10] At the very least, I think if we carefully examine why we think AMF is better, weâll find option (2) more defensible than (1). So weâll end up more skeptical of longtermist interventions.
In any case, if all that remains is my bare intuition in favor of AMF, that intuition tells me very little about what is âbetterâ by impartial lights. If I endorse the pro-AMF verdict above, it should be because I endorse (2) (perhaps in some vague form) as a way of weighing up my reasons for decisions. Failing that, I should admit to some non-impartial or non-consequentialist standard that gives me reason to prefer AMF. In which case, I had better actually endorse that standard in its own right!
So what should we do?
Now, weâve got our pile of candidate reasons, which weâve discovered and disentangled by thinking through various cases. How do we arrive at views on decision theory without assuming any of our choices are justified?[11] Simple. We reflect hard on these reasons themselves, asking whether theyâre well-motivated independently of verdict-level intuitions, and consistent with other well-motivated reasons.
Weâve already seen some direct analysis of candidate reasons in the examples:
- Pascalâs mugging: â... I could examine the possible reason that a bounded utility function is meant to formalize, namely âTurning the universe into a utopia isnât a trillion times better than saving a lifeâ. And I could consider what makes utopias, or human lives, more or less valuable.â
- Smoking Lesion and Twin PD: âSo then I ask myself, can I make sense of how this âlogical guaranteeâ can justify my choice, without appealing to news management?â And âOkay ⌠am I an algorithm? It seems like Iâm just, well, a human.â
- Cluelessness: âRoughly, though, the idea is that the long-term consequences of these donations canât be weighed up without making arbitrary calls.â
Here are more examples of this kind of methodology, i.e., decision-theoretic arguments that donât depend on any appeals to verdict-level intuitions (though I donât necessarily endorse them in full):
- In âCan you control the past?â (Sec. III), Carlsmith argues that EDTâs âmanaging the newsâ seems less objectionable if we reject libertarian free will. Yes, as the CDTist insists, decision theory should be about making a difference in some sense. But if the past and future are both fully determined by physics, why is it any stranger to regard yourself as making a difference to the past than to the future? (See Clifton for an alternative to EDT thatâs also grounded in this kind of reasoning.)
Spohn (2012) argues for one-boxing in Newcombâs problem, not because itâs highly intuitive upon reflection, but because: When you face the boxes and deliberate about what to do, instead of âdecidingâ at that time, youâre learning what you had decided all along â and so even CDT recommends one-boxing. And he tries to show that this interpretation of the decision situation is plausible.[12]
- Cluelessness entails incomplete preferences. Tarsney et al. (2025) argue against incompleteness by giving independent motivations for two principles that are incompatible with incompleteness. As I discuss in the conclusion, though, I think this kind of argument still goes backwards in a different way.
- To argue for bounded utility functions, Russell and Isaacs (2021) donât merely appeal to the intuition against paying in Pascalâs mugging. Instead, they argue that applying a risk-neutral decision rule to an unbounded utility function violates a form of the sure-thing principle.
Objections and responses
Donât our âreasonsâ also require further justification, by your standard? And then those reasons, and so on?
Response: Some reasons do indeed seem to need justification. For example, if an impartial consequentialistâs reason to donate to AMF is that this donation has positive EV, we should ask why they have credences under which the EV is positive.
That doesnât mean my view leads to an infinite regress. While I canât do justice to the whole foundationalism vs. coherentism debate that this touches on (see Huemer (2022, ch. 4-5) or this summary), two key points will suffice.
First, suppose you think all reasons need justification, contra foundationalism. This is consistent with my core claim, because you could still think that reasons justify each other via their mutual coherence. Itâs just that there canât be mutual justification between reasons and verdict-level intuitions, for the reasons Iâve already argued.
Second, if instead we accept foundationalism, it seems very intuitive that some reasons donât need further justification. (Again, my previous arguments show why this isnât true of verdict-level intuitions. See footnote[13] for more on how my view relates to foundationalism generally.) Imagine that the friend from the intro story had said:
The game is simple. You pay $1, and the house pays you $3 if a six-sided die comes up even. I have plenty of money, and it just seems morally obvious that I should be scope-sensitively altruistic. So I should value more money linearly if Iâm only playing a few games. By the principle of indifference, it seems arbitrary to put anything other than 1/6 probability on each possible side of the die. It also seems arbitrary for my decision rule not to give equal weight to equally likely possible outcomes. So I should play the game because its EV is positive.
This casino would go bankrupt in a heartbeat, but set that aside. The friend seems to have given a sufficient justification for their choice, more or less. Itâs not an infallible justification, e.g., maybe infinite ethics weirdness calls that last step about the decision rule into question. But foundational justifications can be fallible (again, see Huemer (2022, ch. 4)).
What if our verdict-level intuitions are tracking good reasons that we canât articulate?
For example, in the Twin PD case, perhaps weâre missing other reasons to cooperate, besides âThe relevant expected values are evidentialâ and âI can âlogically causeâ certain outcomesâ. That seems pretty likely, given how subtle decision theory is. I might, then, appeal to my verdict-level intuition as a predictor: Perhaps this very strong intuition is evidence that, if I had a lot more time to think, the reasons I canât yet articulate would on balance favor this verdict.
Response: To summarize, the problem is that I have little reason to think my verdict-level intuitions track good reasons on net â more so than bad reasons, or various spurious things orthogonal to the good reasons. At least, this seems true when weâre aiming to predict reasons that weâre missing after having already thought a lot. So if someone claims that their verdict-level intuitions are strong evidence about the balance of reasons, I think they have the burden of proof.
Crucially, our intuitions about reasons themselves donât face this critique, because those are intuitions as normative expressions, not as predictors. Itâs a category error to ask about the âreliabilityâ of intuitions that arenât meant to make predictions (see this post for more). Take the principle of indifference example from the previous section. When I say, âIt seems arbitrary to put unequal probabilities on symmetric outcomesâ, Iâm not predicting anything. Iâm saying that such arbitrariness is unacceptable on its face.
Letâs unpack this. I see two general ways I might argue that my intuition-as-predictor is strong evidence of something, as in the bolded claim above. And neither seems to work in this case.
First, I could say, âIâve gotten rich feedback about the answers to many similar questions. So probably my intuition about this question is reliable (even if I havenât explicitly verified its reliability).â Here, the âsimilar questionsâ are past decision problems. And the âanswersâ are the reasons Iâve discovered, underlying my verdicts in those problems.
But whether we have indeed gotten rich feedback depends on the kinds of reasons weâre trying to predict:
- We might have a lot of experience discovering the easy-to-articulate reasons behind our intuitions â the low-hanging fruit, like âDecision-making isnât about managing the newsâ. For easy-to-articulate reasons, we have tight feedback loops between first encountering the decision problem, noticing our verdict-level intuition, and discovering the reasons. This is true even if the reasons take much longer to formalize.
- By contrast, we have much less experience discovering the hard-to-articulate good reasons. These are the reasons we only discover, even in vague form, after years of being in the limbo of âThis feels right, but I canât put my finger on whyâ. So the feedback loops are much weaker. And weâre in that limbo right now, when we appeal to our verdict-level intuitions as predictors. Itâs telling that I canât think of many examples of hard-to-articulate reasons! Iâm curious if the reader can.
Second, I might say, âIâve verified that my intuitive guesses of answers to many similar questions were usually correct (perhaps after some âcalibration trainingâ). So probably my intuitive guess about this question is correct.â
I would update on well-designed empirical studies of this. Paying attention to the track record of oneâs intuitions about missing reasons, especially the hard-to-articulate ones, seems like a promising way to become a âphilosophy superforecasterâ. But I havenât seen anyone point to such a track record to justify their verdicts.
Overall, once Iâve probed my verdict-level intuition for the more obvious candidate reasons, I donât think I should expect this intuition to contain further dormant wisdom.
Concluding thoughts on further implications
Iâve argued that our bottom-line verdicts need to be justified by deeper reasons. Weâve seen that this matters in cases like Pascalâs mugging, acausal decision theory puzzles, and cluelessness. But Iâve defended a more general methodology here: âLook at how the structure of justification depends on the content of what weâre justifying.â So I expect this point to matter to a wide range of questions in decision theory and other fields of philosophy. Hereâs a quick preview of its potential implications.
First, just as thereâs a direction of justification from reasons to verdicts, thereâs also a direction from different kinds of reasons to others. The reasons for our choices are our preferences under uncertainty (among other things). In turn, the reasons for our preferences include our values and beliefs.
Take the example of Tarsney et al. (2025) from âSo what should we do?â. Their argument comes down to: âIncomplete preferences violate a very intuitive principle about how to make choices across multiple (hypothetical) situations. So we should reject these preferences.â I definitely feel the force of the intuitions theyâre appealing to. But this argument doesnât seem to tell us whatâs wrong with the preferences themselves, or the values and beliefs that motivate them, in a given choice situation. It doesnât tell us what about option A makes it preferable to option B. It just tells us that if we have incomplete preferences, the pattern of choices weâll make across multiple situations is supposedly problematic. I donât think we should retrofit our preferences to intuitively âprincipledâ patterns of choices like this. (Wouldnât my reasoning here undermine most money pump and Dutch book arguments? Thereâs a lot more to say on this, but arguably, it would!) See this appendix for more.
Second, just like in decision theory, arguably we need deeper reasons for our verdicts in ethics, epistemology, and more. Why should we count as a datum the brute intuition that, say, the repugnant conclusion is unacceptable? Or that we know weâre not Boltzmann brains? Iâm not concluding that we should reject these intuitive verdicts, here. All I ask is that we go beyond âone personâs modus ponens is anotherâs modus tollensâ, or âMoorean shiftâ. Instead, letâs look much more closely at the structure of our intuitions, and what that says about the justificatory work different intuitions can do.
Acknowledgments
Many thanks to Jesse Clifton for various past discussions that have helped me clarify my views presented here (this doesnât imply his endorsement of all my claims). Thanks also to Lukas Finnveden, Michael St. Jules, Sylvester Kollin, Niels Warncke, Joseph Ancion, Brandon Sayler, Francis Rhys Ward, Eric Olav Chen, Clare Harris, and Claude for helpful comments.
- ^
Compare these claims to the views expressed in McMahan (2013, p. 114), Nye (2015, pp. 628-630), and Singer (1974). However, these authors can be read as arguing that general principles take priority over intuitions about particular cases. As I note in âAgainst verdict-level curve-fitting (and reflective equilibrium)â, my argument is not about âparticular vs. generalâ but about âintuitions about the bottom line vs. reasons for that bottom lineâ. Though, the authors do gesture at the latter to some extent.
- ^
Iâve written about some such questions in previous work, most notably here, here, and here.
I could just as well have called this post âHow to not do normativity backwardsâ, encompassing things like ethics and epistemology as well. I decided to focus on decision theory here both to keep the scope manageable, and because it sounds less highfalutin than ânormativityâ. But see the conclusion for brief thoughts on other domains.
- ^
Quote: âAs a final note, I think it is important to not delude ourselves with terms like âsuccessâ, âfailureâ, âwinsâ and âpreferableâ (which I have used in this post) in relation to decision theories; UEDT and UCDT are both the âcorrectâ decision theory by their own lights (just as all decision theories): the former maximises expected utility with conditionals from the earlier logical vantage point, and the latter does the same but with logical counterfactuals, and that is thatâthere is no objective performance metric. See The lack of performance metrics for CDT versus EDT, etc. by Caspar Oesterheld for more on this.â
- ^
Iâve encountered this view in a few personal communications. Some public examples, besides those in footnote 9 (emphasis mine):
- Briggs (2010): âUnlike EDT, CDT yields the right answer in The Smoking Lesion. ⌠But several authors, Eells (1985), Bostrom (2001), Egan (2007), have formulated examples where EDT gets things right, while CDT gets things wrong.â
- Demski: âI give a variant of the smoking lesion problem which overcomes an objection to the classic smoking lesion, and which is solved correctly by CDT, but which is not solved by updateless EDT. ⌠If CDT fails Newcomb's problem and EDT doesn't, it seems CDT is at best a hack which repairs some cases fitting the above description at the expense of damaging performance in other cases.â
- Treutlein: âThe Coin Flip Creation problem is intended as an example of a problem in which EDT would give the right answer, but all causal and logical decision theories would fail.â
- ^
First, âprinciplesâ connotes something formal or precise. Yet a reason can be as vague as âRational decision-making is about causing good outcomesâ, or âIt seems irrational to be dynamically inconsistentâ. Conversely, FDTâs recommendations match many of my verdict-level intuitions (letâs say), and itâs a formal decision theory. But I donât think this makes FDT a good decision theory. Also, see footnote 11.
Second, âprinciplesâ connotes generality or simplicity. Yet we can have reasons with respect to particular cases, and these reasons might be complex. Indeed, compared to many rationalist curve-fitters, Iâm much more comfortable with certain supposedly âcomplexâ moves. E.g., to avoid money pumps or Dutch books, I can augment my beliefs, values, and decision theory with certain commitments, instead of endorsing some new beliefs, values, or decision theory. And this isnât ad hoc, because such commitments are desirable according to my current beliefs, values, and decision theory! (See here and here.)
- ^
Examples of reflective equilibrium-sympathetic writings that support this interpretation, i.e., âverdict-level intuitions are treated as giving direct justificationâ:
- From the SEP article (emphasis mine): âThe method of reflective equilibrium has been advocated as a coherence account of justification [âŚ]. For example, a moral principle or moral judgment about a particular case [âŚ] would be justified if it cohered with the rest of our beliefs about right action (or correct inferences) on due reflection and after appropriate revisions throughout our system of beliefs.â
- In Rechnitzerâs (2022) reflective equilibrium analysis of the trolley problem(s), she considers several bottom-line verdicts (âinput commitmentsâ), e.g., âIC 3 In Case 3, the bystander at the switch may divert the trolley so that one workman dies instead of five.â She then says (emphasis mine): âThe second strategyâreversing the commitment, accepting that the bystander must not turn the trolley, and instantly establishing consistencyâlooks like the least laborious solution. However, it cannot be vindicated by the RE criteria at this stage. The problem is that this conflicts with the criterion to respect input commitments. For Thomson, the bystander commitment, IC 3, has a high independent credibility: she has a strong intuition that the bystander may divert the trolley âŚâ
- ^
Under orthodox expected utility theory, this isnât how we construct utility functions. The orthodox approach takes our preferences under uncertainty as given, and derives the utility function from those preferences via a representation theorem. So much the worse for the orthodox approach, then! See Peterson (2008), Meacham and Weisberg (2011), and Easwaran (2014) for more thorough arguments.
- ^
I can hear the distant keyboards clacking as readers type, âOkay, hereâs my bank account, $5000 please.â This is a separate conversation, one I have complicated thoughts on. Basically, though, I think in more realistic cases than this one, there plausibly are good reasons not to pay: a combination of my response here, the idea of âbracketingâ, and accounting for perverse incentives set by paying actual muggers. If I didnât buy those reasons, perhaps I still wouldnât pay, but if I were honest Iâd admit this was simply irrational.
- ^
Examples:
- Levinstein and Soares (2017), who say, âIn brief, FDT suggests that you should think of yourself as instantiating some decision algorithmâ, and then primarily argue for FDT by showing that it gets various cases ârightâ;
- comment by Baker;
- the last paragraph of the section âObjection: comparing ontologies by comparing decision theoriesâ in SMKâs âFDT is not directly comparable to CDT and EDT. To be clear, SMK themselves rejects this approach: â[I]t is arguably reasonable to think of ontology as something that is not fundamentally tied to the decision theory in and of itself. Instead, we might want to think of decision theories as objects in a given ontology, meaning we should decide on that ontology independentlyâor so I will argue.â
- ^
See Lewis for an example of the âAMF has higher expected value than MAWFâ view.
As far as Iâm aware, Kollin et al.âs (2025) âbracketingâ is the only theory that attempts to make (2) precise. Unfortunately, bracketing has a pretty significant open problem. Also, in unpublished work, I show that bracketing violates the sure-thing principle. Though Iâm undecided as to whether this is a huge problem, compared to the independent motivations for bracketing.
- ^
Apparently, in the context of epistemology, this question is a classic challenge to âmethodismâ (as opposed to âparticularismâ): âHow do we arrive at reasonable criteria for knowledge without assuming we know any particular facts?â. My perhaps-naĂŻve feeling is that this isnât so challenging at all, since my response in the next sentence seems kind of obvious to me. So I worry I could be missing something.
(Iâm not sure whether my view in this post is inconsistent with particularism, though. Reasons can be particular in a sense, e.g., âItâs logically guaranteed that the other guy cooperates if and only if I doâ. But the paradigm examples of âparticularsâ I usually see in introductions to particularism vs. methodism seem to be verdict-level intuitions.)
- ^
From p. 103 (emphasis mine):
Perhaps, you start deliberating on the matter only when standing before the boxes, because you are informed about the plot only then. This does not necessarily mean, though, that you are deciding only then. Perhapsâindeed, this is what I am suggestingâyou were committed to one-boxing all along, and by deliberating you discover to be so committed all along. In any case, this is the only way how [the model of Newcombâs problem in which your decision is causally upstream of the prediction] makes sense: You are decided early enough to one-box, simply by being rational, and this influences the predictorâs prediction, presumably simply by his observation of your consistent and continuous rationality. [...] Being committed or decided all along without ever having reflected on the matter? This sounds strange, and this may be the weak part of my account of NP, but, as I would insist, the only weak part [...] On the other hand, it is not so strange perhaps. You will grant that you have many beliefs without ever having reflected on them, for instance, about the absence of ice bears in Africa. You have never posed the question to yourself; but for a long time your mind is fixed how to respond. Similarly, I trust, your introspection will reveal that often your reflection does not issue in a decision, but rather finds that you were already decided or committed.
- ^
Some philosophers evidently do take certain verdict-level intuitions as foundational. E.g., Huemer seems to endorse putting a lot of weight on verdict-level intuitions-as-normative-expressions, based on (e.g.) his objections to utilitarianism.
One reason we might be tempted to consider some verdict-level intuitions foundational is by analogy to sense perception. Arguably, if you see a table in front of you, your visual perception is foundational justification for believing thereâs actually a table, absent defeaters. (So says phenomenal conservatism.) Iâm happy to grant that here. But the disanalogy is that, as Iâve said, a verdict just is a claim that you have reason to choose some act. This isnât true of a perception of a table. In particular, my arguments in âAgainst verdict-level curve-fitting (and reflective equilibrium)â are defeaters of the verdict-level intuition.
SummaryBot @ 2026-03-17T21:21 (+4)
Executive summary: The author argues that decision theory should not start from strong intuitions about what one should choose and then justify them, but instead should ground choices in independently compelling reasons, using verdict-level intuitions only to help discover those reasons.
Key points:
- The author claims that a âverdict-level intuitionâ (a brute sense that one should choose a particular action) is not itself a reason, because such a verdict already presupposes that there are underlying reasons for that choice.
- They argue that decision theory should proceed by identifying candidate reasons suggested by intuitions and then evaluating those reasons on their own merits, rather than treating intuitions as direct justification.
- The author contends that reflective equilibrium, when interpreted as allowing mutual justification between intuitions and principles, still relies on the same mistaken use of verdict-level intuitions as justificatory.
- In cases like Pascalâs mugging, the correct method is to assess reasons such as whether utility should be bounded, rather than inferring those reasons from the intuition not to pay.
- The author argues that verdict-level intuitions are weak as predictors of unarticulated good reasons, especially in domains with poor feedback and where hard-to-articulate reasons are involved.
- They suggest that this methodological point generalizes beyond decision theory to ethics and epistemology, where brute intuitions about conclusions should likewise be replaced with analysis of underlying reasons.
This comment was auto-generated by the EA Forum Team. Feel free to point out issues with this summary by replying to the comment, and contact us if you have feedback.