Is the Conjunction Fallacy Irrational? Take a Guess.

Kevin Dorst
Science and Philosophy
10 min readJul 18, 2020

(This post is co-written with Matt Mandelkern, based on our joint paper on the topic.)

It’s February 2024. Three Republicans are vying for the Presidential nomination, and FiveThirtyEight puts their chances at:

  • Mike Pence: 44%
  • Tucker Carlson: 39%
  • Nikki Haley: 17%

Suppose you trust these estimates. Who do you think will win?

Some natural answers: “Pence”; “Pence or Carlson”; “Pence, Carlson, or Haley”. In a Twitter poll earlier this week, the first two took up a majority (53.4%) of responses:

​But wait! If you answered “Pence”, or “Pence or Carlson”, did you commit the conjunction fallacy? This is the tendency to say that narrower hypotheses are more likely than broader ones — such as saying that P&Q is more likely than Q––contrary to the laws of probability. Since every way in which “Pence” or “Pence or Carlson” could be true is also a way in which “Pence, Carlson, or Haley” would be true, the third option is guaranteed to be more likely than each of the first two.

Does this mean answering our question with “Pence” or “Pence or Carlson” was a mistake?

We don’t think so. We think what you were doing was guessing. Rather than simply ranking answers for probability, you were making a tradeoff between being accurate (saying something probable) and being informative (saying something specific). In light of this tradeoff, it’s perfectly permissible to guess an answer (“Pence”) that’s less probable — but more informative — than an alternative (“Pence, Carlson, or Haley”).

Here we’ll argue that this explains — and partially rationalizes — the conjunction fallacy.

Good Guesses

We make guesses whenever someone poses a question and we can’t be sure of the answer. “Will it rain tomorrow?”, “I think it will”; “What day will the meeting be?”, “Probably Thursday or Friday”; “Who do you think will win the nomination?”, “I bet Pence will”; and so on.

What sorts of guesses are good guesses? The full paper argues that there are a variety of robust and intricate patterns, drawing on a fascinating paper by Ben Holguín. Here we’ll just focus on the main patterns in our lead example.

Suppose you have the probability estimates from above (Pence, 44%; Carlson, 39%; Haley, 17%), and we ask you: “Who do you think will win?” As we’ve seen, a variety of answers seem reasonable:

(1) “Pence”✓
(2) “Pence or Carlson”✓
(3) “Pence, Carlson, or Haley” ✓

In contrast, a variety of answers sound bizarre:

(4) “Carlson” ✘
(5) “Carlson or Haley ( = “Not Pence”) ✘
(6) “Pence or Haley” ✘

We’ve run examples like this by dozens of people, and the judgments are robust — for instance, in a similar Twitter poll in which “Pence or Haley” was an explicit option, it was the least-common answer (6.7%):

What’s going on? How do we explain why (1)–(3) are good guesses and (4)–(6) are bad ones?

Our basic idea is a Jamesian thought: making good guesses involve trading off two competing goals. On the one hand, we want to avoid error — to be accurate. On the other, we want to get at the truth — to be informative. These two goals compete: the more informative your guess, the less likely it is to be true. A good guess is one that optimizes this tradeoff between accuracy and informativity.

More precisely, we assume that guesses have answer values that vary with their accuracy and informativity. True guesses are better than false ones, and informative true guesses are better than un informative true ones. Given that, here’s our proposal:

Guessing as Maximizing: In guessing, we try to select an answer that has as much answer-value as possible — we maximize expected answer-value.

To make this precise, we can clarify the notion of informativity using a standard model of a question. Questions can be thought of as partitioning the space of open possibilities into a set of complete answers. For example: the set of complete answers to “Will it rain tomorrow?” is {it will rain, it won’t rain}; the set of complete answers to “Who will win the nomination?” is {Pence will win, Carlson will win, Haley will win}; and so on.

In response to a question, a guess is informative to the extent that it rules out alternative answers. Thus “Pence” is more informative than “Pence or Carlson”, which in turn is more informative than “Pence, Carlson, or Haley”.

Given this, we can explain why (4)–(6) are bad guesses. Consider “Carlson”. It’s exactly as informative as “Pence” — both rule out 2 of the 3 possible complete answers — but it is less probable: “Carlson” has a 39% chance of being true, while “Pence” has a 44% chance. Thus if you’re trying to maximize expected answer value, you should never guess “Carlson”, since “Pence” is equally informative but more likely to be accurate.

Similarly, consider “Pence or Haley”. What’s odd about this guess is that it “skips” over Carlson. In particular, if we swap “Haley” for “Carlson”, we get a different guess that’s equally informative but, again, more probable.

On the other hand, we can see why (1)–(3) can all be good guesses. The basic point: if you really care about being informative, you should choose a maximally specific answer (“Pence”); if you really care about being accurate, you should choose a maximally likely answer (“Pence, Carlson, or Haley”); and intermediate ways of weighting these constraints lead to good guesses at intermediate levels of informativity (“Pence or Carlson”).

(For a formal exposition of all this, see the Appendix or the paper.)

The conjunction fallacy

With this account of guessing in hand, let’s apply it to our opening observation: guessing leads to the conjunction fallacy.

Recall: this is the tendency to rate narrower hypotheses (like “P&Q”) as more probable than broader ones (like “Q”). It’s the star of the show in the common argument that people’s reasoning is systematically irrational, lacking an understanding of the basic rules of probability and instead using simple heuristics.

The most famous example is from the original paper by Tversky and Kahneman:

Linda is 31 years old, single, outspoken and very bright. She majored in philosophy. As a student, she was deeply concerned with issues of discrimination and social justice, and also participated in anti-nuclear demonstrations. Which of the following is more likely?

​a)Linda is a bank teller and is active in the feminist movement.
b) Linda is a bank teller.

A majority of subjects said “feminist bank teller” (FT) was more probable than “bank teller” (T). Again, this violates the laws of probability: every possibility in which Linda is a feminist bank teller is also one in which she’s a bank teller — but not vice versa.

What’s going on here? Our proposal:

Answer-Value Account: People commit the conjunction fallacy because they rank answers according to their quality as guesses (their expected answer-value), rather than their probability of being true.

​In other words, we think the Linda case is analogous to the following one:

It’s 44% likely that Pence will win, 39% likely that Carlson will, and 17% likely that Haley will. Which of the following are you more inclined to guess?

a) Pence or Carlson will win.
​b) Pence, Carlson, or Haley will win.

Given the above survey, we can expect that around half of people would choose (a). (Since 35/(35 + 38.2) = 47.8%.) The crucial point is that in both the Linda and Pence cases, option (a) is less probable but more informative than option (b) with respect to the salient question — e.g. “What are Linda’s social and occupational roles?” or “Who will win the nomination?”

In particular, our model of expected answer value predicts that you should rate “feminist bank teller” as a better guess than “bank teller” whenever you’re sufficiently confident that Linda is a feminist given that she’s a bank teller — whenever P(F|T) is sufficiently high, where the threshold for “sufficient” is determined by how much you value being informative (see the Appendix).

Why is this conditional probability P(F|T) what matters? Because although the probability of “feminist bank teller” is always less than that of “bank teller”, how much less is determined by this conditional probability, since P(FT) = P(T)•P(F|T). Thus when P(F|T) is high, switching from “bank teller” to “feminist bank teller” has only a small cost to accuracy — which is easily outweighed by the gain in informativity. Our account therefore makes the following prediction:

Prediction: Rates of ranking the conjunction AB as more probable than the conjunct B will tend to scale with P(A|B).

This prediction is borne out by the data; a clean example comes from Tentori and Crupi (2012). They give a vignette in which they introduce Mark, and then say that he holds X tickets in a 100-ticket raffle — where X varies between 0 and 100 across different experimental conditions. They then ask subjects which of the following is more likely:

a) Mark is a scientist and will win the lottery.
b) Mark is a scientist.

The rates of ranking “scientist and will win the lottery” (WS) as more likely than (or equally likely as) “scientist” (S) scaled directly with the number of tickets Mark held, i.e. with the probability that Mark wins the lottery given that he’s a scientist, P(W|S) (which equals P(W), since S and W are independent):

This is exactly what the answer-value account predicts.

​So is the conjunction fallacy irrational?

Suppose the answer-value account is right: people commit the conjunction fallacy when they rate answers for their quality as guesses rather than for their probability of being true. What would that imply about the conjunction fallacy, and its role in the debate about human rationality more broadly?

The answer is subtle.

One the one hand, it turns out that questions like “What do you think is (most) likely?” are standard ways of eliciting guesses — which in turn have a very different normative profile than probability judgments. For example, when we asked “Who do you think will win?” in our opening question, answering “Pence” is not irrational — nor would it be so if we tweaked the wording to “What’s most likely?”; “What do you bet will happen?”; etc. These are all ways of eliciting guesses. (Note: our second Twitter poll used “What do you think’s likely to happen?”)

Since these prompts are standard ways of eliciting answers in conjunction-fallacy experiments, this complicates our assessment of such answers. The pragmatic upshot of the question that’s being asked is not a literal question about probability: people will hear these questions are requests to guess––to trade off probability against informativity — rather than to merely assess probability. And reasonably so.

On the other hand, using such guesses to guide your statements and actions can lead to mistakes. This is clearest in experiments that elicit the conjunction fallacy while asking people to bet on options. Of course, “What do you bet will happen?” is a natural way of eliciting a guess in conversation (“I bet Pence’ll win”). Nevertheless, if we actually give you money and you choose to let it ride on “Pence or Carlson” rather than “Pence, Carlson, or Haley”, then you’ve made a mistake. Moreover, experiments show that people do have a tendency to bet like this — though the rates of the fallacy diminish somewhat.

In scenarios like this, the conjunction fallacy is clearly a mistake. The crucial question: What does this mistake reveal about human reasoning?

If our account is right, it does not reveal — as is commonly claimed — that human judgment works in a non-probabilistic way. After all, what’s happening is that people are guessing and then acting based on that guess — and guessing requires an (implicit) assessment of probability. Instead, the conjunction fallacy reveals that people are bad at pulling apart judgments about pure probability from a much more common type of judgment — the quality of a guess.

Why are people bad at this? Our proposal: because guessing is something we do all the time. Moreover, it’s something that makes sense to do all the time. We can’t have degrees of belief about all possibly-relevant claims — no system could, since general probabilistic inference is intractable. So instead, we construct probability judgments about the small set of claims generated by the question under discussion, use them to formulate a guess, and then reason within that guess.

There’s empirical evidence that people do this. For example: poker players decide what to bet by guessing what hands their opponents might have; doctors decide what tests to perform by guessing their patients’ ailments; scientists decide what experiments to run by guessing which hypothesis is true; and so on.

People do this, we think, because probability alone doesn’t get you very far. The most probable answer is always, “Something will happen, somewhere, sometime.” Such certainly-true answers don’t help guide our actions — instead, we need to trade off such certainty for some amount of informativity.

If this is right, the error revealed by the conjunction fallacy is in some ways like that revealed by the Stroop test. Watch the following video and try — as quickly as possible — to say aloud the color of the text presented (do not read the word):

It’s hard! And the reason it’s hard is that it requires doing something that you don’t normally do: assess the color of a word without reading it. Yet throughout most of life, what you do when presented with a word — what makes sense to do — is read it. In short: a disposition that involves sophisticated processing, and is rational in general, can lead to errors in certain cases.

Likewise, it’s hard not to commit the conjunction fallacy because that requires doing something that you don’t normally do: assess the probability of an uncertain claim without assessing it as a guess. Yet throughout most of life, what you do when presented with such a claim — what makes sense to do — is assess its quality as a guess. In short: a disposition that involves sophisticated processing, and is rational in general, can lead to errors in certain cases.

Upshot: although the conjunction fallacy is sometimes a mistake, it is not a mistake that reveals deep-seated irrationality. Instead, it reveals that when forming judgments under uncertainty, we need to trade off accuracy for informativity — we need to guess.

What next?
If you enjoyed this post, please consider retweeting it, following us on Twitter (Kevin, Matt), or signing up for the newsletter. Thanks!
If you’re interested in the details, including other potential applications of guessing to epistemology, philosophy of language, and cognitive science, check out the full paper.
If you want to learn more about the conjunction fallacy, Tversky and Kahneman’s original paper is fantastic, as is this 2013 paper by Tentori et al. — which provides a good overview as well as its own interesting proposal and data.

Originally published at https://www.kevindorst.com.

--

--

Kevin Dorst
Science and Philosophy

Philosopher at University of Pittsburgh, working on the question of how irrational we truly are.