Human Reasoning
Human Reasoning
Human Reasoning
Human
types of mental processing, one rapid and intuitive and shared
with other animals, and the other slow and reflective and more
characteristic of human beings. Finally, we discuss implications
Reasoning
for human rationality.
HUMAN REASONING
David E. Over
University of Durham
Jonathan St B. T. Evans
University of Plymouth
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Shaftesbury Road, Cambridge CB2 8EA, United Kingdom
One Liberty Plaza, 20th Floor, New York, NY 10006, USA
477 Williamstown Road, Port Melbourne, VIC 3207, Australia
314–321, 3rd Floor, Plot 3, Splendor Forum, Jasola District Centre,
New Delhi – 110025, India
103 Penang Road, #05–06/07, Visioncrest Commercial, Singapore 238467
www.cambridge.org
Information on this title: www.cambridge.org/9781009495363
DOI: 10.1017/9781009495349
© David E. Over and Jonathan St B. T. Evans 2024
This publication is in copyright. Subject to statutory exception and to the provisions
of relevant collective licensing agreements, no reproduction of any part may take
place without the written permission of Cambridge University Press & Assessment.
When citing this work, please include a reference to the DOI 10.1017/9781009495349
First published 2024
A catalogue record for this publication is available from the British Library.
ISBN 978-1-009-49536-3 Hardback
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
DOI: 10.1017/9781009495349
First published online: May 2024
David E. Over
University of Durham
Jonathan St B. T. Evans
University of Plymouth
Author for correspondence: David E. Over, [email protected]
1 Introduction 1
2 Deductive Reasoning 2
4 Non-deductive Reasoning 23
5 Conditionals 31
8 Overall Conclusions 60
References 62
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Human Reasoning 1
1 Introduction
Reasoning is the inferring of conclusions from premises, and this Element is
an account of what psychologists have discovered about human reasoning.
We have not found it hard to select topics for it. Two broad contemporary
themes in the psychology of reasoning have been influenced by logicians and
philosophers and clearly raise philosophical questions.
The first of these themes is the emergence of a new paradigm in the
psychology of reasoning, replacing with Bayesian subjective probability
theory the traditional presupposition that classical logic and its conditional
set the normative standard for human reasoning. This theme is primarily found
in Sections 2–5. Section 5 has at its core the psychological findings on the
hypothesis that the probability of the natural language conditional, P(if p then q),
is the conditional probability of q given p, P(q|p). This hypothesis was first stated
by philosophical logicians and has long been of great interest to them, and
its confirmation in psychological experiments, for many conditionals, in turn
supports the new Bayesian paradigm. More generally, logicians and philosophers,
going back at least to de Finetti and Ramsey, have substantially influenced the
new paradigm, as will become clear in our references.
The second theme is research in psychology on dual-process theories of the
mind. These state that there are two types of mental processes, one rapid and
intuitive and shared with other animals, and the other slow and reflective and
more characteristic of human beings. Section 6 is on this topic, explaining how
the psychological understanding of it has been deepened and refined over time.
Philosophers have helped in this process, and no doubt will continue to do so.
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
2 Deductive Reasoning
The importance of classical logic (Shapiro & Kissel, 2022) is widely recog-
nised. It is binary in having truth and falsity as its only truth values, and the
conclusions of its valid inferences are true assuming that the premises are true.
Psychologists working on human thinking and reasoning in the mid twentieth
century presupposed that classical validity was the only logical standard for
human rationality, and this had a profound influence on the psychology of
reasoning. It led to hundreds of published experiments using the deduction
paradigm (Evans, 2002) based on classical logic. To contrast it with later
developments, we will sometimes call this the traditional approach or para-
digm. This field of work was directly or indirectly inspired by philosophical
writing, and it is of importance that philosophers be aware of the findings that
were reported. A detailed review of this period of work is in Evans et al. (1993)
and a brief summary follows in this section. This traditional approach domin-
ated the field until around 1990, when a number of authors started to call for
a different approach, now known as the ‘new paradigm psychology of reason-
ing’ (Oaksford & Chater, 2020; Over, 2009, 2020). This new approach made
degrees of belief and so subjective probability judgements central to the study of
human reasoning. The relevant normative standards for it are probability theory
and probability logic (Demey et al., 2023).
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
tive accounts (see Section 7) was not addressed by psychologists until most of
the deduction studies had been published (but see Elqayam & Evans, 2011,
Oaksford & Chater, 2007, and Stanovich, 1999). The great bulk of studies in
the deduction paradigm have used Aristotelian syllogisms or conditional
inferences, and we will briefly summarise this work and its main findings in
Sections 2.2–2.6.
This argument is clearly invalid. It is possible that some psychologists are also
astronomers but there is no necessary reason why they should be, given the
premises. However, participants are much more likely to say this argument is
valid than invalid when it is given in an abstract form, such as Some A are B, All
C are A, therefore Some C are B.
The premises and conclusions may each take one of four different forms,
classically known as A (All A are B), E (No A are B), I (Some A are B) and
O (Some A are not B). The mood of the syllogisms is determined by these forms.
For example, the above argument has the mood IAI. Syllogisms can also have
figures, depending on how the terms are arranged. The above syllogism is in figure
one: M-P, S-M, S-P. Assuming that the conclusion is always of the form S-P, there
are three other ways to arrange the terms, for example, P-M, S-M, S-P is in figure
two. With 64 possible moods and four possible figures, there are 256 distinct
syllogisms. Of these a mere 24 are valid in Aristotelian logic, even allowing weak
conclusions, such as Some S are P when All S are P could have been concluded.
Some S are P does not validly follow from All S are P in contemporary classical
logic (Shapiro & Kissel, 2022), which implies that All S are P is true when S is an
empty term. But this difference will not affect our psychological points here. From
a psychological point of view, we can double the number of syllogisms by
switching the positions of the first and second premises. Though this does not
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
affect the logic, it can and does affect the inferences ordinary people draw.
The psychological studies divide into those using abstract terms (typically the
letters A, B and C) and those using realistic content. Both have provided
interesting findings, and we consider the abstract reasoning first. The standout
finding of the abstract studies is that people endorse the conclusions of most
syllogisms, even though the great bulk of them are invalid. To our knowledge,
only one published study (Evans et al., 1999) presented every possible combin-
ation of moods and figures, with participants asked in one group to judge
whether the conclusion was necessary (equivalent to valid), and in another
whether it was possible, given the premises. (The authors provided an appendix
showing the percentage saying ‘Yes’ in each case.) Asked to judge the necessity
(validity) of conclusions given the premises, participants accepted about
72 per cent that were valid conclusions, around 45 per cent that were invalid
but possible conclusions, but only 10 per cent that were logically impossible
given the premises. So clearly there was some understanding of the logic being
Human Reasoning 5
shown, but with very high error rates. When asked if the same conclusions were
possible, these numbers increased in all cases, so that ‘possible’ seemed to
participants simply to be a cautious form of ‘necessary’. The effect was greatest
for syllogisms with possible conclusions, but still present for necessary and
impossible ones. Basically, and contrary to both logic and the instructions,
participants like to endorse conclusions unless there is a clear reason not to do
so. The overwhelming finding in this research is the very high endorsement rate
of fallacies. Abstract syllogistic reasoning is clearly very difficult for (typically)
university student participants when not trained in logic.
Both the mood and the figure of the syllogisms also bias the judgement of
validity. It was claimed in the earliest study (Woodworth & Sells, 1935) that
people like to endorse conclusions which match the mood of the premises, the
so-called atmosphere effect. While this finding has been broadly replicated in
later studies (Evans et al., 1993), more nuanced psychological accounts of the
precise nature of the bias have been developed (e.g., Chater & Oaksford, 1999).
The figure of the syllogisms also biases reasoning, in that people generally
prefer conclusions in which the order of the terms corresponds with their order
of presentation in the premises. This has been observed both when people
generate their own conclusions (Johnson-Laird & Bara, 1984) and when con-
clusions are given for evaluation (Morley et al., 2004). In brief, abstract
syllogistic reasoning is very difficult for ordinary people, who make many
logical mistakes, frequently endorse fallacious conclusions as necessary and
are systematically biased by logically irrelevant variants of the presentation (see
Oaksford & Chater, 2020, for new paradigm studies of syllogisms).
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
1. Validity effect. People endorse more valid than invalid arguments, that is
(VB+VU) > (IU+IB)
2. Belief effect. People endorse more arguments with believable than
unbelievable conclusions, that is (VB+IB) > (VU+IU)
3. Interaction effect. Belief bias is larger for invalid arguments, that is
(IB-IU) > (VB-VU).
All three effects are about equally large and highly reliable. Figure 1 shows the
endorsement rates of the combined experiments of Evans et al., 1983. Most
replication studies show a statistically significant belief bias for valid as well as
invalid arguments, but of smaller size, as Figure 1 clearly illustrates.
To illustrate the findings with some examples, the participants in the Evans
et al. (1983) study found the following argument quite compelling, with
71 per cent declaring it valid:
By contrast they were very unimpressed with this argument; only 10 per cent
thought it valid:
100
90
80
70
60
50
40
30
20
10
0
VB VU IB IU
The two arguments have precisely the same logical form and neither is valid. The
difference is that people believe the first conclusion and disbelieve the second.
However, the belief bias effect in syllogistic reasoning (demonstrated with
other reasoning tasks) does not mean that participants ignore the logic, when
clearly instructed to assess the necessity of the conclusion given the premises.
Rather it suggests that there is some kind of conflict between the logical task and
influence of belief. Evans et al. (1983) showed that this was not due, as one
might suppose, to some participants using logic and others belief, but rather that
the conflict was within individual reasoners. That is, people would sometimes
go with logic but other times with belief. This was one of the key findings that
lent support to dual-process theories of reasoning discussed in Section 6. The
interaction of belief with validity may relate to the general finding that people
are all too willing to endorse fallacies generally. One reason to withhold an
inference might be perception of its invalidity, but another could be the
unbelievability of its conclusion. Various authors have suggested that people
might be more motivated to disapprove of an argument and look for
a counterexample case when they disbelieve its conclusion (Evans, 2007a).
De Neys (2012) surprisingly found that belief–logic conflict may be detected
rapidly and preconsciously, reflected in response times, confidence levels and
autonomic responses, leading to an argument that there may be ‘logical intu-
itions’ (but see Section 6.3). Moreover, the logic of the problem can interfere
when people are instructed to respond on the basis of belief (Handley et al.,
2011). These findings seem very odd. How could one know whether an argu-
ment was valid or not, without first engaging in a slow process of reasoning?
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
However, such studies have generally used very much simpler logical argu-
ments than classical syllogisms (Section 6.3), which may not engage working
memory (Evans, 2019). What all these studies demonstrate beyond doubt,
however, is that most people find it difficult or impossible to disregard their
prior beliefs completely when engaged in a deductive reasoning task.
Traditional studies of syllogistic reasoning did point the way to the new para-
digm by showing clearly that this reasoning could not be fully explained without
taking account of people’s beliefs. Sections 3–5 will be on the new paradigm topics
of how belief should influence reasoning and how it actually does.
content. They are somewhat simpler, however, since they relate only two rather
than three terms. Participants are usually tasked with the following four argu-
ment types to evaluate:
Of these inferences, MP and MT are valid, and DA and AC are invalid, in both
classical logic and probability logic (Evans & Over, 2004). While frequencies of
endorsement vary somewhat between different studies in the literature, there are
some very clear trends when materials are abstract, as in the above examples.
MP is endorsed very highly, with a number of studies reporting 100 per cent
acceptance rates. MT, the other valid inference, is endorsed far less frequently,
about 60 per cent of the time. The two fallacies, AC and DA, are frequently
endorsed but the extent varies considerably across studies. The latter finding
could be accounted for by assuming that people adopt a biconditional interpret-
ation or accept invited pragmatic inferences (Evans & Over, 2004). This cannot
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
(Braine & O’Brien, 1991; Rips, 1994). Mental logic theorists argued that people
use an innate ‘natural logic’ comprised of natural deduction inference rules,
some of which take a simple, direct form. An example of such a rule is and-
elimination, inferring p from p & q (see also Section 3.2). Such inferences are
said to be made automatically and without effort, drawing on implicit rules that
are effectively part of the language processing system. Other inferences can
only be made by high effort, and presumably conscious, indirect reasoning
strategies. For example, as in classical natural deduction theories proposed by
logicians (Shapiro & Kissel, 2022), there is no direct rule for MT in mental
logic. It can only be derived indirectly using reductio ad absurdum reasoning. In
the above example, one would suppose that there was an A, and if there were an
A then there would be a 4. Since there is not a 4, the supposition of that there is
an A must be false by a reductio. However, this indirect reasoning is hypothe-
sised to be slow and error prone compared with the direct MP. The latter follows
immediately from an inference rule built into the natural language meaning of if
in the mental logic account.
Mental model theory, by contrast, denies that there are any inference rules in
the head, and proposes instead that deduction is based on a simple semantic
principle: an inference is valid if there is no counterexample to it. In this account,
people construct and use mental models in reasoning to represent possible states
of the world and search for counterexample models to try to infer or test the
validity of inferences. Presumably this is at least in part a conscious process
because core to the theory and its predictions is the claim that mental models load
working memory, increasingly so when they are more complex or more than one
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
needs to be considered. MP is simple and immediate because the p & q case is the
default mental model that is used to represent if p then q, along with initially
unspecified other possibilities. When presented with the premise not-q, no infer-
ence will follow unless the individual tries to ‘flesh out’ other states of the world
(models) that could be consistent with the conditional statement. These additional
models are: not-p & q and not p & not q, with only p & not-q being excluded. On
this basis, only the second of these models is consistent with a conditional
statement in which not-q also holds, and it yields the not-p conclusion. But
again, this reasoning process is supposed to be indirect, difficult, and prone to
error (see Section 3.1 for more on the conditional in mental model theory).
Conditional inferences have also been studied with a variety of realistic
content, which can affect the rates with which the conclusions are inferred in
the four inferences (Byrne, 1989; Stevenson & Over, 1995). Of particular import-
ance are the real-world beliefs that participants hold about the relationship
between p and q, for a conditional statement if p then q. Studies have manipulated
whether the occurrence of p appears to be either a sufficient or necessary
10 Philosophy of Mind
their degree of belief in the conclusion. All participants were given a test of
general intelligence and split into subgroups high and low on this measure.
Under pragmatic reasoning instructions, participants of all abilities showed
a similar belief effect, in that they assigned lower ratings to conclusions
drawn from unbelievable conditional statements. Under deductive reasoning
instructions, however, the belief bias effect was restricted to lower-ability
participants only. This finding is important as few studies of deductive rea-
soning have included intelligence measures. It indicates that people of higher
cognitive ability are better able to comply with instructions to assume the truth
of the premises. Participants of lower cognitive ability are less able to do this
and tend to display belief bias.
In a new paradigm approach (Section 2.6), we would ask how far being less
able to follow instructions to assume premises is a serious bias that limits
rationality (Section 7). But these findings are consistent with the broader
individual differences research programme of Keith Stanovich, who argues
that higher intelligence facilitates the ability to ‘decouple’ beliefs and reason
Human Reasoning 11
Taken as a whole, the above findings suggest that either much real-world reasoning
is irrational, or else that the yardstick for rationality should not simply be classically
valid inferences from assumptions. Of course, being unable to suspend one’s beliefs
when instructed to do so by a psychologist in an experiment cannot, in itself, be
irrational. As a result, this experimental work, along with that in other fields of
reasoning and decision-making, has initiated a great rationality debate (see Evans,
2021, and Section 7). Dissatisfaction with the traditional approach and its associated
deductive paradigm has also led to the new paradigm in the psychology of reasoning,
which we continue to describe in the next and later sections.
in which all propositions are assumed to be true or false, with no account taken
of their subjective probabilities. The original use of ‘new paradigm psychology
of reasoning’ was to refer to studies that placed subjective probability judge-
ments at the centre of the field (Over, 2009). But for some psychologists, the
problem is not just a demand to reason from assumptions in classical logic, but
the use of a strong normative theory to assess human reasoning, obscuring the
psychological study. It has been argued that the psychology of reasoning should
fall in line with the rest of cognitive psychology by focussing on what people do,
rather than recording errors and biases (Elqayam & Evans, 2011; Evans, 2002).
This change encourages a much deeper focus on pragmatics in reasoning
(Bonnefon, 2013).
However, there is also a strong tradition and contemporary practice of
proposing alternative normative models to classical logic for reasoning (Over,
2020). A sustained research programme of this type has been run by the
psychologists Mike Oaksford and Nick Chater, dating from their seminal
paper on the Wason selection task, in which they argued that the standard
‘erroneous’ responses could have a rational explanation from a different view-
point (Oaksford & Chater, 1994). Oaksford and Chater (2007) argued influen-
tially that Bayesian subjective probability theory could replace classical logic as
the norm for human reasoning. Philosophers contributed to the paradigm shift.
L. Jonathan Cohen’s (1981) critique of the psychology of reasoning and deci-
sion-making, which will be covered in Section 7, directly challenged psycholo-
gists’ use of standard normative theories. Also very influential was Edgington
(1995) on the natural language conditional if p then q (the first author learned
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
much from having her as his PhD supervisor). She explained the problems of
claiming that if p then q is logically equivalent to the classical material
conditional, itself logically equivalent to not-p or q. This claim was implicitly
assumed to be correct in much psychological work at that time (but see Braine &
O’Brien, 1991, and Rips, 1994, for dissenting voices).
Edgington (1995) argued strongly (see originally Adams, 1965) for what we
will call the Equation and, in experimental contexts, the conditional probability
hypothesis. The Equation/conditional probability hypothesis states that the
subjective probability of the natural language conditional P(if p then q) is the
conditional subjective probability of q given p:
A book published by the current authors (Evans & Over, 2004) drew together
philosophical and psychological work on conditionals and proposed
a suppositional theory of conditionals (see Section 5.1), linked to broader theory
of hypothetical thinking, in which P(if p then q) = P(q|p). Following this work,
a major field of study of conditionals within the new paradigm has developed,
with a range of logical ideas being taken from philosophy and applied in
a psychological context (see Sections 3, 5, and 6). The traditional deduction
paradigm has not been entirely abandoned, however. For example, it is still
relevant to the study of dual processes in reasoning (Section 6), as it requires
what is known as Type 2 processing – high effort, loading on working memory –
to comply with the instructions to assume the premises of inferences, disregard-
ing belief, and find conclusions that necessarily follow. Only some people with
special training, or relatively high cognitive ability, are able follow such
instructions fully and reliably.
why the new paradigm rejected the traditional presupposition that the natural
language conditional is equivalent to the classical material conditional, and also
cover the logical and philosophical foundations of the new paradigm.
Johnson-Laird and Byrne held that (3.1) is true when Arthur is in Edinburgh, and
Carol is in Glasgow, and false when Arthur is in Edinburgh, and Carol is not in
Glasgow. They then asked, ‘But, suppose its antecedent is false, i.e. Arthur is not
14 Philosophy of Mind
in Edinburgh, is the conditional true or false?’, and answered, ‘It can hardly be
false, and so since the propositional calculus allows only truth or falsity, it must be
true’ (Johnson-Laird and Byrne, 1991, p. 7). The propositional calculus is the
branch of classical logic for reasoning with the connectives not, and, and or, and
so Johnson-Laird and Byrne were clearly presupposing here, without argument,
that the correct logic for the natural language conditional (3.1) is bivalent and
truth functional. They agreed with many other theorists that (3.1) is true when d is
true and g is true, and false when d is true and g is false. But then Johnson-Laird
and Byrne appealed to classical logic to infer that (3.1) is true when d is false and
g is true, and true when d is false and g is false. They are presupposing that (3.1) is
logically equivalent to not-d or g. This material conditional analysis of the natural
language conditional is summarised in Table 1.
Johnson-Laird and Byrne (1991, pp. 73–74) accepted the paradoxes of the
material conditional analysis. The first paradox is that this analysis implies the
logical validity of inferring if p then q from not-p, and the second paradox is that
it implies the logical validity of inferring if p then q from q. They rightly wrote
that we must decide ‘ . . . to abandon this analysis of the conditional . . ., or to
accept the validity of these apparently paradoxical inferences and to explain
why they seem improper’, and they added, ‘We shall embrace the second
alternative’. They went on to argue that these inferences appear paradoxical
because they ‘throw semantic informative away’. In other words, the conclusion
is less informative than the premise.
It is true that in cooperative communication we would not, generally, want to
deny our hearers information by making an assertion that we had inferred from
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
one with more semantic information (compare Grice, 1989, on these paradoxes).
But we do not have to worry about communicating well with other people when
we are making inferences from our own degrees of belief, for our own benefit. For
example, suppose we are wondering whether to invest in a cryptocurrency
scheme that promises a large profit, and consider:
pq if p then q
TT T
TF F
FT T
FF T
(3.2) If we invest in the scheme (i), then our money will be in honest hands (m).
validity preserves truth. For inferences with more than one premise, we need to
define the uncertainty of a premise p as 1 – P(p) and of a premise if p then q as
1 – P(q|p), and then an inference is probabilistically valid, p-valid, if and only if
the uncertainty of its conclusion cannot exceed the sum of the uncertainties of
its premises (Adams, 1998). More informally, a valid inference cannot increase
our uncertainty: it cannot take us from less uncertainty to more uncertainty. An
example of a two-premise inference is and-introduction, inferring p & q from
p and q as separate premises. This inference is clearly classically valid, since its
conclusion cannot be false when its premises are true, and p-valid, because its
conclusion cannot be more uncertain than its premises. This definition of
probabilistic validity, p-validity, plays a fundamental role in probability logic
(Adams, 1998) and in new paradigm psychology of reasoning (Oaksford &
Charter, 2020; Over, 2020).
In the traditional paradigm, reviewed in Section 2, psychologists usually
asked their participants to assume the truth of the premises in a reasoning
experiment while setting aside any relevant beliefs they might have. These
premises could sometimes be highly abstract and were usually detached from
belief acquisition, practical decision-making, or scientific prediction, and could
be unbelievable when they were not abstract. Such premises are of little use in
everyday or scientific reasoning, where there is little or no point in arbitrarily
assuming what is irrelevant and unbelievable. Participants could find it hard to
make such assumptions and could automatically use what they judged to be
relevant beliefs for their inferences, with the result that they could be charged
with belief bias in a traditional approach. Using a reductio ad absurdum
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
necessary to cover all the uses of probability theory. In this respect, most new
paradigm psychologists, including ourselves, side with Ramsey and not de
Finetti.
Both authors argued, however, that people’s degrees of belief can be meas-
ured by the bets they are willing to make, and both made use of the notion of
a conditional bet, for example, we might say that, if a certain coin is spun, we bet
it will come up heads. For this bet, suppose we observe that the coin has a worn
edge and decide that we will pay 60 Euro cents to make the bet under the
following conditions. We will win the bet and get 1 Euro when the outcome
turns out to be heads and will lose the bet and get nothing when the outcome
turns out to be tails. If the coin is not spun, the bet, being only conditional, is
‘void’, and we get our 60 cents back. The conditional subjective probability for
us is then .60 that a head will be the result given that the worn coin is spun, and
.60 is the expected value of the bet. Section 5.1 has more on the relation between
conditional bets and the use of natural language conditionals.
18 Philosophy of Mind
substituted a double-tailed coin for the fair one before the random selection,
then invariance will not hold, and we will conclude that P2(h) = 1. Non-
deductive inferences, the topic of Section 4, are present even in this rudimentary
example.
For a less artificial example, recall our earlier statement:
We might have a high degree of belief in (3.3) because we know that Arthur was
going to catch the London train to Edinburgh but, by the train times, might be just
short of Edinburgh. If we learn, however, that Arthur is not in Scotland as a result of
missing his train in London, we will not appeal to (3.3) and conclude, using MT and
double negation (inferring p from not-not-p), that he is in Edinburgh. It will be clear
to us that invariance has failed and that (3.3) has a probability of 0. We saw in
Section 2.4 that, although MP and MT are both valid inferences in classical and
probability logic, the endorsement frequency of MT is significantly lower than that
of MP. Whether this fact can be explained by some awareness of the greater
fragility of dynamic MT is yet to be explored in the psychology of reasoning.
Strict conditionalisation is called ‘strict’ because we use it when we are
certain of the evidence. But perhaps there is some uncertainty about whether
Carol is in Glasgow, and so we cannot increase P2(g) to 1. Let us suppose we can
increase P2(g), so that P2(g) > P1(g). In this case, we can use a more general
Bayesian procedure than strict conditionalisation to go from the prior to the
22 Philosophy of Mind
3.5 Diagnosticity
In our example of Bayesian reasoning about Arthur and Carol, we took the
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
3.6 Conclusion
In recent years, most psychologists have abandoned the traditional paradigm
view that the conditional in natural language is a material conditional when
people fully model its meaning. Surprisingly for a psychological paradigm, the
traditional approach paid little attention to inferences from beliefs, except as
a biasing factor, and made no attempt at all to explain changes in degrees of
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
4 Non-deductive Reasoning
We introduced Bayesian reasoning in Section 3.4 and explained that the proof of
Bayes’ theorem is itself an instance of deductive reasoning from the axioms of
probability theory. That it is the starting point for Bayesian accounts of
24 Philosophy of Mind
equal intervals, and they typically asked about three numbers consistent with
their hypothesis, for example, 5 7 9, in a positive test strategy. In such case they
would invariably receive a ‘Yes’ response, appearing to confirm their hypoth-
esis. The participants did not seem to Wason to be trying to falsify the hypoth-
eses that they considered. They could additionally be said to have assigned too
high a prior probability, P1(h), to the hypothesis that appeared to be suggested
by the context.
Influenced by Popper and helping to initiate the traditional paradigm, Wason
inferred that his participants had a serious failing in rationality. When Wason
and Johnson-Laird (1972) reviewed the 2–4–6 task, they charged the partici-
pants with having a confirmation bias, the tendency to seek verification in the
task and not falsification. They pointed out that a negative test of their hypoth-
esis, such as 1 5 6 would have falsified it. But the problem is that, in general,
positive testing can lead to falsification of hypotheses and is frequently used in
science as well as normal life (see Evans, 1989, 2016). If one predicts an
Human Reasoning 25
experimental result that is not observed, that is a falsification. Hence, one cannot
infer a confirmatory attitude simply from the observation of positive testing, as
a number of later authors have noted. The difficulty of the 2–4–6 task arises
from the fact that the participant is cued by the biased example to adopt
a hypothesis which is a more specific case of the actual rule of any ascending
sequence. In this particular situation, all positive tests will get positive feed-
back. The hypothesis is not so much incorrect as insufficiently general.
Their mistake is like stating, truly, that iron expands when heated, without
testing if this property applies to metals in general.
Consider a Bayesian analysis of this task in the new paradigm. Because of the
way it is presented to them, the participants may assign a high prior probability
that Wason’s hidden rule was that to increase the three numbers by two. Letting
this hypothesis be ht, their implicit judgement was that P1(ht) was high at the
start of the task. The conditional probability of a ‘Yes’ answer, y, given ht and
the use of 5 7 9 as a guess is 1, for y can be logically deduced from ht and the use
of 5 7 9. Let us just express this judgement as P1(y|ht) = 1, with the understand-
ing that it has made in the context of 5 7 9 as the guess. The participants might
have been so focused on their hypothesis ht that they made no judgement about
the probability of y given its negation, P1(y|not-ht), but it is clear that a ‘No’,
not-y, reply to 5 7 9 should result, by strict conditionalisation, in a posterior
judgement that ht was certainly false, P2(ht) = 0, as P1(not-y|ht) should be 0
when P1(y|ht) is 1. The hypothesis would be falsified, and a high degree of prior
belief in it should fall to 0. A ‘Yes’ answer would confirm ht in a Bayesian
analysis, increasing the degree of belief in it, in this positive test strategy.
A negative test strategy would be to guess a triple that should receive a ‘No’
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
answer given the hypothesis, like 1 2 3. With the understanding that this triple is
the guess, P1(not-y|ht) = 1, and then falsification, from a ‘Yes’ reply, or
confirmation, from a ‘No’ reply, could again result.
More precision in a Bayesian analysis of the task is impossible when P1(ht)
and P1(y|not-ht) are not specified. At Wason’s suggestion, some studies of it did
eventually help participants to think about both a hypothesis, ht, and its neg-
ation, not-ht, or an instance of its negation, with dual goal instructions (Evans,
2016). This resulted in much higher solution rates on the 2–4–6 task, although
participants still followed a positive test strategy for a complementary hypoth-
esis which was only implicitly negative for the focal hypothesis. Note that
a negative test strategy is not normally needed, or even advantageous, in
many contexts (see Klayman & Ha, 1987; Poletiek, 2001). In a fully Bayesian
and new paradigm version of the task, the participants’ judgements P1(ht) and
P1(y|not-ht) would have to be measured, and the question would be how closely
their posterior judgement would be to the P2(ht) that was implied by Bayes’
26 Philosophy of Mind
theorem and conditionalisation. But the existing studies do support the implica-
tion of the singularity principle (Evans, 2007a): that people will tend to focus on
a single hypothesis that is of importance to them, until they have a reason to give
it up. This is also consistent with the cognitive miser hypothesis (Stanovich,
2018), positing a system which evolved to minimise cognitive effort.
Of course, conforming to probability theory and being logically coherent
are not enough for effective reasoning and belief updating in the real world.
People must not only go beyond probability theory by using strict or Jeffrey
conditionalisation and making inferences about invariance. They must also
have a strategy for considering hypotheses and updating them. There is no
proof in probability theory that obliges us, on pain of incoherence, to have
a positive or negative test strategy or some combination of the two. Someone
could be logically coherent forever by asking Wason about numbers that
increased by two.
There are then lessons to be learned from Wason’s 2–4–6 task, but it can still
be argued that it does not correspond well to scientific investigations of hypoth-
eses or to most ordinary belief updating. It could be charged that it is artificial
for people to have to trust someone to tell them whether a hidden rule had been
conformed to by some data. But we do have to depend in ordinary reasoning on
experts to tell us what has been confirmed and well established. We face too
many complex questions and problems in our contemporary world, even in our
everyday reasoning and decision-making, to rely only on whatever highly
limited expertise we may possess (see Sloman & Fernbach, 2017, on how we
‘never think alone’). Even if we derive P1(h|e) using Bayes’ theorem, we may
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
well have to trust an expert to tell us that e holds and concluding that someone is
trustworthy has to be mainly the result of non-deductive reasoning. In the next
section, we will introduce experiments on relying on experts in arguments.
4.2 Argumentation
The study of argumentation in psychology seeks to account for people’s infor-
mal and ordinary use of arguments in reasoning (Hahn & Oaksford, 2007;
Oaksford & Chater, 2020). Many common arguments are deductively valid,
with both classical and p-validity. The best example is simply MP, which is so
common, and often so automatic, that its use can pass unnoticed. But the
premises of MP and other valid inferences usually have to be supported by
arguments that are not deductively valid. A Bayesian new paradigm approach
reveals that some of these invalid inferences can be strong, well-justified
reasoning in some contexts, with highly probable conclusions, and not the
unqualified ‘fallacies’ of a traditional classification. For example, the classical
Human Reasoning 27
probability that the hypothesis is asserted given that it holds, that the asserter is
an expert, and that the asserter is trustworthy. The model is simplified by
assuming that h, ex, and tr are independent of each other. Independence, and
conditional independence, assumptions are of importance in simplifying
Bayesian models. In this instance, the independence assumption allows Harris
et al. to predict what the P1(h|ha) judgements should be, according to the model,
from judgements about P1(ha|h & ex & tr) and the other relevant probabilities.
Their two experiments gave good, though somewhat noisy, support to their
model.
This model is computational, as Harris et al. emphasise. They do not try to
give an account of the underlying psychological processes. These may
be relatively simple in some cases. We might believe that, if Ivermectin is
not effective for treating COVID-19, then the USA Food and Drug
Administration (FDA) will not approve it for treating COVID-19, and
infer from this that, if the FDA does not approve it for treating COVID-
19, then it is not effective for treating COVID-19. Inferring high confidence
28 Philosophy of Mind
There is a .02 probability that a medical expert will be picked out from
a group in a certain context, P1(hmex) = .02. The probability is 1 the person
picked out will state that Veklury can be an effective treatment for COVID-19
if they are a medical expert, P1(hsv|hmex) = 1, and a probability of .5 that the
picked out person will state Veklury can be an effective treatment for
COVID-19 if they are not a medical expert, P1(hsv|not-hmex) = .5. What is
the probably the person is a medical expert if they state that Veklury can be an
effective treatment for COVID-19, P1(hmex|hsv)?
The correct answer to the above question, using Bayes’ theorem, is about .04,
that is, P1(hmex|hsv) is about .04. Studies of problems like the above have been
run for a long time, following on from Tversky & Kahneman (1974), and it has
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
been found that participants tend not to answer them correctly, responding that
P1(hmex|hsv) is much higher than .04. They seem to be suffering from the
fallacy of base-rate neglect. In above example, this fallacy would be not to
take account, or insufficient account, of the fact that the prior probability, the
base rate, that the person picked out is a medical expert is so low, with P1
(hmex) = .02. It is true that medical experts will apparently make the statement
about Veklury with certainty, and the non-experts will respond with this
statement at random, but there are many more non-experts than experts in
this sample, and there will be many more statements made about Veklury as an
effective treatment for COVID-19 by non-experts than by experts. The fact
that someone makes this statement will therefore not be strong evidence that
they are a medical expert.
Influenced by Kleiter (1994), Gigerenzer and Hoffrage (1995) found that
Bayesian reasoning can be improved if experimenters use what they called
natural frequencies in the Bayesian word problems given to their participants.
Human Reasoning 29
For the above problem, a natural sampling (Kleiter, 1994) version using natural
frequencies would go like this:
In a sample of 100 people, 2 are medical experts and 98 are not medical
experts. The two who are medical experts will state that Veklury can be an
effective treatment for COVID-19. Out of the 98 who are not medical experts,
49 will state that Veklury can be an effective treatment for COVID-19. Out of
all those who state that Veklury can be an effective treatment for COVID-19
how many will be medical experts, ___ out of ___?
The correct answer to this problem is 2 out of 51, and it is clearly easier for most
people to come by than .04, or more accurately .039, or an even more accurate
longer decimal expansion. Cosmides and Tooby (1996) used an evolutionary
argument to try to explain why such versions of Bayesian problems
are relatively easy to solve. They argued that our human ancestors, who lived
as hunter-gatherers for almost all of our evolutionary history, acquired, by
natural selection, an adaptive dedicated module for easily processing ‘natural
frequencies’. For example, hunter-gatherers might, arguably, have concluded
that 47 out of 52 of their group members who ate a certain red berry recovered
from a cold within a week.
We doubt that an evolutionary hypothesis about a dedicated module for
natural sampling can explain how a ‘natural frequency’ format facilitates
Bayesian reasoning in a psychological word problem (see Evans, 2007a and
Over, 2003). Even if a group of hunter-gatherers had the linguistic means to
represent a number like 47 or 52, this sample from their group might well not tell
them that eating the red berry explains a recovery in a week (Sebben & Ullrich,
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
2021). The sample could be biased, and even if it is not, people may recover
from a cold within a week whether they eat the berry or not.
The term ‘natural frequency’ suggests that there is an objective frequency in
play here. But natural sampling only gives us sample frequencies, which must
be processed further in non-deductive reasoning for high degrees of confidence
in objective frequencies and causes. Bayesian belief updating, with its non-
deductive aspects, is itself needed to take us from a sample frequency of, for
instance, five heads in a row to a high degree of confidence that a coin has a bias
which results in a high objective probability of coming up heads.
We do not doubt, of course, that hunter-gatherers inferred the existence of some
objective frequencies and causes through non-deductive reasoning that increased
their reproductive success. They could notice that there was rapid new berry and
other growth that attracted game after a fire cleared a forest. They apparently
started fires themselves for this purpose for tens of thousands of years (Zimmer,
2021). These interventions could have, in effect, confirmed for them a Bayesian
30 Philosophy of Mind
model for such causal reasoning (Pearl, 2000; Sloman & Lagnado, 2015). Some
simple natural sampling would have played a part in this process. Still, solving
a word problem in a psychological experiment is a long way from starting fires to
encourage the growth of blueberries, and a dedicated module for natural sampling
could help with the latter, but not the former.
For us, Tversky and Kahneman (1983) had the key insight for explaining
why participants in psychological experiments can solve Bayesian word
problems when these are in a ‘natural frequency’ format. In their article,
as we have reported in Section 3.2, they found that people sometimes
commit the conjunction fallacy of, for example, judging that it is more
probable than a man is over 55 and has had one or more heart attacks than
that the man has had one or more heart attacks. But they also found that
instances of this fallacy significantly decreased when it was expressed using
sample frequencies. In one study, they told their participants that there was
a sample of 100 males, asked how many of these 100 men had one or more
heart attacks, and how many of these 100 men both were over 55 years and
had had one or more heart attacks. Given these questions, we would avoid
the conjunction fallacy by replying, for example, that 10 of these 100 men
had had one or more heart attacks and that 8 of these had had 1 or more heart
attacks and were over 55. Tversky and Kahneman pointed out that, in this
version of the task, the class inclusion relation can be ‘readily perceived and
appreciated’. The class of men who have had one or more heart attacks and
are over 55 is necessarily included in the class of men who have had one or
more heart attacks, and this relationship can be literally seen when we draw
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Euler circles in which the circle for former class is inside the circle for the
latter class (Barbey & Sloman, 2007).
In mathematical set theory, when one class includes another, the latter is also
said to be a subset of the former, and we can say that, when this relation holds,
the subset and the set that includes it are ‘nested’, as Euler circles can be nested.
The nested-sets hypothesis implies that a nested-sets format is what helps
people to answer questions about sample frequencies, avoiding the conjunction
fallacy and base-rate neglect. This hypothesis has been confirmed in many
studies (Barbey & Sloman, 2007). The natural sampling format puts its samples
in a nest-sets relationship, and it now seems to be generally accepted that
a ‘natural frequency’ problem has to have a nested-sets structure to be trivially
easy to solve (McDowell & Jacobs, 2017).
There is a necessary connection between elementary set theory and logical
relations, as we effectively illustrated when we introduced the conjunction
fallacy in Section 3.2. This can be seen in the Euler circle representation of
logical relations. The circle of p & q truths is inside the circle of p truths,
Human Reasoning 31
representing the logical fact that the set of cases in which p & q is true is necessarily
a subset of the cases in which p is true. Specifically, the rows of a truth table in which
p & q is true is a subset of the rows of a truth table in which p is true. Making logical
relations clear with a set-subset and nested-sets presentation helps people to avoid
fallacies, both in classical logic and the logic of probability. When we give our
participants the free gift of a nested-sets representation in an experimental word
problem, they can do well. But in the real world, people do not usually receive this
free gift, and then how often they do well with non-deductive reasoning about
sample frequencies is a different question.
4.4 Conclusion
Effective non-deductive reasoning for reproduction, argumentation, or other
goals has to be more than logical inferences about sets and subsets. This is just
to repeat, in other words, that effective reasoning from degrees of belief and for
belief change has to be more than deductive inferences in probability theory.
There are inferences about invariance, which are usually non-deductive, to
account for. And Wason’s participants could have been logically coherent by
naively accepting the hypothesis suggested by the context of his 2–4–6 task and
asking indefinitely about triples increasing by two. Choices must be made about
which hypotheses to test and how to test them. But no matter what normative
principles are proposed for people’s reasoning, they will have to simplify it.
They can sometimes do this by trying to rely on experts, but identifying experts
is sometimes itself hard for people, and they can fail miserably at it. There can
be effective and efficient ways or heuristics to make inferences, but these are not
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
perfect and can also cause problems. That must have been true even when we
were in the happy state of being hunter-gatherers.
We have continued to assume in this section that the Equation, P(if p then q) =
P(q|p), holds for natural language conditionals. Deductive and non-deductive
inferences from p to q, and the strength of these inferences, can be ‘summed up’
in if p then q, and in P(q|p). It is now time for a fuller account of if p then q which
implies the Equation.
5 Conditionals
The new paradigm in the psychology of reasoning identifies the probability of
the natural language conditional, P(if p then q), with the conditional subjective
probability of q given p, P(q|p), and aims to integrate the study of conditional
reasoning with studies of subjective probability and utility judgements. The
origin of the new paradigm can be traced back to de Finetti’s and Ramsey’s
foundational studies of subjective probability (Section 3.3).
32 Philosophy of Mind
pq if p then q
TT T
TF F
FT u
FF u
described there, the probability it will be won is P(q|p), which is its expected
value. Baratgin et al. (2013) and Politzer et al. (2010) confirmed this close
relationship in people’ judgements.
In an extension of de Finetti’s analysis, in which the third value u expresses
uncertainty about if p then q, u becomes the conditional probability P(q|p) itself
(Sanfilippo et al., 2020). It is the expected value, or ‘prevision’ to use de
pq if p then q
TT T
TF F
FT P(q|p)
FF P(q|p)
(5.2) If Anne invests in the cryptocurrency scheme (a), then Carol will too (c).
Suppose our degree of belief in (5.2) is P(c|a) = .6, and our degree of belief in a is
P(a) = .7, and consider MP, inferring c from if a then c and a. Let us also say that,
after making this inference, we judge P(c) to be .6. Our degree of belief in c, .6, is
coherent and does not violate the p-validity of MP as the uncertainty of c, (1 – .6),
is not greater than the sum of the uncertainties of the premises, (1 – .6) + (1 – .7).
The paradoxes of the material conditional are prime examples of inferences
that are not p-valid for suppositional conditionals. P(q|p) can be low when P
(not-p) is high, and when P(q) is high, and experiments have shown that people
do not at all accept these inferences as p-valid in their probability judgements
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
(Cruz et al., 2017). The formal system for p-validity (Adams, 1998) is some-
times called System P (Gilio et al., 2020; Pfeifer & Kleiter, 2009). It descends
from de Finetti’s (1936/1995) logic of probability and Ramsey’s (1926/1990)
logic of partial belief.
As we saw in Section 3.3, coherence in the logic of probability corresponds to
consistency in classical logic but is more general. It is of particular relevance in
the new paradigm, as are the logical coherence intervals that can be derived for
inferences (Pfeifer & Kleiter, 2009). A coherence interval for an inference
specifies the range in which the probability of the conclusion must fall for
coherence. For example, the coherence interval for MP follows from the total
probability theorem:
She remarked that many people might judge (5.3) as ‘ . . . not acceptable, or
even, false’. The Equation/conditional probability hypothesis implies that
(5.3) will be assessed as highly probable and not definitely false, since
P(o|n) is high. In general, P(q|p) is high when P(q) is high and p and q are
independent. Pragmatically unacceptable conditionals like (5.3) are some-
times called ‘missing-link’ conditionals, but this term can be misleading, as
there are pragmatically acceptable conditionals with ‘missing-links’ in some
sense, for example, (3.3) above, ‘If Arthur is not in Edinburgh, then he is in
Scotland’. In sharp contrast, conditionals like (5.3) are not pragmatically all
right, and we will call them Walrus conditionals (Cruz & Over, 2024; Over,
2023b). In Lewis Carroll’s nonsense poem, ‘The Walrus and the carpenter’,
the Walrus says, ‘ . . . the time has come to talk of many things: of shoes, and
ships, and sealing-wax, of cabbages, and kings’. The Walrus’s ‘conversation’
Human Reasoning 37
(5.4) If Mark presses the power button on his TV, then the TV will be turned on.
(5.5) If Mark is wearing socks (w), then his TV will be working (v).
(5.6) If Mark presses the power button on his TV, then the screen will remain
blank.
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
independence conditional. Consider the normal form of (5.5), that is, if w then
(w & v).Now P(v|w) = P(v|not-w) = P((w & v)|w) = P(v), and yet we have
P((w & v)|w) > P((w & v)|not-w) when P(v) is not 0. Therefore, ‘If Mark is
wearing socks, then he is wearing socks and his TV will be working’ is
a dependence conditional, but that is intuitively not sufficient to make it more
pragmatically acceptable than ‘If Mark is wearing socks, then his TV will be
working’.
As we have already illustrated with (3.3) and (5.6), there are pragmatically
acceptable conditionals which are independence and not dependence condition-
als. For another example, doctors could say to parents whose children are
known not to have autism:
(5.7) If your children are vaccinated, they will not get autism.
The pragmatic implication of (5.7) is that whether children will get autism is
independent of whether they have vaccinations. Knowing about independence,
and conveying information about it using conditionals, is of great utility in
Human Reasoning 39
human reasoning and decision-making (Cruz & Over, 2023; Over, 2023b).
It is to the advantage of suppositional theories that they give unified accounts
of dependence and independence conditionals.
(5.8) The butler stole the jewel (b) or the cook did (k).
On this basis, they can also infer with some confidence that:
(5.9) If the butler did not steal the jewel, the cook did.
Let us use P1 for the police’s probability judgements early in their investigations
of this case, and P2 for their later probability judgements. Notice immediately
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
would happen if the police were to discover that the jewel was not stolen, but
only misplaced. In that case, P2(k) and P2(k|not-b) would both be 0.
Suppose, however, that the police become more and more convinced in
dynamic reasoning that the butler stole the jewel. At some point in this process,
they might start to use counterfactual conditionals, beginning with ‘If the butler
had not stolen the jewel . . .’, but they could reject as improbable:
(5.10) If the butler had not stolen the jewel, then the cook would have.
Using the Ramsey test (Section 5.1), the police could reason that the cook would
not be turned into a criminal supposing something intervened to prevent the
butler from stealing the jewel (see Pearl, 2013, on Ramsey’s ‘idea’, and
Kaufmann, 2023, for a temporal analysis of counterfactuals). If they got still
more evidence and started to go backwards, losing their confidence that the
butler is the culprit, their counterfactual thought (5.10) could return to the
indicative (5.9) and their previously high confidence in that. In recent years,
psychologists of reasoning have used Bayesian networks (Pearl, 2000) to
represent people’s conditional reasoning about causes (Oaksford & Chater,
2020), but much more research is needed on how people move in their reasoning
from indicative conditionals to counterfactuals and, sometimes, back again
(Over & Cruz, 2023).
There is much research in psychology on counterfactuals and the emotions,
especially regret (see Corbett et al., 2023, on how the two are linked in human
beings even at a young age). Recall yet again the cryptocurrency scheme and
suppose it does turn out to be a scam. We could express regret in this way:
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
(5.11) If we had not invested in that scheme, we would not look like fools.
But assume our decision about the scheme was a long time ago, and we have
numerous investments, making it hard to remember the details. When we consult
our accountants, they tell us that we did not in fact make that investment, and we use
(5.11) as the major premise of MP to infer that we do not look like fools. Then our
regret vanishes. People do conform to MP when the major premise is
a counterfactual (see Thompson & Byrne, 2002, and Over & Cruz, 2023, for
comment). But psychologists do not have an account of how, or why, logical
reasoning with counterfactuals is so closely tied to people’s emotions, and philo-
sophers have so far paid no attention to this link in their theories of counterfactuals.
5.4 Conclusion
The compliance of humans with the Equation or conditional probability hypoth-
esis, P(if p then q) = P(q|p), has been strongly confirmed in psychological
Human Reasoning 41
Table 4 Dichotomies and features that have been associated with dual processes
in thinking by various authors
Type 1 Type 2
Intuitive Reflective
Non-conscious Conscious
Fast Slow
Automatic Controlled
Associative Rule based
Implicit knowledge Explicit knowledge
Contextualised Abstract
High capacity Low capacity
Independent of general intelligence Correlated with general intelligence
Independent of working memory Imposes load on working memory
but the notion of dual systems is problematic. A number of authors have suggested
that the two systems could have roots in evolutionary distinct systems, System 1
being associated with more animal-like cognition and System 2 more distinctively
human (e.g., Epstein, 1994; Evans & Over, 1996; Reber, 1993; Stanovich, 2004),
but this seems to us now to be an unhelpful way to use the word ‘system’. We now
prefer to describe such approaches as two minds theories, a concept explored in
detail by Evans (2010). System 1 cannot be a single system as there are multiple
kinds of Type 1 process (Evans & Stanovich, 2013), and there are also arguments
for multiple Type 2 systems (Evans, 2019). Hence, the Type 1 and 2 terminology
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
as discussed in Section 4.) He believed that people were trying to prove the
statement true rather than false, and hence looking for the confirming case of
A and 3. In the original paper he also proposed that people have a ‘defective
truth table’ for the conditional (see Section 5), in which the statement was seen
as irrelevant to letters other than A. This explains why the D card is hardly ever
selected, but not why people often choose the 3 card, unless they take the
statement as a biconditional. The alternative account was matching bias. That
is, people chose A and 3 simply because they matched the content of the
conditional statement. In order to demonstrate this, negations have to be
introduced into the conditional. For example, suppose the problem is the same
as shown in Figure 2 except that the conditional statement is ‘If there is an A on
one side of the card, then there is not a 3 on the other side of the card’.
If cards are chosen to verify, then the participants should select the A and 7
(not the 3) cards, but if they are matching, they will continue to choose A and 3,
even though this is now the logically correct choice, as it can uncover
a falsifying case. The latter is what participants actually do, as originally
Human Reasoning 45
demonstrated by Evans and Lynch (1973) and replicated many times since.
Good Popperian as he was, Wason immediately accepted that his verification
bias account must be wrong, he but was nevertheless puzzled by the finding.
The main reason for his curiosity was that he had recently shown that when
asked to provide a short verbal justification for their card choices, people
appeared to show understanding of the logic (Goodwin & Wason, 1972). That
is, they described their choices as verifying or falsifying the statement, consist-
ent with their card selections. As a result, Wason and Evans (1975) ran an
experiment combining the two methods. They used both the affirmative and
negative conditional forms, but also asked participants to provide a short,
written justification for each card choice. Participants performed both tasks,
half affirmative followed by negative and half vice versa. The results of this
simple and single experiment launched the dual-process theory in the psych-
ology of reasoning. (Their findings were never contested and were eventually
replicated in all essential details by Lucas and Ball, 2005.)
First, Wason and Evans replicated the matching bias account. Participants’
predominant choices were the equivalent of A and 3 on both versions of the task
(of course, actual lexical content was varied randomly). However, they also
replicated Goodwin and Wason, in that explanations offered were always
logically consistent with the card choices made. This meant that on the negative
tasks, people generally showed ‘insight’, saying that they were choosing A and
3 in order to prove the statement false, but on the usual affirmative versions they
continued to justify their choices as seeking a confirming combination of A and
3. This applied equally to those receiving the negative version first. So how
could a negative stimulate ‘insight’ which was then immediately lost when the
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
discussed in detail the origin and nature of these two forms of dual-process
theory and their potential for reconciliation. The original rationalisation form of
the theory is still preferred by contemporary authors who believe that reasoning
evolved for the purpose of argumentation rather than problem-solving (Mercier
& Sperber, 2011, 2017).
The second form, which we might call the conflict theory, was stimulated by
the study of belief bias in syllogistic reasoning already described (Sections 2.3
and 2.4; Evans et al., 1983). In view of our discussion of the new paradigm, the
reader may wonder if it is still right to refer to this effect as a cognitive ‘bias’. It
is now held to be permissible, and even obligatory, for people to take account of
their prior beliefs in reasoning, as long as this is consistent with Bayesian
principles (Section 3). But the term ‘belief bias’ persists in the contemporary
literature and strictly speaking it is a bias when observed in experiments with
instructions requiring participants to disregard prior belief. It also turns out that
the ability to comply with such instructions – or not – is a key factor in dual-
process accounts. These are the main reasons ‘belief bias’ continues to be used
in some contemporary studies. In deference to the new paradigm, however, we
will refer to the belief effect in this section, avoiding the term ‘bias’.
The reader may recall that in the study of Evans et al. (1983) participants were
influenced by both the logical validity of the conclusions offered and their prior
believability. Also, this belief effect was stronger for invalid than valid argu-
ments. One account, favoured by the original authors as well as later in a mental
model theory (Oakhill et al., 1989), is known as selective scrutiny. That is,
participants are more motivated to examine the logical basis for unbelievable
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
standing research programme by Keith Stanovich and Ricard West from the 1990s
onwards (Stanovich, 1999, 2011; Stanovich & West, 2000). We should also note
that there have now been many studies of individual differences in measured
working-memory capacity (WMC), which also correlated with performance on
a huge range of cognitive tasks (see Baddeley, 2007). These measures are them-
selves highly correlated with general intelligence, although the size of the correl-
ation is disputed. At the very least, however, we can say that those high in general
intelligence are likely to have higher working memory capacity and are therefore
also likely to be better at Type 2 reasoning.
Stanovich and West have studied a large range of reasoning and decision
tasks drawn from cognitive and social psychology usually using SAT scores as
a surrogate for general intelligence (SAT is, in psychometric speak, highly
g-loaded). Among many other similar findings, they have shown that perform-
ance on the abstract Wason selection task is related to general intelligence
(Stanovich & West, 1998) as in the ability to resist the influence of belief in
syllogistic reasoning (Stanovich et al., 1999). In a recent huge scale and
comprehensive study, they showed that general intelligence is implicated in
the vast majority of known cognitive tasks on which well-established errors and
cognitive biases are demonstrated (Stanovich et al., 2016). However, there is
one important exception: myside bias. This is the tendency to value and propose
only arguments which accord with one’s core values and worldview. Myside
bias is equally marked in those of higher intelligence, education level and
rational thinking style. This exceptionality is so important that Stanovich has
recently written a whole book about myside thinking (Stanovich, 2021). We
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
are relevant, as described by the singularity principle (Section 3.4) and consist-
ent with the cognitive miser hypothesis (Stanovich, 2018). All of these things
have been pointed out repeatedly by both Evans and Stanovich (e.g., Evans,
2007b; Stanovich, 2011). So how did these authors contribute to the fallacy?
The answer is by focussing much dual-process research on particular laboratory
tasks that are novel, abstract, and difficult, and to which real world experience is
hard to apply. Deductive reasoning can be difficult, for example, and prior
knowledge is biasing by definition in the old, traditional paradigm (see
Section 2). Without logical training, people will frequently commit fallacies,
and without high general intelligence, they will find it hard to follow instruc-
tions to assume premises and disregard their prior beliefs (see Section 2). These
laboratory tasks are selected for dual-process research precisely because Type 2
reasoning is necessary (but not sufficient) for their solution, so that we can make
it more or less easy to apply, for example by the use of time limits, instructional
sets or working memory loads.
50 Philosophy of Mind
measures. It has also been claimed that the logic of a reasoning problem can
interfere with judgements to decide the believability of a conclusion (Handley
et al., 2011), in a reversal of the evidence normally claimed to show that
‘belief bias’ research supports dual-process theory. There are also papers
showing that people of higher intelligence may have more accurate intuitions,
which cannot be attributed to better Type 2 reasoning ability (Thompson et al.,
2018).
In commenting on these studies (Evans, 2018, 2019) has pointed to the very
simple nature of the tasks used in these experiments, in which the normative
response could easily be produced rapidly. For example, Handley et al.’s
findings showed that it was more difficult to judge a conclusion unbelievable
when it was derived by a Modus Ponens inference. But MP is arguably
provided automatically by the language module (Braine & O’Brien, 1991).
To infer that a correct answer necessarily results from Type 2 reasoning is, of
course, the normative fallacy. A recent paper by Ghasemi et al. (2022)
confirms this interpretation of the studies (see also Ghasemi et al., 2023).
They showed that intuitive inferences can be rapidly generated from structural
features (as well as beliefs), but that such inferences may or may not be
logically accurate. For example, the conclusion of AC has ‘pseudovalidity’
such that it will interfere with conflicting belief judgements, in just the same
way as occurs for the actually valid MP inference. Intuitions flow from
structure as well as beliefs, but of course we already knew that from examples
such as atmosphere effects in syllogistic reasoning (Section 2.2) and matching
bias on the selection task.
In conclusion, we agree with critics that the ‘received’ version of dual-
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
of confirmation bias (later disputed) and other evidence of logical error and bias
led him directly to conclusions of irrationality, writing later, ‘It could be argued
that irrationality rather than rationality is the norm. People all too readily
succumb to logical fallacies’ (Wason, 1983). From the early 1970s onwards,
the revolutionary research of Daniel Kahneman and Amos Tversky had massive
influence on the study of judgement and decision making. JDM researchers in
the 1960s thought that human decision-making was fairly rational, provided that
probabilities were allowed to be subjective, but Kahneman and Tversky dem-
onstrated a whole range of cognitive biases in a series of highly influential
articles.
Their general thesis was that people judge probabilities by applying heur-
istics such as those based on representativeness (Section 5.3) and availability
(Tversky & Kahneman, 1974). Although often useful, these heuristics could
lead to biases. For example, the availability heuristic is used when we judge
frequency by the ease with which examples can be brought to mind. Thus,
Human Reasoning 53
doctors when confronted with a patient’s symptoms might recall similar cases in
their experience and use these to determine a probable diagnosis. However, as
Kahneman and Tversky demonstrated, this heuristic can often lead to biases due
to how our memories work. If shown a list of names that includes famous as
well as unknown people, for example, participants will overestimate the fre-
quency of the famous ones as they are easier to recall. Doctors could also be
biased if they recalled a patient with similar symptoms but an unusual diagnosis
who was especially memorable, say, as a close family friend. Following the
early papers of Kahneman and Tversky, a very large field of research built up
demonstrating many judgemental biases both in the laboratory and in real world
settings (Gilovich et al., 2002; Kahneman et al., 1982).
Kahneman and Tversky were more cautious than Wason about attributing
irrationality to their participants, but many of their followers were less
restrained. Cohen (1981) read widely in the psychology of reasoning and
JDM and came to the conclusion that psychologists were overstating their
case. In fact, he claimed that irrationality could never be demonstrated by
such experiments, a position later dubbed Panglossian (Stanovich, 2011). As
the second author has observed (e.g., Evans, 2021) there were three main
arguments in Cohen’s paper which Evans terms:
The interpretation problem is that the participants may not understand the
problem in the way in which the experimenter intended. Take as an example,
the conjunction fallacy described in Sections 3.2 and 4.3 (Tversky &
Kahneman, 1983). Given a stereotype suggesting that Linda is a radical
young woman, people may judge that ‘Linda is a feminist and a bank teller’
to be more probable than ‘Linda is a bank teller’. This judgement might seem
to be logically incoherent and a fallacy, but when the two sentences are
included in the same list, it is possible that participants take the second to
mean implicitly ‘Linda is a bank teller who is not a feminist’, in which case
there is no fallacy. The interpretation problem is real and must be attended to
by the experimentalists, but it cannot plausibly explain away the vast
amounts of evidence for cognitive biases. The weakest of Cohen’s claims,
in our view, is the external validity problem: that psychological experiments
are unrepresentative of the real world. They can be artificial, of course, but
there are very many reported experiments that are carefully controlled
pointing to clear evidence of cognitive biases. There have also been many
demonstrations that biases demonstrated in the laboratory occur in real world
contexts (see a number of the papers in the collection edited by Gilovich
et al., 2002).
How far psychologists of reasoning should make normative evaluations of
people’s reasoning is a debatable question in psychological research, as we
first discussed many years ago (Evans & Over, 1996). However, when
Elqayam and Evans (2011) went so far as to propose a descriptivist approach
in which we would essentially do away with normative evaluations in the
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
human, which has been suggested by various authors (e.g., Evans & Over, 1996;
Reber, 1993; Stanovich, 2004) is not strictly correct as pointed out by the
biologist Toates (2004, 2006). Higher animals have working memory and two
different kinds of learning, one based on association and the other on encoding
of individual, episodic events (Schacter, 1987). We think it more accurate to say
that the second system (which does Type 2 processing) became uniquely
developed in human beings, due the evolution of greatly enlarged frontal
lobes, high working memory capacity, and capacities for language and suppos-
itional reasoning (Evans, 2010).
We refer to this development as the new mind, which was added to the old
mind which still shares many features with the minds of other animals, includ-
ing basic emotions and associative learning systems. Humans consequently
have a unique ability to reason about hypothetical possibilities. Stanovich
(2011) has made a similar distinction between an autonomous and a reflective
mind. Our instrumental rationality derives partially from associative learning
systems in the old mind that we share with other animals, as well as innate
cognitive modules underlying, for example, the visual system and language
processing which operate entirely outside of consciousness. But it also depends
on the epistemic rationality of the new mind with its highly developed capacity
for acquiring, storing, retrieving, and reasoning about explicit knowledge and
belief.
Evans (2013) has written about the differences in rationality of the two
minds. Old mind rationality in humans and other animals is driven by the
past. We evolved cognitive modules by adaptation to past environments, and
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
we learn within our lifetimes to adapt our behaviour to the current environment.
Much of this learning is Skinnerian: we repeat activities which have been
reinforced in the past and avoid those which were punishing. This learning
process can lead to helpful habits as well as destructive addictions, for example,
pathological gambling, and phobias. New mind rationality is facilitated by
hypothetical thinking or ‘cognitive decoupling’ (Evans & Stanovich, 2013)
meaning that it can ignore the current context, and imagine scenarios not
grounded in current beliefs about the actual state of the world. Such thinking
can engage with both counterfactual possibilities and yet to be determined
future possibilities. Hence we can, in the manner envisaged by decision theory,
anticipate the future consequences of our actions and change our behaviour
accordingly. But when the two minds are in conflict, the old often wins (Evans,
2010). For example, we go for short-term gratification and not the future benefit
we prefer at a higher level. Yet, it is the new mind that makes us distinctively
human and able to do many things unknown elsewhere in the animal kingdom,
for example in the arts and sciences. To do these things we must maintain, at
56 Philosophy of Mind
least to some extent, logically coherent belief systems and reason with them
well enough for our purposes.
Keith Stanovich (e.g., Stanovich, 2004, 2011, 2018) has written extensively
about rationality within a dual-process framework. His autonomous mind con-
tains a mixture of innate cognitive modules, implicit knowledge acquired by
associative learning, and knowledge which was once explicit and has become
automated by practice. He has argued, intriguingly, that other animals are often
more instrumentally rational than humans, at least as conventionally measured by
normative decision theory (Stanovich, 2013). This is because animals always
pursue immediate goals, while humans have layers of complexity of higher-order
values and conflicting goals in their reflective minds. A good example of a human
bias is sunk costs. Imagine you bought expensive opera tickets months ago but
when the night comes you are exhausted, and your favourite football team is on
TV. Right now, you would rather stay home and watch the game than go to the
opera. In normative decision theory, that is what you should do because the cost of
the tickets is sunk, the same whether you go out or stay in. But most people would
nevertheless go to the opera because they would not like the thought of ‘wasting’
money, or perhaps because they like to think they are sophisticated people who
prefer opera to football. Of course, we could argue that their choice is rational in
a sense not captured by the instrumental rationality of pursuing immediate goals.
Stanovich has argued that normative problem-solving and decision-making
depends on a number of factors. You must first recognise that the problem in
front of you cannot be solved by habit or pattern recognition. Then you must
apply the necessary cognitive effort, you must have sufficient cognitive
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
capacity to solve the problem and you must have relevant mindware, or
knowledge of how to solve it. Mindware is a key concept (Stanovich, 2018):
rationality depends on education and training as well. Other authors have, by
contrast, praised the role of intuition and ‘gut feelings’ in rational decision
making (e.g., Gigerenzer, 2007; Gladwell, 2005). Both perspectives have their
merits. Intuitions can save lives (see Klein, 1998, for examples) and pattern
recognition, derived from experiential learning, has a key role in expert
problem-solving. But Stanovich emphasises that in a modern technological
world, reliance on intuitions can be dangerous and a rational thinking style
highly advantageous. Our cognitive miserliness evolved in a very different
world from that in which we live now.
remaining coherent in our beliefs, then we will all get closer and closer to
agreeing that the coin is double-headed. Our separate priors will be ‘washed
out’, and there will be convergence in our beliefs. There are even arguments that
such convergence is what is meant by scientific objectivity or truth (see
Sprenger & Hartmann, 2019, and the references given there for these conver-
gence arguments and proofs and critical evaluations of them).
Well-confirmed scientific hypotheses and theories, by Bayesian belief
updating, can obviously be used to achieve many goals, some of which have
the greatest importance for our survival as individuals and a species. Even so,
not everyone has the goal of scientific truth in the first place. Scientists should
have it but sometimes do not. Fanatics and propagandists can be indifferent or
even hostile to it. Perhaps propagandists have to use some Bayesian belief
updating to find the best ways of persuading people of some point of view.
But fanatics could simply refuse to acknowledge the existence of scientific
evidence. They might, for example, refuse to accept the word of scientific
experts and appeal to their own ‘experts’, as defined by their myside bias.
58 Philosophy of Mind
Social media algorithms also have the effect of encouraging users to access only
the opinions of others that accord with their own, with or without probative
evidence.
Stanovich (2021) considers at length whether myside bias is really a bias at
all and whether it should be considered to be irrational. As mentioned earlier
(Section 6.2), what sets myside bias measures apart from all other cognitive
biases is that they are not related to general intelligence or rational thinking
measures. The extent of the ‘bias’ is actually predicted by something different:
the strength of the core belief which is influencing the thinking. Referring to the
Bayesian framework, Stanovich points out that the diagnosticity (Section 3.5)
of the evidence should be assessed independently of prior belief, before the two
are combined to conform to Bayes’ theorem (Section 3.4). But whether or not
Bayesian procedures are properly followed, prior beliefs can vary widely
between people and have a major impact on their reasoning.
Many psychologists, for example, would reject apparently strong statis-
tical evidence for extra-sensory perception due to a firm prior belief that ESP
is intrinsically most unlikely. Myside bias is particularly evident for topics
on which there is a deep binary division of opinion, such as the case for the
UK leaving or remaining in the European Union in 2016. In such cases,
Stanovich notes that each side often holds that the other must be less
educated or intelligent, or gullible in accepting weak or unsubstantiated
arguments. But the evidence can suggest that they are not: they may simply
have differing beliefs and values. Myside bias is not about factual beliefs.
Compare our personal beliefs that (a) the UK has left the EU and (b) that the
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
UK should one day re-join the EU. Only the former is verifiable against
objective evidence. There are no clear objective grounds for establishing the
‘should’ in (b).
Rationality can reside in the beliefs people hold rather than the reasoning they
do. Consider again people with actual clinical delusions such as those who have
paranoia (Rhodes et al., 2020). They may ‘hear’ voices, and their new mind,
Type 2 processing may have the goal, which only a human being could have, of
explaining these occurrences. The result could be the inference that they are
being spied on, giving them the satisfaction of goal achievement and to that
extent instrumental rationality. Experiments have found that people with delu-
sions can tend to jump to conclusions given limited evidence (see Rhodes et al.,
2020, for points about these results), but in some instances, the inference to the
explanation of being spied on could satisfy Bayesian principles, with the
problem simply being hallucinatory premises. As we have already noted,
paranoid delusions may prevent some people from achieving many life goals,
but that might not hold for other individuals if their paranoia is not very severe.
Human Reasoning 59
How far people with even clinical delusions fail to achieve instrumental ration-
ality, caused by violating Bayesian or other normative principles, has to be
investigated on an individual basis.
Some limited conformity to logical and Bayesian principles is a necessary
condition for us to have hypothetical thoughts and Type 2 processing to begin
with, before these are assessed for their contribution to instrumental rationality.
As we reported in Section 3.2, people do have an above chance tendency to
conform to p-validity and be coherent, and MP has an especially high endorse-
ment rate. The existence of hypothetical thought most clearly depends on some
correct use of MP in particular, even though that that might sometimes take
place when there is not full logical coherence.
7.4 Conclusions
Psychologists and philosophers alike have been engaged in a debate about
the implications for human rationality of experimental research on human
reasoning and judgement and decision-making. The frequently reported
conflicts between human responses and presupposed normative systems
initially led to concerns that humans might be fundamentally irrational.
However, the past fifty years or so that debate has become much more
complex and nuanced, in part triggered by the contribution of philosophers
(Cohen, 1981), but also by a growing conviction among psychologists that
humans could not have achieved what they have if they were irrational. This
has led to detailed scrutiny of presupposed normative systems, as evidenced
by the new paradigm psychology of reasoning. It is now more common to
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
a cognitive bias. A key aim in the new mind is the attempt to maintain at least
a locally coherent belief system, and it appears that cannot be achieved without
creating a cognitive frame which encourages myside thinking to some extent.
On the other hand, it is clear that myside thinking can undermine instrumental
rationality, for example, when people act on the advice of ‘experts’ defined by
their myside thinking instead of experts more reliably identified (Section 4.2) or
seek out only the views of those who already agree with them, a process
facilitated in the modern age by social media.
8 Overall Conclusions
Psychologists have been studying human reasoning intensively since the 1960s
with much use of the traditional deduction paradigm for the first thirty to forty
years that period. This method instructed people to ignore prior beliefs and draw
only conclusions that necessarily followed from assumed premises, and it used
standard binary logic as a normative system for deciding accuracy. The review
of hundreds of such studies in Evans et al. (1993) made it clear that humans are
poor reasoners by those standards, even those of higher intelligence and educa-
tion levels. First, they make many logical errors, especially by endorsing
inferences with conclusions that do not necessarily follow given the premises.
Second, they are systematically biased by irrelevant linguistic and structural
features of the arguments presented. Finally, people have great difficulty in
ignoring the prior beliefs that are stimulated by the content and context of the
problems presented. The last finding is strikingly the case for the those of
measured lower cognitive ability, even within university student populations.
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
These kinds of results caused a major debate about the rationality of human
reasoning. Philosophers and psychologists were both engaged, and they pre-
sented a paradox to reasoning researchers. How could human have achieved
their astonishing advances in science, engineering, and the arts, which so clearly
set them apart from other animals, if they were fundamentally irrational? The
conclusion eventually reached by most psychologists was that the problem lies
in the presupposed normative system: binary logic is simply not a good yard-
stick for rational reasoning in the real world. Human reasoning is in fact
inherently belief based and probabilistic. This conclusion led in turn to what
is known as the new paradigm in the psychology of reasoning.
The new paradigm recognises that human reasoning usually consists of
drawing probable conclusions from degrees of belief as premises, or from
some contextually relevant hypothetical possibilities, and is often directed
towards revising or updating degrees of belief. It aims to account for this
reasoning using subjective probability theory and the Bayesian account of belief
Human Reasoning 61
Corbett, B., Feeney, A., & McCormack, T. (2023). Prosocial risk taking and
interpersonal regret in children: An individual differences study. Social
Development, 32, 171–187.
Cruz, N. (2020). Deduction from uncertain premises? In S. Elqayam, I. Douven,
J. St. B. T. Evans, & N. Cruz (Eds.), Logic and uncertainty in the human mind
(pp. 27–41). Abingdon: Routledge.
Cruz, N., Baratgin, J., Oaksford, M., & Over, D. E. (2015). Bayesian reasoning
with ifs and ands and ors. Frontiers in Psychology, 6, 192.
Cruz, N., & Over, D. E. (2023). Independence conditionals. In S. Kaufmann,
D. E. Over, & G. Sharma (Eds.), Conditionals – Logic, linguistics and
psychology (pp. 223–233). London: Palgrave Macmillan.
Cruz, N., & Over, D. E. (2024). From de Finetti’s three values to conditional
probabilities in the psychology of reasoning. In P. Egré, & L. Rossi (Eds.),
Handbook of trivalent logics. Cambridge, MA: MIT Press.
Cruz, N., Over, D., Oaksford, M., & Baratgin, J. (2016). Centering and the
meaning of conditionals. In A. Papafragou, D. Grodner, D. Mirman, &
J. C. Trueswell (Eds.), Proceedings of the 38th annual conference of the
cognitive science society (pp. 1104–1109). Austin, TX: Cognitive Science
Society.
Cruz, N., Over, D., & Oaksford, M. (2017). The elusive oddness of or-
introduction. In G. Gunzelmann, A. Howes, T. Tenbrink, & E. Davelaar
(Eds.), The 39th annual meeting of the cognitive science society (pp.
663–668). London: Cognitive Science Society.
De Finetti, B. (1936/1995). The logic of probability. Philosophical Studies, 77,
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
181–190.
De Finetti, B. (1937/1964). Foresight: Its logical laws, its subjective sources.
In H. E. Kyburg, & H. E. Smokier (Eds.), Studies in subjective probability
(pp. 55–118). New York: Wiley.
De Neys, W. (2012). Bias and conflict: A case for logical intuitions.
Perspectives on Psychological Science, 7, 28–38.
De Neys, W. (2022). Advancing theorizing about fast-and-slow thinking.
Behavioral and Brain Sciences, 46, e111.
Doherty, M. E., Mynatt, C. R., Tweney, R. D., & Schiavo, M. D. (1979).
Pseudodiagnosticity. Acta Psychologica, 43, 11–21.
Douven, I., Elqayam, S., & Krzyżanowska, K. (2023). Inferentialism:
A manifesto. In S. Kaufmann, D. E. Over, & G. Sharma (Eds.),
Conditionals – Logic, linguistics and psychology (pp. 175–221). London:
Palgrave Macmillan.
Deary, I. J. (2020). Intelligence: A very short introduction. Oxford: Oxford
University Press.
64 References
Demey, L., Kooi, B., & Sack, J. (2023). Logic and probability. The Stanford
Encyclopedia of Philosophy. E. N. Zalta, & U. Nodelman (Eds.), https://
plato.stanford.edu/archives/fall2023/entries/logic-probability/.
Edgington, D. (1995). On conditionals. Mind, 104, 235–329.
Egré, P., Rossi, L., & Sprenger, J. (2021). De Finettian logics of indicative
conditionals. Journal of Philosophical Logic, 50, 187–213.
Elqayam, S., & Evans, J. St. B. T. (2011). Subtracting ‘ought’ from ‘is’:
Descriptivism versus normativism in the study of human thinking.
Behavioral and Brain Sciences, 34, 233–290.
Elqayam, S., & Evans, J. St. B. T. (2013). Rationality in the new paradigm: Strict
versus soft Bayesian approaches. Thinking & Reasoning, 19, 453–470.
Engle, R. W. (2002). Working memory capacity as executive attention. Current
Directions in Psychological Science, 11, 19–23.
Epstein, S. (1994). Integration of the cognitive and psychodynamic
unconscious. American Psychologist, 49, 709–724.
Evans, J. St. B. T. (1989). Bias in human reasoning: Causes and consequences.
Brighton: Erlbaum.
Evans, J. St. B. T. (2002). Logic and human reasoning: An assessment of the
deduction paradigm. Psychological Bulletin, 128, 978–996.
Evans, J. St. B. T. (2005). How to do research: A psychologist’s guide. Hove:
Psychology Press.
Evans, J. St. B. T. (2006). The heuristic-analytic theory of reasoning: Extension
and evaluation. Psychonomic Bulletin & Review, 13, 378–395.
Evans, J. St. B. T. (2007a). Hypothetical thinking: Dual processes in reasoning
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Evans, J. St. B. T., Handley, S. J., Harper, C., & Johnson-Laird, P. N. (1999).
Reasoning about necessity and possibility: A test of the mental model theory
of deduction. Journal of Experimental Psychology: Learning Memory and
Cognition, 25, 1495–1513.
Evans, J. St. B. T., Handley, S. J., & Over, D. E. (2003). Conditionals and
conditional probability. Journal of Experimental Psychology: Learning
Memory and Cognition, 29, 321–335.
Evans, J. St. B. T., & Lynch, J. S. (1973). Matching bias in the selection task.
British Journal of Psychology, 64, 391–397.
Evans, J. St. B. T., Newstead, S. E., & Byrne, R. M. J. (1993). Human reason-
ing: The psychology of deduction. Hove: Erlbaum.
Evans, J. St. B. T., & Over, D. E. (1996). Rationality and reasoning. Hove:
Psychology Press.
Evans, J. St. B. T., & Over, D. E. (2004). If. Oxford: Oxford University Press.
Evans, J. St. B. T. (2013). Two minds rationality. Thinking & Reasoning, 20,
129–146.
66 References
Evans, J. St. B. T., & Stanovich, K. E. (2013). Dual process theories of higher
cognition: Advancing the debate. Perspectives on Psychological Science, 8,
223–241.
Evans, J. St. B. T., Thompson, V., & Over, D. E. (2015). Uncertain deduction
and conditional reasoning. Frontiers in Psychology, 6, 398.
Evans, J. St. B. T., & Wason, P. C. (1976). Rationalisation in a reasoning task.
British Journal of Psychology, 63, 205–212.
Frankish, K., & Evans, J. St. B. T. (2009). The duality of mind: An historical
perspective. In J. St. B. T. Evans, & K. Frankish (Eds.), In two minds: Dual
processes and beyond (pp. 1–30). Oxford: Oxford University Press.
Frederick, S. (2005). Cognitive reflection and decision making. Journal of
Economic Perspectives, 19(4), 25–42.
Feeney, A. (2018). Forty years of progress on category-based inductive reason-
ing. In L. J. Ball, & V. A. Thompson (Eds.), International handbook of
thinking and reasoning (pp. 167–185). Hove: Psychology Press.
Ghasemi, O., Handley, S. J., Howarth, S., Newman, I. R., & Thompson, V. A.
(2022). Logical intuition is not really about logic. Journal of Experimental
Psychology: General, 151, 2009–2028.
Ghasemi, O., Handley, S. J., Howarth, S. (2023). Illusory intuitive inferences:
Matching heuristics explain logical intuitions. Cognition, 235, 105417.
Gigerenzer, G. (2007). Gut feelings: The intelligence of the unconscious.
London: Penguin.
Gigerenzer, G., & Hoffrage, U. (1995). How to improve Bayesian reasoning
without instruction: Frequency formats. Psychological Review, 102, 684–704.
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Gilio, A., & Over, D. E. (2012). The psychology of inferring conditionals from
disjunctions: A probabilistic study. Journal of Mathematical Psychology, 56,
118–131.
Gilio, A., Pfeifer, N., & Sanfilippo, G. (2020). Probabilistic entailment and
iterated conditionals. In S. Elqayam, I. Douven, J. St. B. T. Evans, & N. Cruz
(Eds.), Logic and uncertainty in the human mind (pp. 71–101). Abingdon:
Routledge.
Gilovich, T., Griffin, D., & Kahneman, D. (2002). Heuristics and biases: The
psychology of intuitive judgement. Cambridge: Cambridge University Press.
Gladwell, M. (2005). Blink. London: Penguin.
Good, I. J. (1983). Good thinking: The foundations of probability and its
applications. Minneapolis: University of Minnesota Press.
Goodwin, R. Q., & Wason, P. C. (1972). Degrees of insight. British Journal of
Psychology, 63, 205–212.
Grice, P. (1989). Studies in the way of words. Cambridge, MA: Harvard
University Press.
References 67
Press.
Johnson-Laird, P. N., & Bara, B. G. (1984). Syllogistic inference. Cognition, 16,
1–61.
Johnson-Laird, P. N., & Byrne, R. M. J. (1991). Deduction. Hove: Erlbaum.
Johnson-Laird, P. N., Khemlani, S., & Goodwin, G. P. (2015). Logic, probabil-
ity, and human reasoning. Trends in Cognitive Science, 19, 201–214.
Kahneman, D. (2011). Thinking, fast and slow. New York: Farrar, Straus and
Giroux.
Kahneman, D., Slovic, P., & Tversky, A. (1982). Judgment under uncertainty:
Heuristics and biases. Cambridge: Cambridge University Press.
Kaufmann, S. (2023). How fake is fake past? In S. Kaufmann, D. E. Over, &
G. Sharma (Eds.), Conditionals – Logic, linguistics and psychology (pp.
389–423). London: Palgrave Macmillan.
Keren, G., & Schul, Y. (2009). Two is not always better than one: A critical
evaluation of two-system theories. Perspectives on Psychological Science, 4,
533–550.
68 References
Newstead, S. E., Handley, S. J., Harley, C., Wright, H., & Farelly, D. (2004).
Individual differences in deductive reasoning. Quarterly Journal of
Experimental Psychology, 57A, 33–60.
Oakhill, J., Johnson-Laird, P. N., & Garnham, A. (1989). Believability and
syllogistic reasoning. Cognition, 31, 117–140.
Oaksford, M., & Chater, N. (1994). A rational analysis of the selection task as
optimal data selection. Psychological Review, 101, 608–631.
Oaksford, M., & Chater, N. (2007). Bayesian rationality: The probabilistic
approach to human reasoning. Oxford: Oxford University Press.
Oaksford, M., & Chater, N. (2020). New paradigms in the psychology of
reasoning. Annual Review of Psychology, 71, 305–330.
Oberauer, K., & Wilhelm, O. (2003). The meaning(s) of conditionals:
Conditional probabilities, mental models and personal utilities. Journal of
Experimental Psychology: Learning Memory and Cognition, 29, 680–693.
Oberauer, K., Weidenfeld, A., & Fischer, K. (2007). What makes us believe a
conditional? Thinking & Reasoning, 13(4), 340–369.
Osman, M. (2004). An evaluation of dual-process theories of reasoning.
Psychonomic Bulletin and Review, 11(6), 988–1010.
Over, D. E. (2003). From massive modularity to metarepresentation: The evolution
of higher cognition. In D. E. Over (Ed.), Evolution and the psychology of
thinking: The debate (pp. 121–144). Hove: The Psychology Press.
Over, D. E. (2009). New paradigm psychology of reasoning. Thinking &
Reasoning, 15, 431–438.
Over, D. E. (2020). The development of the new paradigm in the psychology of
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
University Press.
Reber, A. S. (1993). Implicit learning and tacit knowledge. Oxford: Oxford
University Press.
Rhodes, S., Galbraith, N., & Manktelow, K. (2020). Delusional rationality. In
S. Elqayam, I. Douven, J. St. B. T. Evans, & N. Cruz (Eds.), Logic and
uncertainty in the human mind (pp. 178–191). Abingdon: Routledge.
Rips, L. J. (1994). The psychology of proof. Cambridge, MA: MIT Press.
Rothschild, D. (2023). Living in a material world: A critical notice of Suppose
and tell. Mind, 132, 208–233.
Sanfilippo, G., Gilio, A., Over, D. E., & Pfeifer, N. (2020). Probabilities of
conditionals and previsions of iterated conditionals. International Journal of
Approximate Reasoning, 121, 150–173.
Schacter, D. L. (1987). Implicit memory: History and current status. Journal of
Experimental Psychology: Learning Memory and Cognition, 13, 501–518.
Sebben, S., & Ullrich, J. (2021). Can conditionals explain explanations?
A modus ponens model of B because A. Cognition, 215, 104812.
References 71
Shapiro, S., & Kouri Kissel, T. (2022). Classical logic. The Stanford Encyclopedia
of Philosophy, E. N. Zalta & U. Nodelman (Eds.), https://plato.stanford.edu/
archives/win2022/entries/logic-classical/.
Singmann, H., Klauer, K. C., & Over, D. E. (2014). New normative standards
of conditional reasoning and the dual-source model. Frontiers of
Psychology, 5, 316.
Skovgaard-Olsen, N., Kellen, D., Krahl, H., & Klauer, K. C. (2017). Relevance
differently affects the truth, acceptability, and probability evaluations of
‘and’, ‘but’, ‘therefore’, and ‘if then’. Thinking & Reasoning, 23, 449–482.
Skovgaard-Olsen, N., Singmann, H., & Klauer, K. C. (2016). The relevance
effect and conditionals. Cognition, 150, 26–36.
Sloman, S. A. (1996). The empirical case for two systems of reasoning.
Psychological Bulletin, 119, 3–22.
Sloman, S. A., & Fernbach, P. (2017). The knowledge illusion: Why we never
think alone. New York: Riverhead.
Sloman, S. A., & Lagnado, D. A. (2015). Causality in thought. Annual Review
of Psychology, 66, 223–247.
Smedslund, J. (1970). Circular relation between understanding and logic.
Scandinavian Journal of Psychology, 11, 217–219.
Smith, E., & DeCoster, J. (1999). Associative and rule-based processing:
A connectionist interpretation of dual-process models. In S. Chaiken, &
Y. Trope (Eds.), Dual-process theories in social psychology (pp. 323–360).
New York: The Guildford Press.
Spearman, C. (1904). General intelligence, objectively determined and
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
van Rooij, R., Krzyżanowska, K., & Douven, I. (2023). Williamson’s abductive
case for the material conditional account. Studia Logica, 111, 653–685.
Wang, M., Over, D. E., & Liang, L. (2022). What is required for the truth of
a general conditional? Quarterly Journal of Experimental Psychology, 75, 11.
Wason, P. C. (1960). On the failure to eliminate hypotheses in a conceptual task.
Quarterly Journal of Experimental Psychology, 12, 12–40.
Wason, P. C. (1966). Reasoning. In B. M. Foss (Ed.), New horizons in psych-
ology I (pp. 106–137). Harmandsworth: Penguin.
Wason, P. C. (1983). Realism and rationality in the selection task. In
J. St. B. T. Evans (Ed.), Thinking and reasoning: Psychological approaches
(pp. 44–75). London: Routledge & Kegan Paul.
Wason, P. C., & Evans, J. St. B. T. (1975). Dual processes in reasoning?
Cognition, 3, 141–154.
Wason, P. C., & Johnson-Laird, P. N. (1972). Psychology of reasoning:
Structure and content. London: Batsford.
Wilkins, M. C. (1928). The effect of changed material on the ability to do formal
syllogistic reasoning. Archives of Psychology, 16, 102.
Williamson, T. (2020). Suppose and tell: The semantics and heuristics of
conditionals. Oxford: Oxford University Press.
Woodworth, R. S., & Sells, S. B. (1935). An atmosphere effect in syllogistic
reasoning. Journal of Experimental Psychology, 18, 451–460.
Zhao, J., & Osherson, D. (2010). Updating beliefs in light of uncertain evidence:
Descriptive assessment of Jeffrey’s rule. Thinking & Reasoning, 16(4),
288–307.
Zimmer, K. (2021). People may have used fire to clear forests more than 80,000
https://doi.org/10.1017/9781009495349 Published online by Cambridge University Press
Keith Frankish
The University of Sheffield
Keith Frankish is a philosopher specializing in philosophy of mind, philosophy of
psychology, and philosophy of cognitive science. He is the author of Mind and Supermind
(Cambridge University Press, 2004) and Consciousness (2005), and has also edited or
coedited several collections of essays, including The Cambridge Handbook of Cognitive
Science (Cambridge University Press, 2012), The Cambridge Handbook of Artificial
Intelligence (Cambridge University Press, 2014) (both with William Ramsey),
and Illusionism as a Theory of Consciousness (2017).
Animal Minds
Marta Halina
Human Reasoning
David E. Over and Jonathan St B. T. Evans