Factive Theory of Mind Jonathan Phillips
Factive Theory of Mind Jonathan Phillips
Jonathan Phillips
Department of Psychology
Harvard University
&
Aaron Norby
Department of Philosophy
Yale University
1
Abstract
2
Factive Theory of Mind
a single chance to steal food (Santos et al., 2006). While the monkey was
watching, two translucent containers were placed on the ground and a grape
was placed in each. Both of the boxes had jingle bells glued to their lids,
but the ringers had been removed from the bells on one of the boxes, mak-
ing one of the boxes silent and the other noisy as they were handled by
put his head between his knees so that he could not see the monkey or the
boxes. Of the fourteen monkeys who attempted to steal food from the exper-
imenter, twelve of them went to the box without ringers, avoiding alerting
the experimenter. Other times though, instead of putting his head down,
When the experimenter was watching, the preference for taking food from
the silent box reversed itself: eleven of the sixteen monkeys tried to steal
the grape from the noisy box with ringers. They no longer seemed to care
others’ attention by pointing. They use this ability not only to draw adults’
attention to things that they themselves find interesting, but also to di-
rect adults’ attention to things that they believe adults are interested in
3
that were out of the experimenter’s view, and the experimenter then began
in. However, if the experimenter had seen the object move to a new location,
infants did not preferentially point to the object, despite the experimenter’s
One good thing about secrets is that they have a way of making almost
any social situation more interesting. Whether they involve illicit affairs
they’re something that others don’t know—not because they are one more
thing that you do know. When you find out that someone is going to have a
surprise birthday party, for example, it’s certainly not the fact that they will
have a birthday party that one finds interesting; it’s that they don’t know
that they are having a birthday party. And then there’s the intrigue that
goes along with knowing a secret: not letting on that you know something
that others don’t, or trying to secretly find out if someone else already knows,
or subtly indicating to others that you too know. And when things become
most interesting is if someone finds out that you know a secret. They know
that they don’t know, even if they don’t know what they don’t know. What
What’s compelling about each of these cases is that they illustrate the way
in which very different subjects seem to keep track of something about what
other agents know or understand and then use that information to achieve
4
their own goals. Moreover, they all are cases where the subject takes advan-
tage of the fact that others understand the world in a way that differs from
the way that they do. In short, these seem to be the kind of paradigmatic
dom, these cases don’t demonstrate the core ability required for a genuine
theory of mind because they don’t require the ability to represent false be-
liefs (for classic examples of false belief tasks, see, Wimmer & Perner, 1983
started with may be thought of as suggestive or intriguing, but the idea that
they are clear demonstrations of the essential abilities for a theory of mind
We must have taken a wrong turn somewhere. There are cases of clever
behavior that do not demonstrate a genuine ability for theory of mind (such
as gaze-following), but these cases aren’t like that. In some of them, our own
researchers have gone to great lengths to document the way that these sub-
jects can flexibly make use of others’ mental states in a way that is unlikely
the periphery and not at the center of theory of mind research. This seems
odd, and it’s worth taking a moment to look backward and consider how we
got here.
5
1.2 A brief history of false beliefs
ioral and Brain Sciences called ‘Does the Chimpanzee have a theory of
chimpanzees could identify the solution to problems that other agents faced,
that chimpanzees could recognize other agents’ goals or intentions, and thus
gued that this kind of evidence was not sufficient to demonstrate a capacity
for theory of mind (Dennett, 1978; Bennett, 1978; Harman, 1978; Pylyshyn,
1978). The problem, they pointed out, was that the experiments did not
chimpanzees could represent false beliefs. The reasoning was that repre-
senting a false belief requires that you represent a belief that differs from
your own. Otherwise you wouldn’t treat it as false. False beliefs guarantee
the necessary dissociation between representing the world itself and repre-
senting another agents’ mental states. Thus was born the litmus test for
Soon after the commentaries were published, Wimmer and Perner (1983)
their false beliefs. Shortly after that, Baron-Cohen and colleagues developed
6
and published the now classic Sally-Anne version of the false belief task in
their article, ‘Does the autistic child have a ‘theory of mind?’ ’ (Baron-Cohen
and Brain Sciences when explaining why false belief representation was the
appropriate test of the capacity for theory of mind. And now, forty years
later, we find ourselves teaching introductory students that the litmus test
for having (or developing) a capacity for theory of mind is the passing a false
belief test, and we’ve collectively written well over 8,000 papers employing
saying that any of us think that false beliefs are the entirety of theory of
mind; we’re saying that most of us treat them as the most important way
to test for theory of mind, and this seems to be just as true today as it was
in 1978.
The trouble with false beliefs is that they don’t carve the world at its joints.
There’s more to theory of mind than false beliefs, and much of the time
concerned with what others falsely believe. More often, we’re just interested
in keeping track of what they know (or whether they too saw something, or
if they recognize the person you’re pointing out, or whether they remember
that one time when...). Borrowing a term from linguistics, we can call all
7
they’re directly tied to the way we take the world to be. When you know the
in Paris. That’s just not the way that representations of seeing work. There
are no false seeings. The same is true for knowing, recognizing, realizing, and
so on. There are false beliefs though. You can also imagine things that don’t
thing about them is that they aren’t tied to the way you take the world to
actually be.
We seem stuck. On the one hand, we know that there are a lot of phe-
nomena that we all agree seem to involve keeping track of what others take
the world to be like, and then using that knowledge to predict or explain
others’ behavior. Yet, for all the reasons pointed out forty years ago, we’re
wary that these kinds of factive theory of mind don’t pass muster. They
may not be good evidence for genuine theory of mind representations, since
these representations are, by definition, tied to the way one actually takes
the world to be. Then, on the other hand, we have ‘non-factive theory of
mind’ which we are pretty certain is good evidence for genuine theory of
Up til now, most people have tended to agree we’re stuck, and so we’ve
gone with the sure bet. Demonstrating a genuine capacity for theory of
representations, and this has been the standard for how we demarcate, for
example, which species actually have a genuine capacity for theory of mind
and Kiparsky, 1970).
8
(Heyes, 1998; Call and Tomasello, 2008; Drayton and Santos, 2016; Martin
and Santos, 2016), when in the course of the human lifespan a capacity for
theory of mind develops (Onishi and Baillargeon, 2005; Wimmer and Perner,
1983; Kovács et al., 2010; Baron-Cohen et al., 1985), and even which brain
regions are responsible for representing and reasoning about others’ minds
(Saxe and Kanwisher, 2003; Gallagher and Frith, 2003; Gweon et al., 2012;
But we don’t agree that we’re stuck; we just think that we took a wrong
think there’s a principled way of showing when and how factive represen-
tations are genuine theory of mind representations. We also think there are
good reasons to believe that having a capacity for theory of mind doesn’t
and shouldn’t require the ability to represent false beliefs, or even beliefs at
all. And we also think that getting clear on this will help to reorient the way
we think about both the past and the future of theory of mind research.
That’s where we’re going, but we’re going to start by going back to the
basics.
There is still a great deal we don’t know about theory of mind. For example,
there’s some evidence that infants have the ability to succeed on non-verbal
false-belief tasks (Onishi and Baillargeon, 2005), and thus (barring non-
mentalistic confounds, e.g., Heyes 2014a) some evidence that they have a
capacity for theory of mind. But very little is known about the principles by
which infants are able to pass such tests, even at the most abstract level of
9
well. This is not to say that past forty years of theory of mind research have
not been both impressive and productive; they have. The point is rather
theory, simulation theory, and hybrid accounts all are aimed at understand-
ing how theory of mind ‘works’ and what it is. Simulation theory tells us
theory of mind involves simulating the mental states of others and running
the mental states of other agents are formed and influence their behavior
(Gopnik and Wellman, 1992). Hybrid theories tell us that theory of mind
shouldn’t be too hard to see that this is the case. Assume for the moment
that the way theory of mind representations are produced and manipu-
ally theory of mind representations and which are instead some other kind
course be produced by the same sort of simulation (e.g., what part of the
issue we are pointing out concerns what the computational role of these
10
representations must be for them to be theory of mind representations in
particular. It’s also important to notice that the answer to this question is
that support the representations of others’ minds are not themselves theory
involving any genuine theory of mind abilities (Andrews, 2012; Gergely and
Csibra, 2003; Penn and Povinelli, 2007; Perner and Ruffman, 2005; Butter-
fill and Apperly, 2013). The issue here is in no way specific to simulation
are theory of mind representations and which are not? And combining the
two, like hybrid theories do, obviously won’t solve the problem either.
representations, and then proceed to argue over which kind of processes un-
we haven’t actually spent much time working out an account of what the
even exactly what ability one has when one has the capacity for theory of
mind).
11
theory of mind are going to depend on how this issue ends up being re-
solved. As we pointed out earlier, consider the debate about when during
Baillargeon, 2005; Surian et al., 2007). The answer to this question is going
the debate over whether theory of mind is a uniquely human ability (Call
and Tomasello, 2008; Martin and Santos, 2014; Penn and Povinelli, 2007;
Lurz, 2011). And the same is true for which brain regions are responsible for
representing and reasoning about others’ minds (Saxe and Kanwisher, 2003;
Gallagher and Frith, 2003; Frith and Frith, 2012; Koster-Hale and Saxe,
2013). In all of these cases, we suspect that at least part of the disagreement
over these issues comes from the fact that we have been proceeding without
a more worked out idea of what it is exactly we are looking for when we’re
3 Mental Representations
It’s going to be helpful to start by setting aside theory of mind and instead
want to know whether human infants have the ability to represent number.
One obvious place to look is the experimental work that’s been done to
answer this question. Typically, the stimuli used in these studies are arrays
of dots, and the question asked is whether infants represent the number
infants’ behavior indicates that they differentiate between arrays with dif-
12
ferent numbers of dots (for reviews, see Carey 2009; Feigenson et al. 2004).
positions and then on a test trial be shown either a new array containing
seven differently arranged dots or an array that instead contains three dots.
the three-dot array than at the new seven-dot array. But even if the in-
fants do look longer at the three-dot array, what has to be shown is that
something else that covaries with number. Thus, experiments are designed
to vary number while holding fixed properties like density and area of the
array, the size of individual dots, and so on (Dehaene and Changeux, 1993;
Gallistel and Gelman, 1992; McCrink and Wynn, 2007; Xu et al., 2005).
The purpose of this methodology is clear: we want to show that the sub-
jects are tracking number and not something else. And if it can be shown
that their behavior is tracking number and not anything else that just hap-
it is often claimed that this tracking methodology can not only reveal that
number is represented but can also provide information about how number
that there are multiple systems for representing number, one of which pre-
Number cognition is only one example, and although the picture we’ve
in vision science, in neuroscience, and so on. The basic idea is that the
13
ability to respond differentially to some particular feature of the world (and
not something that simply covaries with it) is directly tied to the ability to
represent that thing itself. This typically isn’t taken to be the totality of
being able to take the intentional stance (Dennett, 1978). To take the inten-
Specifically, the beliefs that it would be rational to have given the evidence
that would rationally satisfy its desires (assuming that its beliefs are true).
ideally fit together. Moreover, such an ability will only really come as a
suite: you either have the ability to represent all of these mental states at
a level of relative complexity or you simply won’t be able take this sort of
stance. After all, it wouldn’t do one much good to take the intentional stance
either beliefs or desires in a way that they didn’t allow them to combined
14
On reflection though, we’re not sure that all of these commitments are
things that theory of mind researchers would really want to endorse. Sup-
the entity represented, then to show that infants are truly representing num-
ber and not something lesser, one will additionally need to show that they
understand, e.g., that numbers, being numbers, are by definition the kinds
number. One does not need to be able to understand number at this level
that were the case, we’re not sure whether most adult humans are capable of
representing number. And we can’t see why much the same thing isn’t also
true in the case of theory of mind. Requiring that one be able to represent
the falseness of beliefs is like requiring that one be able to represent the
negativeness of number. Sure, if you could do it, no one would doubt you
could represent others’ minds, but it’d hardly be a reasonable test of whether
you could had the capacity for representing others’ minds in the first place.
tation in the theory of mind domain in the same way that we frame our
to track the minds of others is the first step in being able to think any kind
of complex thought about others’ minds. It is the core of an ability for the-
15
ory of mind. With this alternative in hand, we want to start to ask what a
picture of theory of mind might look like once we let go of false beliefs.
and representing that same property or object, and here we want to turn this
into a full-fledged account of the capacity for theory of mind. The proposal
has two key requirements. The first is that an organism be able to track
that the organism be able to keep the outputs of its tracking mechanism
representation to the other, while using one’s own representation for action.
If both of these are satisfied, then the organism has the capacity for theory
of mind. Our claim is that this is the core of theory of mind, around which
described in § 3), then it’s obvious that the other agents’ understanding of
the situation is the thing that needs to be tracked if one is to represent that
perspective. This makes sense not only theoretically but also from a practical
some species or agent has theory of mind ability is to find out whether they
and thus having some way of representing the content of another agent’s
16
other words, in order to utilize theory of mind, I have to be able to predict
your behavior specifically on the basis of how you understand the world,
won’t have utilized any theory of mind capacity. Predicting what you will
In other words, to represent the minds of other agents in the sense im-
portant to theory of mind research, a subject has to be able not only to track
others’ understanding of the world, but also keep those representations sep-
both tracking and separation. Tracking demonstrates that you have a rep-
this representation plays the functional role that is appropriate for a theory
extra-mental world is like. This is the core ability one has when one has a
My own map is just my representation of the world, the way that I take
the world to be. Other agents have their own maps, each of which captures
their understanding of what the world is like. On this rough analogy, theory
a second map of the world in addition to my own), and keeping that map
17
separate from my own.
To see what this sort of representation could allow one to do, consider a
simple example. Imagine that you can see an opaque box on the floor, and
that you know that there is a banana inside the box. In the maps analogy,
this is to say that your map represents there being a banana inside of a box.
Now imagine that a second person comes along and can also see the box.
First, let’s suppose that you are incapable of constructing multiple maps—
all you have is your own map of what the world is like. That is, you have no
capacity for theory of mind. You could still go some way toward predicting,
manipulating, and understanding the behavior of this other person. You can
do this by (tacitly) treating the other person as though they have the same
map as you. Supposing for example that the person is looking for bananas,
you could predict that the person will go after the banana in the box. What
you won’t be able to do, however, is represent the person as being ignorant
Suppose now that you also have an altercentric map that tracks some
distinct from your own map. Let’s even suppose that it’s an incredibly simple
contradicts your own map. All you can do is either represent the other agent
as realizing the way the world actually is, or else as not being aware of certain
small pieces of it. To stretch the separate maps analogy, we can imagine that
you can construct altercentric maps by ‘removing’ parts of your own map
and then using this altered map to predict the other person’s behavior. With
this ability, you’d now be able to successfully represent the other person as
ignorant of the fact that the banana is in the box. Supposing that the person
is still looking for bananas, you’d have the capacity, for example, to realize
18
that the person won’t immediately look for the banana in the box. Critically
though, this sort of ability requires two functionally separate maps – one for
predicting what the other person will do, and a second one that you base
theory of mind can be had without the ability to represent non-factive men-
tal states. One can, for example, have a capacity for theory of mind without
being able to represent beliefs. If the core capacities of theory of mind are
cient for theory of mind. All that is required is a demonstration that these
it should radically change how we think about what is and is not evidence
for a genuine capacity for theory of mind. One easy way to make this dif-
et al. (2006). In this study, the animals faced a human competitor and two
visually identical boxes, each containing food. One of the boxes would make
noise if opened while the other would not. When the experimenter positioned
himself so that he could not see the monkey or the boxes, but the monkey
the silent rather than from the noisy box. In fact, this is something that
the animals figured out on their very first attempt after being presented
19
with the boxes. This latter fact is important, because it strongly suggests
that the animals were making inferences about what to do based on what
the experimenter knew, rather than applying some simpler behavioral rule
or associative connection that would link silent boxes with being able to
retrieve food.4
task? An obvious answer is that they must infer that they are more likely
to get food if they attempt to steal from the silent as opposed to the noisy
box. But how do they figure that out? Going after the silent box looks like a
good strategy only if the monkey realizes that its approach will not be part
of the competitor’s understanding of the world. That is, only if the monkey
realizes that the competitor’s representation of things will not include its
Thus, among other things, the monkey must be able to track what things
are like from the competitor’s perspective. Moreover, the animal’s own map
of the world must be kept separate from this representation of the com-
petitor’s understanding: even though the monkey’s picture will include the
approaching the silent box and that the experimenter doesn’t realize that
tation of what things are like from the competitor’s perspective, and using
20
ing that representation as a complete picture of the world, i.e., keeping it
functionally separate from the way the monkey takes the world to actually
be.
Importantly, the ability that we’ve argued these monkeys seem to be ex-
ing of the situation as including some facts but not others, while simulta-
only provide a capacity for factive theory of mind. This ability allows one
to represent which things others know and which things they don’t, but
there would still be things one couldn’t do. One couldn’t, for example, pass
It’s not hard to see why. Returning to the earlier example, suppose that
the other person originally saw the banana being put into the box, but
then didn’t see the same banana being moved to another location. Factive
theory of mind would allow one to represent the person as not knowing
where the banana is (i.e., attributing to them a map that does not include
the banana). But that won’t help you pass the false belief task; all that
would allow you to do is have no idea where the person would look for the
banana. Factive representations, like knowing, won’t help either. Your own
map doesn’t include the banana being in the box, so you can’t represent
the other agent as knowing that. Neither will it help to represent the agent
includes the banana being in the new location), since this would lead you to
make precisely the wrong prediction about where the agent will look. To be
need to construct an altercentric map that strictly contradicts the way you
take the world to be. Your own map of the world would need to represent
21
the banana not being in the box while you simultaneously construct and
say, you’d need a capacity for non-factive theory of mind. Without it, you
couldn’t represent the person as falsely believing that the banana is still in
the box, and you wouldn’t be able to predict that this is where they’ll look
Given how much we think hangs on this, it’s going to be worth getting
clear on exactly what the difference between factive and non-factive theory
of mind boils down to (and how they both differ from not having a capacity
for theory of mind at all). In a certain light, factive and non-factive theory
of mind can seem quite similar. In the former, you’re systematically tracking
what the person does not represent as being the case; in the latter, you’re
systematically tracking what else the person takes as being the case. When
understood this way, it’s easy to wonder what the real difference is supposed
on this issue is in large part what has led us down the wrong track in theory
of mind research, and that once we get clear on what the difference actually
is, a number of important (and missed) distinctions will begin to come into
focus.
5 Theory of Mind
the proposal, but alongside that, we’re also going to provide a slightly more
precise enough to get these details right turns out to be important, so we’ll
keep these formal details nearby for anyone who wants to see how all this
22
gets worked out. For the most part though, all of these details can pretty
our account and its implications for theory of mind at a broad conceptual
level.
Start with the basics. Take your own map to be the way you take the
relevant part of the world (call it the situation) to actually be. Not too
much will depend on it, but for the purposes of simplicity, let’s suppose that
we can represent each of the various things you take to be true about the
situation as propositions. We can now think of your map as just the set of all
in the box. We’ll take the proposition that there is a banana in the box,
and assume that your map includes that proposition, along with all of the
other things you take to be true about the situation. We can also think of
the map that represents the other agent’s understanding of the situation in
With this way of characterizing maps, we can now restate what the
core abilities of theory of mind are. Tracking requires that one dynamically
update the other agent’s map to reflect changes in the way that this agent
one must be able to add or remove propositions from one’s own map in a
way that does not demand a corresponding change to the altercentric map,
and vice versa. With just these few pieces, we have everything we need to
characterize the difference between (1) not having a capacity for theory of
mind, (2) having a capacity for factive theory of mind, and (3) having a
23
5.1 No theory of mind vs. factive theory of mind
If one does not have a capacity for theory of mind (either because one does
not track the other agent’s understanding of the situation, or because one
does not keep that map separate from one’s own), then the other agent’s map
will simply be identical to one’s own map.5 To the extent that one predicts
have to draw on a single map which represents both one’s own understanding
altercentric map that is not identical to one’s own. For example, the other’s
map may be a proper subset of one’s own. In this case, your map may contain
the proposition that the banana is in the box, but the altercentric map may
not. This is what it means to represent another person as not knowing that
the banana is in the box. This capacity to represent two non-identical maps
of the same situation, however, would not by itself allow for your own map to
situation itself does not involve contradictions, no subset of the things you
of the situation. Thus, when you take some thing to be the case, factive
thing, but it does not allow you to represent someone as representing that
24
5.1.1 What the difference is not
The distinction we are pointing to will likely become clearer when compared
Santos argued that the best way to make sense of non-human primates’
successes on a variety of theory of mind tasks is that they are able to track
other agents’ awareness of the situation. Stated in the terms we’ve been
the situation and can keep track of whether other agents are aware of that
situation—only using that map to predict their actions in cases where they
are aware.7 As Martin and Santos suggest, this capacity would not actually
about other agents (Martin and Santos, 2016, pp. 380-381). In cases where
another agent is aware of the situation, one can treat the other as having
one’s own map, and in cases where the agent is not aware, one simply has
no predictions about what the other agent will do (or just relies on one’s
priors over potential behaviors in that situation). On our view, this proposal
suggests that non-human primates have half of a theory of mind: they can
track whether or not an agent is aware of a situation, but they only have a
single map of the world, and so they definitely can’t keep the other agent’s
7
While Martin and Santos did not formalize their proposal, the most straightforward
interpretation would be that non-human primates have some way of tracking whether the
agent has awareness of the relevant situation, and if that condition is met, they attribute
their entire representation of that situation to the agent; otherwise, they do not attribute
any understanding of the situation to the agent. That is:
(
MS If other is aware
MO =
∅ Else
Obviously, in cases where the other agent is aware of the situation, our proposed charac-
terization of not having theory of mind, ∀p : p ∈ MO ⇐⇒ p ∈ MS , holds; in cases where
the agent is not aware, there is no representation, MO .
25
map separate from their own. This might be a particularly clever way of
An important thing that this comparison should help make clear is that
we don’t think theory of mind is a matter of whether or not you are able to
accurately predict others’ behavior based on the situation they are in. And it
you take someone to share your understanding of the situation. You’ve still
only got one representation of the world, and a genuine capacity for theory
But if this isn’t enough for theory of mind, then what is? Suppose that what
one does is not track just whether another agent is aware of the situation,
but instead tracks which parts of the situation another agent is aware of and
which parts the agent is not aware of. Stretching the maps analogy further,
suppose that you can do this by starting with your own map and simply
removing all of the parts of the map that the agent isn’t aware of (or is
mistaken about), leaving you with just a bunch of pieces of your original
map. Further, suppose that you then use only these pieces to predict what
the other agent will do but not the rest of the map. As long as you don’t lose
your own map of the situation in constructing this map for the other agent,
then as far as we’re concerned, you’ll end up with two functionally distinct
maps. Here’s why: take a case where you remove pieces from your own map
and then predict the other agent’s behavior based only on the remaining
parts. Great—but now, how do you decide what you yourself should do? If,
in making your own decisions, you rely on information that was not part
26
of the other agent’s map, then you must have retained the rest of your
own map in one way or another. How else could you have a more complete
the situation and keep that representation separate from your own, and
of mind: tracking and separation. Together, these two capacities give you
To see where the distinction we’re pointing to matters, let’s return to our
boxes of bananas. Suppose that I saw you put a banana in one of the boxes
(Box 1), but then when I wasn’t looking, you also put another banana in the
other box (Box 2). If you have only the abilities suggested by Martin and
Santos’ proposal (2016), then you have two possible approaches to predicting
where I would will look for a banana. On the one hand, you might no longer
entire situation, i.e., having your own understanding of the situation. But
if that’s the case, you should also predict that I might go to either box to
retrieve a banana, since I’ll understand that both boxes have bananas in
them.8 What you can’t do, however, is simultaneously know that there is
8
Let q1 be the proposition that there is a banana in Box 1 and q2 be the proposition
that there is a banana in Box 2. If you take me to be unaware of the situation, then
MO = ∅, and trivially, q1 6∈ MO , so you can’t predict my actions based on my knowing
q1 . If you take me to be aware of the situation, then MO = MS , and then obviously, if
27
a banana in both boxes but predict that I will retrieve the banana from
the Box 1 but not from Box 2.9 Being able to do that would require, at a
We’ve said something about the difference between not having a genuine
capacity for theory of mind, and having a capacity for factive theory of
mind, but even if you have a capacity for factive theory of mind, there will
still be things you cannot do. While factive theory of mind would allow you
own, it will not allow for your own map to be inconsistent with another’s
map, so you still couldn’t represent (false) beliefs. After all, as long as your
obviously no subset of the things you take to be true of the situation will
you take some thing to be the case, factive theory of mind allows you to
represent someone as not representing that thing, but it does not allow you
q2 ∈ MS , then q2 ∈ MO so you also can’t predict my behavior based on my not knowing
q2 .
9
It may initially be tempting to object by arguing that the proposal of Martin and
Santos (2016) need not be so all-or-nothing, and that a more charitable interpretation
would be that their proposal allows one to represent others as being aware of some situ-
ations, while not being aware of others (i.e., their awareness representation operates over
situations rather than entire maps). The trouble with this response is that this version of
Martin and Santos’s proposal is equivalent to our own, and ours straightforwardly allows
for representations of knowledge and ignorance per se, which Martin and Santos explicitly
deny their proposal allows for (Martin and Santos, 2016, pp. 380-381). It shouldn’t be too
hard to see why this must be the case. Suppose an agent is not aware of one situation,
S1 , but remains aware of other situations, {S2 , S3 , ...Sn }. To not be aware of a given sit-
uation, S1 is just to not be aware of the set of the facts that make up that situation, i.e.,
∀p : p ∈ S1 , p 6∈ MO . Similarly, to be aware of a situation is just to be aware of the facts
that make up that situation, i.e., ∀p : p ∈ S2 , p ∈ MO . Thus, in cases where the agent is
represented as being unaware of S1 , but aware of S2 , then straightforwardly, SO 6= ∅ and
∃p : p ∈ MS ∧ p 6∈ MO , which is a more precise way of saying that one has an ability for
factive theory of mind—an ability to represent knowledge and ignorance per se.
28
to represent someone as representing that thing not being the case.10
To be able to do that, one must have the capacity for non-factive theory
of mind.11 With this ability, the other agent’s map can both be non-identical
to your own and can also be inconsistent with it. Non-factive theory of mind
allows for it to be the case that your own map contains some proposition
and simultaneously, the other agent’s map contains the negation of that
proposition (or some set of proposition that are jointly inconsistent with the
of mind is that you can do something more than understanding that the
another person does not represent something: you can now understand that
With this more precise way of thinking about factive and non-factive
theory of mind in hand, it should be easy to see what the difference between
from your own; both factive and non-factive theory of mind require this.
The difference is just a matter of whether one can construct and maintain a
that is inconsistent with the way you take the world to actually be.
When seen for what it is, the difference between factive and non-factive
29
others’ understanding of the world. That is, it is not at heart a difference in
theory of mind. Rather, it’s a difference that concerns what kind of content
One easy to way to see that the ability that allows for non-factive the-
notice that the same ability also allows for other completely non-mental
what would happen if a certain state of affairs were to occur; when reasoning
of affairs had occurred rather than what actually happened. In both cases,
one makes predictions about what would happen when conditions are differ-
ent from the way you take them to actually be. However, only counterfactual
have happened if something had been different than it actually was — you
sentation in precisely the same way as success on the false belief task. In
both, you must construct and update a representation of the situation that
is inconsistent with the way you take the situation to actually be and then
tion that is inconsistent with your own. One can reason about what would
happen if p were the case (adding p to a hypothetical map) while simply not
13
See Byrne (2017) for a review of the empirical evidence for this distinction in online
processing.
14
See Kratzer (2012); Peterson and Bowler (2000); Grant et al. (2004) for related dis-
cussions.
30
having any belief as to whether or not p is actually the case. Just as with
is different from the way you take the situation to actually be, but it does
not require that one represent a situation that is inconsistent with your own
understanding.
The key point here is that the ability that allows one to move from
kind of ability that allows one to move from factive to non-factive theory
non-factive states of affairs. And it should not be surprising that if you can’t
represent states of affairs that are inconsistent with the way you take the
world to be, then you can’t represent another person representing states of
affairs that are inconsistent with the way you take the world to be. Perhaps
then, it also should not be so surprising that young children who cannot pass
simple verbal counterfactual reasoning tasks also cannot pass the verbal false
belief task, though they can pass strikingly similar hypothetical reasoning
tasks (Rafetseder et al., 2010; Riggs and Peterson, 2000; Peterson and Riggs,
1999; Riggs et al., 1998)15 In much the same way, it also should not be
surprising that people with Austism Spectrum Disorder have difficulty not
only with false belief reasoning but also with counterfactual reasoning (e.g.,
Peterson and Bowler 2000), while they have very little trouble with theory
of mind tasks that do not require representing non-factive content (e.g., Tan
The suggestion here isn’t that the ability that allows for both non-factive
31
ability; it clearly is an incredibly productive ability. Rather, the argument
We’ve illustrated that factive theory of mind allows for you to track other
still two different ways in which others’ maps could differ from yours. One
way would be for you to represent the other agent as being ignorant of
for you to represent the other agent as knowing something you are ignorant
agent as not knowing where you placed a banana while they were out of
the room. The ability to represent egocentric ignorance, on the other hand,
allows you to represent the other person as knowing where they placed the
represent both the agent as knowing more and less than you, since neither
32
resentations of altercentric ignorance. Here’s why: to represent altercentric
ignorance, you can simply take your map, remove the parts the other agent
doesn’t know, and then attribute that map to the other agent. But what
when they know more than you do? It’d be nice if you just take your own
map and then add the propositions that the other agent knows and then at-
tribute that map to the other agent. But obviously, you can’t do that. You
have no way of knowing which propositions the agent knows (if you did,
norance. If that’s right, though, then how do you do it? The solution to this
problem requires that you instead construct and attribute a more complex
To see why, just suppose that you don’t know where the banana is, but
you know that the other agent knows where it is. If your capacity for factive
theory of mind is working correctly, you should not be surprised if, on the
very first try, the agent looks for the banana where it actually is. Say, the
agent looks for the banana behind the door and finds it there. The reason
you would not be surprised by this is that you understood that the agent’s
map included a banana behind the door. Attributing such a map to the other
agent is a step in the right direction, but by itself, it isn’t yet enough to fully
capture your representation of the other agent’s knowledge. After all, you
also wouldn’t have been surprised if the agent looked for the banana in a box
instead and found it there. So you must have also been representing it as
possible that the agent’s map was one in which the banana was in that box.
The same thing is true for any location in the room where the banana might
be hidden, since you don’t know where the banana actually is. Generalizing
33
then, what we are left with is a set of all of the different maps that you
think the agent might have. We could think about this as a map of all of
the different possible maps. And this is precisely the kind of more complex
What you track and update is not a single map, but a map of maps.18
The basic insight is that in cases of egocentric ignorance, you know that
the agent knows something you don’t, but you don’t know what they know,
more complex than representing altercentric ignorance, since that just in-
volves attributing a single map to the other agent. Neither of these, however,
Unlike the distinction between attributing true and false beliefs, the
almost completely un-studied. And yet, once we’ve seen why the focus on
false beliefs was misplaced from the beginning, it should also be apparent
34
subject is not making predictions based on their own understanding of the
world. In both cases, these are important distinctions in the different kinds of
are distinctions we can make within a more general capacity for tracking
others’ representations of the world and keeping them separate from one’s
own. That is, they are both distinctions one can make within a more general
capacity for theory of mind, but what they are not, are distinctions that
out is how we should move forward empirically. We think the right way to
start is by first setting all of these distinctions to one side, and developing
a way to test for the core abilities of theory of mind: the capacity to track
another’s understanding of the world and keep it separate from your own.
In the next section, we lay out what we think the minimal version of such a
test looks like. With this test in hand, we then return to these distinctions
and show that it’s easy enough to extend the test in ways that allow for it
to distinguish between the different kinds of maps one can construct and
attribute to others.
As we laid out at the beginning of the paper, the current state of things
is that the litmus test for theory of mind is passing the false belief task.
However, given that the false belief task tests not only for an ability for
we want to propose an alternative task, which more precisely tests for the
35
core abilities of theory of mind without also making the representation of
these things a name, we’ll call it the diverse-knowledge task, for reasons that
will become obvious if they aren’t already. There are many different ways
will help to clarify how to test for theory of mind without relying on false
beliefs.
the diverse-knowledge task requires two responses from subjects: one which
provide evidence that they are tracking the other agent’s understanding of
the situation, and another which provides evidence that they are keeping it
One easy way to implement this is with the kind of situation illustrated
in Figure 1. The subject in the experiment is situated such that she can
see into two rooms, Room 1 and Room 2, each of which have two empty
boxes. The other agent in the experiment is instead placed such that she
can see what happens in Room 2 but not Room 1. With this setup in place,
a banana is placed in one of the two boxes in Room 1 (in view of only the
subject), and another banana is placed in one of those two boxes in Room 2
(in view of both the subject and the agent), and all of the boxes are closed.
Two tests are then required. The first (Fig. 2, left) is one that tests whether
she does, then when the agent is in Room 2, she should predict that the
agent will look for the bananas in the box where they actually are. At the
same time though, when the agent is in Room 1, the subject should be at
36
chance in her predictions of where the agent will look for the bananas. The
second test (Fig. 2, right) instead asks whether the agent has kept her own
retrieve the bananas from the box in which they were placed in either room.
Obviously, it’s not possible for a single map to give rise to both different
It may be helpful to compare this task and the original false belief task.
The false belief task was proposed as a way of testing whether representa-
the world (Dennett, 1978; Bennett, 1978; Harman, 1978; Pylyshyn, 1978).
the false belief task requires evidence for a single complex representation
agent’s own beliefs), we simply require that there be two non-identical rep-
37
Figure 2: Schematic depiction the Separation and Tracking conditions of the
diverse-knowledge task. Dotted arrows indicate the equal preference that
the agent or subject should have between multiple locations; solid arrows
represent the agent’s unequal preference for one location over another.
resentations of the world, one that the subject uses to guide her own actions
and one that the subject uses to predict the actions of others.
Once we have the most basic form of this paradigm in hand, though,
it’s not hard to see how it can be extended to test for the other distinctions
and egocentric ignorance. The previous test is sufficient for establishing the
ficient for factive theory of mind). To test for the additional capacity to
the initial setup. The subject should only be able to see that the agent can
see where the object is placed, but not where the object itself is placed. If
subject should not be surprised when the agent retrieves the desired object
38
from any location in the room the subject couldn’t see into. Passing this task
would provide prima facie evidence for the capacity to represent egocentric
ignorance.
The paradigm can also be extend to test for the capacity to represent
non-factive content by having the agent be in the room during the first
placement of the desired object, but out of the room when the location of
the object is switched. This is the equivalent of a standard false belief task.
And of course, it’s not hard to combine non-factive content with egocentric
ignorance either. One would first have the subject see only that the agent can
see where the desired object placed. Then, after the agent has left the room,
the object would be removed from that location, and the subject would now
While these ways of extending the diverse-knowledge task may help de-
termine the kind of content that subjects can represent in theory of mind
tasks, it bears reemphasizing that these are not tests of whether or not the
subject has a genuine capacity for theory of mind. For that, all one needs is
that separate from your own understanding of the situation. Evidence for
The long answer is that throughout we’ve clearly been making the ide-
tation of these tasks is not better explained by some behavioral rule (or
any other nonmentalistic feature) that happens to covary with the agent’s
understanding of the situation. What’s essential for any version of these ex-
39
periments is to detect whether and to what extent the subject is tracking
another agent’s understanding of the world and not anything else that hap-
pens to be confounded with that understanding (as laid out in §3). As great
as we think the diverse knowledge task is, it’s not magic, and ruling out
these kinds of confounds is just as critical for the diverse knowledge task as
it is for any other theory of mind task, including, of course, the false belief
task (see, e.g., Heyes, 2014b; Perner and Ruffman, 2005; Povinelli and Vonk,
7 Looking backward
From the beginning, our aim has not simply been to develop a new test for
theory of mind. Rather, it has been to offer a way for us all to let go of
While we’ve sketched one simple way to move forward by testing for only
the essential parts of an ability for theory of mind (§6), it’s also worth
taking a moment to look backward. The past forty years of theory of mind
research have, by and large, been conducted with the assumption that false
mind. Yet, if tracking and separation are what is actually essential, then the
literature from the past forty years should begin to take on a very different
light.
In some cases, research that has been taken to provide clear evidence
for theory of mind because it involves false beliefs should start to look more
40
that have sought to demonstrate theory of mind by showing that the calcula-
(see, e.g., Kovács et al., 2010; Apperly, 2010; van der Wel et al., 2014). While
these findings are clearly intriguing, they are not well-suited to provide ev-
the reason: to show an effect, these paradigms require that participants fail
to keep other agent’s understanding of the world separate from their own.
And if all of the evidence is evidence of a lack of separation, then it is, ipso
facto, not good evidence for what is essential for having theory of mind.
In other cases though, research that has not been seen as providing
good evidence for theory of mind (because it hasn’t involved false beliefs)
12 months will help an adult find an “adult” object (e.g., a stapler) that
has been hidden while the adult was gone, but not when the object was
hidden in the presence of the adult (Liszkowski et al., 2006, 2008). Or, to
to track others’ understanding of the world and keep it separate from their
own when stealing food (Santos et al., 2006). Similar patterns have also been
found with other non-human primate species using both auditory and visual
It’s worth reemphasizing that infants, great apes, and even monkeys
keys have any capacity for non-factive representations (Martin and Santos,
21
See §Section IV in the supplement to this article where we consider alternative expla-
nations and show how our framework provides straightforward ways of augmenting these
paradigms to ensure that they actually do provide clear evidence for factive theory of
mind.
41
2014, 2016), only very limited and mixed evidence in the case of great apes
(Kaminski et al., 2008; Krachun et al., 2009; Krupenye et al., 2016), and a
growing debate of over the case of non-human infants (Powell et al., 2018;
ting aside the potential for an inability for non-factive theory of mind across
all of these cases, the studies we’ve pointed to provide promising examples
of a capacity for both tracking what another agent understands and keep-
ing that representation separate from their own understanding of the world.
In other words, these tasks seem to provide good evidence for a genuine
Of course, the claim here is not that these infants or non-human primates
understand what mental states are in any kind of complex or reflective way,
that they are taking some kind of intentional stance, or that they have the
theory of mind in these terms is exactly what led us astray in the first place.
Rather, our claim is that there is good evidence that they can track others’
their own understanding. And if that is not theory of mind, we are not sure
what is.
42
Acknowledgements. Among many other people, we would like to thank
Lindsey Drayton, Lindsey Powell, Alia Martin, Jessie Munton, Laurie San-
tos, Brian Leahy, Enoch Lambert, the Yale Cognitive Science Reading Group
43
References
Psychology Press.
124.
Baron-Cohen, S., Leslie, A. M., and Frith, U. (1985). Does the autistic child
Sciences, 1(4):557–60.
Press.
28(2):141–72.
44
Dehaene, S. and Changeux, J.-P. (1993). Development of elementary nu-
5(4):390–407.
1(4):568–70.
Dörrenberg, S., Rakoczy, H., and Liszkowski, U. (2018). How (not) to mea-
sure infant theory of mind: testing the replicability and validity of four
Feigenson, L., Dehaene, S., and Spelke, E. (2004). Core systems of number.
45
Gergely, G. and Csibra, G. (2003). Teleological reasoning in infancy: The
1:158–71.
Gweon, H., Dodell-Feder, D., Bedny, M., and Saxe, R. (2012). Theory of
1868.
Hare, B., Call, J., and Tomasello, M. (2006). Chimpanzees deceive a human
science, 17(5):647–659.
46
Heyes, C. M. (1998). Theory of mind in nonhuman primates. Behavioral
Kaminski, J., Call, J., and Tomasello, M. (2008). Chimpanzees know what
Kovács, A. M., Téglás, E., and Endress, A. D. (2010). The social sense:
330:1830–34.
Krachun, C., Carpenter, M., Call, J., and Tomasello, M. (2009). A com-
petitive nonverbal false belief task for children and apes. Developmental
Science, 12(4):521–535.
Krupenye, C., Kano, F., Hirata, S., Call, J., and Tomasello, M. (2016). Great
apes anticipate that other individuals will act according to false beliefs.
Science, 354(6308):110–114.
Leahy, B., Rafetseder, E., and Perner, J. (2014). Basic conditional reasoning:
810.
47
Liszkowski, U., Carpenter, M., Striano, T., and Tomasello, M. (2006). 12-
130:300–8.
Nagel, J. (2017). Factive and nonfactive mental state attribution. Mind &
Language, 32(5):525–544.
University Press.
48
Onishi, K. H. and Baillargeon, R. (2005). Do 15-month-old infants under-
Perner, J. and Ruffman, T. (2005). Infants’ insight into the mind: How
Powell, L. J., Hobbs, K., Bardis, A., Carey, S., and Saxe, R. (2018). Repli-
81(1):376–389.
49
Ramsey, W. M. (2007). Representation reconsidered. Cambridge University
Press.
Macaca mulatta, know what others can and cannot hear. Animal Be-
haviour, 71(5):1175–1181.
19(4):1835–1842.
Surian, L., Caldi, S., and Sperber, D. (2007). Attribution of beliefs by 13-
van der Wel, R. P., Sebanz, N., and Knoblich, G. (2014). Do people au-
Cognition, 130(1):128–133.
50
Wimmer, H. and Perner, J. (1983). Beliefs about beliefs: Representation and
Xu, F., Spelke, E. S., and Goddard, S. (2005). Number sense in human
51