Proust2007 Article MetacognitionAndMetarepresenta
Proust2007 Article MetacognitionAndMetarepresenta
Proust2007 Article MetacognitionAndMetarepresenta
DOI 10.1007/s11229-007-9208-3
Joëlle Proust
A good part of our mental life is devoted to evaluating our mental performance, and
predicting how well (or badly) we can do, have done, or are doing in a new job, a new
task, or a new social situation. This is the domain of metacognition: thinking about
one’s own thinking. It is exemplified in all cognitive activities in which one is trying
to appreciate, in retrospect, a cognitive achievement (did I do this task right?, haven’t
I forgotten something?), to remember the source of the information one is using, or
to predict whether one will be able to attain some cognitive goal (learn new material,
retrieve a proper name within seconds, or make efficient plans in a new context).
J. Proust (B)
Institut Jean-Nicod (CNRS-EHESS, ENS),
29 rue d’Ulm, 75005 Paris, France
e-mail: [email protected]
123
272 Synthese (2007) 159:271–295
Most researchers in theory of mind have taken for granted that metacognition, in this
sense, necessarily involves the second-order representation, i.e., the metarepresentation,
of first-order cognitive contents. In other terms, modularists as well as theory-theorists
have assumed that the capacity of three-year-old children to attribute knowledge to
themselves, as well as the capacity to distinguish a real from an apparent property
of the world (reviewed in Flavell 2004) develop as a consequence of the children’s
mindreading skills, and more specifically, are made possible by their having metarep-
resentational ability.
When trying to explain why children attribute true or false beliefs to others (solve a
false belief task—or FBT) at around 4 and a half, most theoreticians have emphasized
that children only succeed when they are able to metarepresent the corresponding
mental states. Radical simulation theorists (Gordon 1996) have rejected this assump-
tion. Their claim is that children succeed at FBTs when they are able to simulate the
perspective of another child on a given situation, when it is different from the one they
know themselves to hold. Radical simulation theorists, however, have not produced a
fully adequate explanation of children’s understanding of false belief.1 Non-radical or
“hybrid” simulation theorists, on the other hand, have recognized that decoupling sim-
ulations and reasoning across them—rather than reasoning within a simulation—was
a key ingredient in a mindreading capacity.2 Such a decoupling has been shown to be
semantically equivalent to forming a metarepresentation of a first-level simulated con-
tent. For reasons that will not be discussed here, exploiting a simulation from without
amounts to forming a metarepresentation in the broad sense of the term (decoupling
the interpreted content from a larger set of contexts where it can receive different truth
evaluations).3
On the basis of this large consensus, this article will take it for granted that meta-
representation is necessary to ensure the kind of decoupling involved in mindreading.
Furthermore, it may be a precondition for certain forms of metacognitive effort, in par-
ticular those that are regulated by explicit self reference. An interesting and debated
question, however, is whether metarepresentation has to occur for metacognition to be
possible at all. This question is interesting, for if it can be shown that metacognition
can be dissociated from metarepresentational capacity, one might infer that different
evolutionary pressures are at play in metacognition and in mindreading. There is, as
will shall see, some evidence that metacognition may be present in animals without
a theory of mind. These results are still controversial, however, because the animal
data on which the case for non-metarepresentational metacognition is based raise
methodological problems that may interfere with interpretation of the results.
There is another dimension to the question that has theoretical import: metacog-
nition is by definition a form of self-predication; metarepresentation, however, is not
essentially reflexive. As a semantic skill, it can be used in self- or other-attribution, even
1 For critical reviews of radical simulation theory, see Dokic and Proust (2003) and Nichols and Stich
(2003).
2 See Perner (1996), Nichols and Stich (2003), Proust (2003b). The contrast between exercising simulation
and exploiting it allows one to contrast performance in 3-year-olds and 4-year-olds (Recanati 2000). We
will come back to this contrast in Sect. 1.
3 See Recanati (2000) chapters 4–7 for a detailed defense of this claim.
123
Synthese (2007) 159:271–295 273
if it can be argued that having access to one’s own beliefs and desires is more imme-
diate than inferring beliefs and desires in others. We will review below the conceptual
and empirical arguments that can be adduced to show that there are non-metarepre-
sentational forms of metacognition.
The central section of this article will be devoted to an examination of the main
defining features of metacognition. We will further examine how a decision about the
respective structures of metacognition and metarepresentation impacts on issues such
as epistemic transparency and the ability to pursue recursively into higher orders.
We will begin our discussion, however, by observing that an utterance constructed
as a metarepresentation can admit various cognitive interpretations. The distinction
between deep and shallow types of processing will turn out to be quite important in
discussing the relations between metacognition and metarepresentation.
123
274 Synthese (2007) 159:271–295
123
Synthese (2007) 159:271–295 275
123
276 Synthese (2007) 159:271–295
to be able to form the metabelief, and be committed to the truth of the first operation
to infer the truth of the second.
6 Other metacognitive predicates seem also to be subject to the same limitation: “believe”, “be aware”,
“feel”, “doubt”, etc.
123
Synthese (2007) 159:271–295 277
(6) “I know that Anna knows that her father knows that her mother knows that her
grandmother knows that she is invited for lunch on Sunday”.
What kind of account can we offer for this difference? Our previous distinction
provides the beginning of an answer. Our limitation does not reside in an inability to
understand the formal possibility of recursion. Indeed logical systems capture in exact
terms the properties of recursion. A written formula for a recursive proposition is easy
to learn and to reproduce. Nor does the problem reside in an inability to form self-
attributive judgments, for we succeed at the second-order level. It resides rather in our
limited ability to engage first-personally in a self-directed n-order recursive thought
for n greater than one. Although we understand fourth- or fifth-order attributions in
the third-person case, we fail to simulate self–self–self simulations as being distinct
cognitive operations.7 While the third-person case can be accounted for by the recur-
sive properties of metarepresentation, the first-person limitations can be explained by
the properties of metacognition.8 To complete this point, we need to turn to these
properties.
There is a form of engagement that explains “rich” understanding in both the Last
Supper case, and in the mental case of belief/desire self-ascription. Constraints associ-
ated with such an engagement should account for the limitations that apply to recursive
self-attribution. We will now attempt to list the various properties that jointly define
metacognitive engagement. Our goal is to discuss these various properties to get a
clearer picture of the relationship between metacognition and metarepresentation.
7 Richard Carter has objected here that a sentence including different tenses works better: “I know that
I knew (yesterday) that I knew (the day before) etc.” On his view, this suggests that what makes it hard
to find any context or interpretation in which “I know that I know that I know that P” is true, is that the
three “knows” have present reference. Such contexts however can be constructed: “I know that I know that
I have always known/that I will always know that P” has a clear interpretation. Carter finally observes that
the limitations also hold for the corresponding third-person thoughts: “John knows that he knows that he
knows that P” (where the two “he’s” refer to the same John). Carter’s objections can both be addressed in a
similar way. In both cases (adjunction of tensed contexts, and third-person attribution), metacognition and
metarepresentation overlap. When representing my epistemic relation to P at other times, and comparing
them, I am not scrutinizing my present cognitive adequacy (as in “Do I presently know P?” or in “May I
predict whether I will remember P”?). I am not therefore performing a metacognitive action; I am rather
using the mental concept of “knowing” in its fully general scope across various metacognitive episodes,
which requires using a metarepresentation of various metacognitive states. Conversely, it is as difficult to
understand “John knows that he knows that he knows that P” as the first-person parallel sentence, because in
both cases, the only interpretation that would make sense is the “engaged” one, i.e., an interpretation gained
through metacognition. As simulationists have argued, self-simulation is a necessary ingredient in many
kinds of other-attributions (those that qualify as “engaged”). It is an interesting consequence of the present
view that overlaps between metacognition and metarepresentation can be apriori predicted/explained on
the basis of the engaged/shallow distinction.
8 Compare with the preceding footnote: each “know” has a subject with a different reference, unlike the sen-
tence above with “John”. This allows creating metarepresentational contexts from a succession of engaged
metacognitive meanings.
123
278 Synthese (2007) 159:271–295
123
Synthese (2007) 159:271–295 279
123
280 Synthese (2007) 159:271–295
As we saw in Sect. 3.1, control and monitoring are the two relata of a closed causal
loop. This property can be rephrased by saying that there is causal contiguity between
the control and monitoring levels (which means that control directly affects monitor-
ing, which in turn directly affects further control). One crucial aspect of this dynamic
reafference is that it retains the causal structure of on-line control-monitoring. What
is retrieved in memory is retrieved as a result of the present intention to retrieve it. The
metacognitive prediction (“that I know the name of Peter’s daughter”) is a response
to a prior command with a closely related content (“Do I know the name of Peter’s
daughter?”). Causal contiguity is the solution that Hugh Mellor14 has developed for
explaining what distinguishes what he calls “subjective beliefs” (such as “I face food
now”) from “objective facts” (such as “X faces food at t”), which are fully explicit
truth-conditional expressions of states of affairs. What makes a belief subjective is that
causal contiguity guarantees that an individual’s belief will make his desires cause him
to act to satisfy them. “Causal contiguity is how a subjective belief refers to whomever
has it, and when”. (Mellor 1991, p. 24).
Causal contiguity, as a general causal structure, belongs to every adaptive control
system where feedback comparison is made to causally respond to command in a sys-
tematic way. In our case, however, the control system is cognitive, which means that
representations—beliefs and desires—are used to control behavior. As philosophers
of action have shown, using various terminologies, causal contiguity relations between
the cognitive subsystems engaged in action are associated with an intriguing concep-
tual relation. If you intend to act, your intention constitutes the condition of satisfaction
of the completed action.15 This property of an intention to be embedded in the con-
tent of an action, is what I will call representational promiscuity.16 In every cognitive
control structure, there is such representational promiscuity. Monitoring instructs the
command, while the command directs or organizes upcoming monitoring (by offering
predictions about what is to be the case if the action develops normally). There is repre-
sentational promiscuity because control and monitoring share the basic informational
pattern that drives the causal loop to the expected goal.
As noted by Mellor (1991), an important consequence of causal contiguity is that
self-reference does not need a self concept to be instantiated. More generally, there can
be token-reflexive dimensions of thinking with no corresponding concept; the modes
of presentation that are expressed by words like “here”, “now”, or “I” can be exhibited
by organisms that do not master the corresponding concepts, but are equipped with the
corresponding causally contiguous control-informational device. In other words, the
very structure of the causal and informational exchanges between control and monitor-
ing builds reflexivity into the operating mode, and into its output. Such “procedural”
or architecture-bound reflexivity is more than a precursor of explicit semantic reflex-
ivity (as expressed by indexical words). It is the locus where all forms of reflexivity
123
Synthese (2007) 159:271–295 281
are generated. Where there is reflexivity, there must be an adaptive control system
engaged in monitoring its commands.
Every token of metacognition can provide illustrations for this procedural reflex-
ivity: a student who is rote-learning a poem, for example, must appreciate whether
she now knows the poem; an aging speaker must appreciate whether she will be able
to remember a proper name in a minute (from now), etc. This dependence of opera-
tion on self-evaluation and various task-relative (temporal and spatial) indexicals does
not need to be achieved in explicit semantic terms as long as self-reference is the
indefeasible background default of the operating system.
To qualify as procedurally reflexive, a further condition must hold, and it deserves
to be spelled out even though it results from the conjunction of causal contiguity and
representational promiscuity. A procedurally reflexive system must not be a mere com-
bination of distinct devices that happen to produce a joint effect via a causal loop; it
must have the intrinsic function of affecting itself, and do so in a causal-representational
way (through informational means). In other words, there is an evaluative dimension
in procedural reflexivity that should appear as a definitory feature of metacognition.
When producing an evaluation of its own operating mode, the system reflects its intrin-
sic reflexivity: being closed under a norm, the cognitive system generates commands
to change itself in order to adjust to a changing environment. As we saw in Sect. 3.1,
the decision to initiate/revise a command is produced by an information-based (sim-
ulatory) estimate of the probability of success/failure at a task. Success or failure are
outcomes that motivate the agent to metacognize (i.e., predict and evaluate) its own
states.
In contrast to (metarepresentational) mindreading, metacognition is always evalu-
ative, rather than merely predictive. You can predict, for example, that Jane intends
to eat bananas rather than pears, but this prediction does not involve your own prefer-
ences (even though it involves an evaluation by Jane of what is best for her). Predicting
future states generally implies anticipating trajectories in internal (mental) or external
(physical or behavioral) dynamic events or event sequences. Evaluating future states
involves in addition appreciating the efficiency of a given course of action, which
means comparing internal resources with objective demands for the task. A judgment
of learning, or an evaluation of one’s emotional level, for example, involve norms of
adequacy: the goal of such judgments is to find an efficient or reliable way of coping
with a set of requirements.
Note that ordinary bodily action also requires flexibility and adjustment to local
conditions, and these corrections are part of what “intending” or “willing to do P”
mean. In metacognition, however, flexibility and adjustment concern the informational
resources presently available. A subject cannot change much of the set of epistemic
or motivational resources available to her at any point in time, but she may acquire
control over pragmatic aspects of her thinking that crucially affect its outcome. She
may plan ahead of time to gather resources for a demanding task. The time she spends
on it (looking, considering, reasoning), the acuity with which she perceives an object,
the chosen distance from and orientation to the perceptual input, the duration of the
mental effort given to collecting information, all these correlate with significant dif-
ferences in the validity of the output. In sum: an agent cannot change the content of her
123
282 Synthese (2007) 159:271–295
attitudes, but she can improve the informational quality and the cognitive adequacy
of her mental processes in a controlled way.17
Critics of metacognition argue that all that is needed for “metacognition” is a capacity
to predict the Bayesian likelihood with which a certain desired outcome will occur.
A dedicated neural reward system encodes disparities between an animal’s expecta-
tions and its experience of success. In other terms, it encodes errors in the predic-
tion of future reward.19 Such outcome evaluation, the criticism goes, can proceed
on a first-order cognitive basis: actions will be selected if they maximize rewards in
123
Synthese (2007) 159:271–295 283
123
284 Synthese (2007) 159:271–295
trials when it was uncertain.(Smith et al. 1995) Animals used the third pedal as human
subjects do. It is now recognized by the authors that the “uncertain” response might
have been reinforced to cope with objective uncertainty. In that case, it would not
qualify as metacognitive. Rhesus monkeys have also been tested in their ability to per-
form, or to bail out from, a visual density discrimination task (Smith et al. 2003). The
animals had to discriminate whether a display contained 2,950 pixels. They responded
in positive cases by moving a cursor to a box, in negative cases by moving it to an
S shape, and to a star to decline the trial. Monkeys have again been found to have a
pattern of response similar to humans.
The uncertain response is known to have a specific profile, both in humans and
in monkeys: it is quite variable across individuals. For this reason, some researchers
consider that it is prompted by “extraserial attitudes and decisional temperaments”
(Smith 2005). A potential problem for this experiment, however, is that the animals
have to respond to an occurrent situation. It is particularly difficult, in a simple per-
ceptual paradigm, to tease apart objectively produced and subjectively appreciated
uncertainty.23
A research project responding to this worry has been conducted using a memory
monitoring paradigm: a serial probe recognition task. In one version of this task by
Hampton (2001), monkeys are presented with an icon, that they will subsequently need
to recognize among distractors after a varying delay, in order to obtain a reward. A
first crucial constraint is that they decide to take the test (after the delay) before seeing
the test stimuli, or rather chose an alternative, easier task. Taking this decision before
seeing the test stimuli, they have to rely on their memory of the icon. As Hampton
emphasizes,24 this task elicits a prospective judgment of memory in monkeys. To solve
it, monkeys must rely on the equivalent of our human feeling of knowing. They cannot
base their response on the perceived familiarity of the test stimuli, for they have not
yet seen them: they cannot consult the world to decide what to do.25 Exposed to the
same material, pigeons show that these two conditions are critical.26 They can only
perform the task when they perceive the test stimuli, not on the basis of their memory
of the icon. These experimental results indicate that animals without a mindreading
capacity, such as monkeys or dolphins, can still succeed in using metamemory. This in
turn suggests that mental metarepresentation might not be the initial input that makes
metacognition possible.
Let us summarize this point. Metacognition, when it is present, draws on a kind
of information that is not delivered by the problem situation, but by the subject’s
own procedural self-knowledge. For that reason, metacognition can deal with novel
decisions, while well-practiced routines remain within the scope of cognition (where
external cues can be used as predictors). These two features seem to provide a divide
23 Smith and colleagues addressed this difficulty since then in Smith et al., 2006. See also Kornell et al.,
2007.
24 Hampton (2005).
25 That absence of memory was causing the response was experimentally controlled by introducing trials
where no icon was presented. The decision to decline is also present in these cases, suggesting that the
animals did not learn to predict result from such cues as noises, grooming, or motivational changes.
26 Inman and Shettleworth (1999).
123
Synthese (2007) 159:271–295 285
between the capacities exemplified in primates and marine mammals, on the one hand,
and in pigeons and rats,27 on the other.
27 For a recent study on rats, suggesting meta-cognition ability, see Foote and Crystal (2007).
123
286 Synthese (2007) 159:271–295
123
Synthese (2007) 159:271–295 287
123
288 Synthese (2007) 159:271–295
It might be objected, however, that when a subject stores the informational source
from which she acquired a reflective belief, as seems necessary in order to utter sen-
tences like (13), she certainly needs to use her own metacognitive competence. For to
store the original conveyor of the information that p, the objection goes, one must be
able to engage in a simulation of the way one acquired that information. Therefore deep
first-person engagement is needed in most reflective, metarepresentational beliefs.
Although this kind of analysis is quite popular among mindreading theorists, it
may not square well with recognition of the contrast stressed in Sect. 3. The point of
contrasting “having a capacity” and “exercising it” was to show that, although one
can use one’s mentalizing abilities in many circumstances, one can also approach a
question in non-mental terms, by simplifying the problem. There are deep forms of
engagement available to an agent who has access to metarepresentational thinking,
because normally such an agent independently possesses metacognitive capacities.
But she does not need to exercise them, in particular when under time pressure, or in
routine situations. For example, simple association of a sentence with an agent [Anna,
p] can replace a full-blown, engaged simulation of the corresponding belief activity
(with its network of normative constraints). A subject who already has these mental
concepts in her repertoire may replace them by their shallow, non-mental counterparts.
Such counterparts, furthermore, have been recognized by several authors. Among
psychologists, Josef Perner has observed that although human adults can use a meta-
representational approach to fully understand the mental import of a conversation, or
of a social context, they don’t need to. In most cases a “situational” approach to the
problem at hand will help one recover the information that needs to be shared and
jointly processed. Processing a social context through “situation theory” amounts to
using the world as the model from which to predict behavior, instead of using another
subject’s representation of the world. We resort to representational theory “only when
we need to”: “remaining a situation theorist whenever possible can save many unneces-
sary complications … it saves the gory details of reshuffling mental representations.”32
The reasons Perner offers for staying with situation theory processing of contexts are
the same we offered in our discussion of having/exercising a capacity to exploit deep
levels of processing: simplicity (and reliability given the task at hand).
Among philosophers, Gareth Evans described a shallow understanding of belief,
through a procedure that has come to be called “the ascent routine”.
123
Synthese (2007) 159:271–295 289
Evans’s point (also made by Gordon 1996), is that non-mentalizers can make shal-
low linguistic use of a metarepresentation such as “I desire that p”, or “I want that p”.
Looking at the world, seeing what is the case, and reporting it as a belief may be done
in conversing with someone without deploying the concept of belief: the procedure
of embedding a content within an epistemic verb may be socially learned before one
has realized what “belief” actually means.
It is interesting to observe that, although Evans’s point was made to account for
a shallow form of metarepresentation, the phenomenon he describes may also be
interpreted as a deep metacognitive (non-metarepresentational) self-engagement. As
suggested in the preceding section, the verbal expression “I believe that p” should be
analyzed differently if it is an explicit answer to a linguistically expressed utterance
(as suggested by “answering” a question, in which case it is metarepresentational) or
to a mental token of metacognition (the question “answered” is raised by the thinker
herself, in which case it is metacognitive and not necessarily metarepresentational).
A non-mentalizer might thus entertain in thought a metacognitive equivalent of “I
believe that p”—in a way that requires neither a metarepresentation nor the mastery
of the concept of belief. This is the case, for example, when the subject has reasons
to be subjectively uncertain of a situation P (where P is not objectively uncertain:
for example her memory for a given event is weak). What forming this metacogni-
tive self-attribution does require, however, is that the subject engage in appropriate
self-simulation (comparing the feedback obtained to a stored norm).
The expression “self-simulation” may be somewhat misleading, if one takes it to in-
volve “turning inward”. In self-simulating herself, a subject also “turns to the world”—
but in a “deeper” way, through dynamic prediction of reafferences and evaluation.33
If these observations are correct, there are several ways of opposing a “precursor”
to a “genuine” “I believe”. The metarepresentational type of precursor “I believe” is
what Evans had in mind. In that case, an utterance of “I believe” merely emphatically
expresses P. In contrast, a metacognitive precursor “I believe” is not verbally expressed,
but rather procedurally exercised by giving attention to the cognitive adequacy of (what
is in fact) the belief that p.
The full-fledged deep metarepresentational form of “I believe” requires exercising
both the capacity to metacognize that I believe and the ability to master the concept
of belief in a general way, i.e., to apply it in third-person attributions.
33 “Turning to the world”, let us note, does not mean only “turning to peripersonal space”. Bodily states
belong to the world, as do the contents of memories which were generated by “external” states of affairs.
It is arguable that the epistemic feelings generated in the monitoring part of the metacognitive loop have a
“mind to world” direction of fit: the thinker who enjoys these feelings extracts information on her mental
dispositions from her bodily states. The fact that, most of the time, we perceive the external world through
vision should not obfuscate the fact that proprioception is always an important mode of perceiving the
world, our body being a part of it. Most metacognitive reafferences are proprioceptive rather than visual.
See however, the fascinating results discussed in Levin (ed.) (2004).
123
290 Synthese (2007) 159:271–295
The preceding section allows us to dissipate the puzzle about recursivity and trans-
parency presented in Sect. 2. On the one hand, transparency holds in metacognition,
not in mindreading: When I know or believe something, I immediately know that I
know or believe it, but I do not immediately know what you know or believe. Your
mental contents are not transparent to me: I need to infer them from your behavior.
This is intuitive, but it should now also be theoretically obvious: the only system that
can be used as a direct model of a knowledge state is the system that instantiates the
corresponding internal model of the world. It is in virtue of this model structure of the
mind that each mind is only equipped to directly simulate its own belief states.
In order to extend simulation of one’s own mental states to the mental states of
others, as one does in mindreading, one will have to modify the model of the world
to be simulated, that is, construct an ad hoc simulation in which some crucial beliefs
are missing or have different truth values. In this process the notion of evaluation will
be lost in part: one cannot compare in an engaged way (metacognitively informed)
the reafferences imputed to another subject with the norm that subject acquired via
his previous experiences. The modified dynamic model does not coincide with the
self-simulation dynamic model either, but only overlaps with it. This is why self-
simulation can only partly account for mindreading, and needs to be supplemented
with third-person kinds of generalizations. In mindreading, the evaluative component
thereby becomes less prominent than the descriptive–predictive component.34
Now the link between transparency and recursivity becomes clearer. The reason
why recursivity is open to modeling multiple others’ embedded mental states rather
than a single individual’s is that this operation relies on metarepresentation, which
relies on the syntactical phenomena of natural languages. It is indeed a universal for-
mal property of human languages that they admit embedded clauses. If our hypothesis
above is correct, metarepresentation is not as such cognitively demanding. It is implic-
itly mastered through language use. Embedded clauses are understood quite early in
life, although children typically make mistakes in processing correctly the information
delivered by such sentences when the clauses conflict in meaning or reference.35 An
experienced speaker, however, can sort out the variations in references and modes
induced by example (6), repeated here:
(6) “I know that Anna knows that her father knows that her mother knows that her
grandmother knows that she is invited for lunch on Sunday”.
Some features of (6) help us in parsing the various attributions of mental contents: first,
a “mental” word followed by the syntactical marker (that) introducing an embedded
clause signals a new attribution; second, the different proper names work as social
cues on which to anchor a new attributional verb and a new content (spatial cues
used in gesturing can also be used to index perspective). Third, the temporal succes-
sion of the clauses is a cue made linguistically available and rehearsable. Iteration
123
Synthese (2007) 159:271–295 291
123
292 Synthese (2007) 159:271–295
To respond in full to these important objections would require more space than
is available here. I will however summarize the various methodological and theo-
retical considerations that can be adduced in defending level 1.5 of metacognitive
processing. Consciousness is often identified with person-level information process-
ing, taken to constitute an adequate grounding for self-engagement. The present
approach suggests a different picture of self-engagement, as well as of the emergence
of a self-representation. The present proposal is that metacognition is an essentially
reflexive mental function allowing an explicit form of self-representation to eventually
emerge.37 There are however forms of self-engagement that do not require the con-
scious mode, as we already know from neuroscientific evidence. Insofar as metacog-
nition is the crucial capacity for building up self-identity, its important property is not
consciousness, understood as the capacity to verbally report one’s mental states or
to attribute them to oneself, but reflexivity, the capacity to evaluate and revise one’s
cognitive states (whether epistemic or conative). Therefore, our definition of meta-
cognition, to be non-circular, should rely on normative forms of self-guidance, rather
than on a full-fledged representation of oneself.
Several researchers have recently defended a metarepresentational view of metacog-
nition in combination with a higher-order-theory of consciousness (Rosenthal 2000a,
b; Dienes and Perner 2002). The view is that conscious seeing, remembering, acting,
etc., require some form of metarepresentation. Metarepresentation, when it is applied
reflexively to one’s first-level thought in an immediate way (not mediated by infer-
ence), makes the first level representation conscious. The second-order thought, in
turn, is made conscious when it is immediately represented by a third-order thought.
In this view, metacognition, being cognition about one’s cognition, is seen as a repre-
sentational mechanism for producing conscious thoughts. A twist in the theory is that a
rich conceptual articulation of the relation of the experience to the event that caused it
is not necessary for consciousness (although it is for mindreading). Self-referentiality
that is inherent to metacognition does not need to be made explicit, a view closely
similar to the view defended here.38
Metacognition, on the view of these higher-order theorists, has two functions.
The first is to make mental states conscious by making them the content of higher-
order thoughts.39 Metacognitions make first-order contents conscious by providing
the non-inferential metarepresentation that makes a mental state conscious. They are
not themselves conscious, however, unless a third-order thought metarepresents them.
The second function is inherent to the semantic structure of a metarepresentation: it
is that of applying [implicit or explicit] concepts to first-order contents, which in turn
provides the inferential structure needed for reasoning about one’s own states as well
as those of others. It is important to note that, on this view, the concepts used are used
tacitly unless they are metarepresented in appropriate third-order thoughts.
123
Synthese (2007) 159:271–295 293
Acknowledgements I am deeply indebted to Radu Bogdan and to Richard Carter for their help with lin-
guistic form and philosophical content for an earlier version of this article. I am also grateful for discussions
with Jérôme Dokic, Paul Egré, Josef Perner, François Recanati, and the participants in the APIC seminar
at the Institut Jean-Nicod.
123
294 Synthese (2007) 159:271–295
References
Carey, S., & Xu, F. (2001). Infants’ knowledge of objects: Beyond object files and object tracking. Cognition,
80, 179–213.
Cary, M., & Reder, L. M. (2002). Metacognition in strategy selection: Giving consciousness too much
credit. In M. Izaute, P. Chambres, & P. J. Marescaux (Eds.), Metacognition: Process, function and use
(pp. 63–78). New York: Kluwer.
Conant, R. C., & Ashby, W. R. (1970). Every good regulator of a system must be a model of that system.
International Journal of Systems Science, 1, 89–97.
Dayan, P., & Abbott, I. F. (2001). Computational and mathematical modeling of neural systems. Cambridge:
MIT Press.
de Villiers, J. G. (2000). Language and theory of mind: what are the developmental relationships? In
S. Baron-Cohen, H. Tager-Flusberg, & D. J. Cohen (Eds.), Understanding Other Minds (pp. 83–123).
Oxford: Oxford University Press.
Decety, J., Grezes, J., Costes, N., Perani, D., Jeannerod, M., Procyk, E., Grassi, F., & Fazio, F. (1997).
Brain activity during observation of action. Influence of action content and subject’s strategy. Brain,
120, 1763–1777.
Dienes, Z., & Perner, J. (2002). The metacognitive implications of the implicit-explicit distinction. In
P. Chambres, M. Izaute, & P.-J. Marescaux (Eds.), Metacognition: Process, function and use (pp. 171–
190). Dordrecht, Netherlands: Kluwer Academic Publishers.
Dokic, J., & Egré, P. (in press) Margin for error and the transparency of knowledge.
Dokic, J., & Proust, J. (Eds.) (2003). Simulation and knowledge of action. Amsterdam: John Benjamins.
Evans, G. (1982). The varieties of reference. Oxford: Clarendon Press.
Flavell, J. H. (2004). Development of knowledge about vision, in Levin, D.T. (ed.), Thinking and seeing.
Visual metacognition in adults and children (pp. 13–36). Cambridge, MIT Press.
Foote, A. L., & Crystal, J. D. (2007). Metacognition in the rat. Current Biology, 17, 551–555.
Gennaro, R. (Ed.) (2004). Higher-order theories of consciousness. An anthology. Amsterdam: John
Benjamins.
Gordon, R. M. (1996). ‘Radical’ simulationism. In P. Carruthers, & P. K. Smith (Eds.), Theories of theories
of mind (pp. 11–21). Cambridge: Cambridge University Press.
Hampton, R. R. (2001). Rhesus monkeys know when they remember. Proceedings of the National Academy
of Sciences USA, 98, 5359–5362.
Hampton, R. R. (2005). Can rhesus monkeys discriminate between remembering and forgetting? In
H. S. Terrace & J. Metcalfe (Eds.), The missing link in cognition. Origins of self-reflective consciousness
(pp. 272–295). New York: Oxford University Press.
Inman, A., & Shettleworth, S. J. (1999). Detecting metamemory in nonverbal subjects: A test with pigeons.
Journal of Experimental Psychology: Animal Behavioral Processes, 25, 389–395.
Koriat, A. (1993). How do we know that we know? The accessibility model of the feeling of knowing.
Psychological Review, 100, 609–639.
Kornell, N., Son, L. K., & Terrace, H. S. (2007). Transfer of metacognitive skills and hint seeking in
monkeys. Psychological Science, 18(1), 64–71.
Leslie, A. M., & Thaiss, L. (1992). Domain specificity in conceptual development: Neuropsychological
evidence from autism. Cognition, 43, 225–251.
Levin, D. T. (2004). Thinking and seeing. Visual metacognition in adults and children. Cambridge:
MIT Press.
Mellor, H. (1991). I and now. In Matters of metaphysics (pp. 17–29). Cambridge: Cambridge University
Press.
Nelson, T. O., & Narens, L. (1992). Metamemory: A theoretical framework and new findings. In
T. O. Nelson (Ed.), Metacognition, core readings (pp. 117–130).
Nichols, S., & Stich, S. (2003). Mindreading. NewYork: Oxford University Press.
Perner, J. (1991). Understanding the representational mind. Cambridge: MIT Press.
Perner, J. (1996). Arguments for a simulation-theory mix. In P. Carruthers & P. K. Smith (Eds.), Theories
of theories mind (pp. 90–104). Cambridge: Cambridge University Press.
Perner, J., & Dienes, Z. (2003). Developmental aspects of consciousness: How much theory of mind to you
need to be consciously aware? Consciousness and Cognition, 12(1), 63–82.
Proust, J. (2003a). Action. In B. Smith (Ed.), John Searle (pp. 102–127). Cambridge, Mass.: Cambridge
University Press.
123
Synthese (2007) 159:271–295 295
Proust, J. (2003b). Thinking of oneself as the same. Consciousness and Cognition, 12(4), 495–509.
Proust, J. (2006a). Agency in schizophrenics from a control theory viewpoint. In W. Prinz & N. Sebanz
(Eds.), Disorders of volition (pp. 87–118). Cambridge: MIT Press.
Proust, J. (2006b). Rationality and metacognition in non-human animals. In S. Hurley & M. Nudds (Eds.),
Rational animals? Oxford: Oxford University Press.
Proust, J. (in press a). What is a mental function? In A. Brenner & J. Gayon (Eds.), French philosophy
of science, Boston Studies in the Philosophy of Science.
Proust, J. (in press b). Is there a sense of agency for thought? In L. O’Brien (Ed.), Mental action. Oxford:
Oxford University Press.
Recanati, F. (2000). Oratio Obliqua, Oratio Recta. Oxford: Blackwell.
Rochat, Ph. (2003). Five levels of self-awareness as they unfold early in life. Consciousness and cognition,
12(4), 717–731.
Rosenthal, D. (2000a). Consciousness, content, and metacognitive jujdgments. Consciousness and Cogni-
tion, 9, 203–214.
Rosenthal, D. (2000b). Metacognition and higher-order thoughts. Consciousness and Cognition, 9,
231–242.
Schultz, W. (1998). Predictive reward signal of dopamine neurons. Journal of Neurophysiology, 80, 1–27.
Searle, J. R. (1983). Intentionality. An essay in the philosophy of mind. Cambridge: Cambridge University
Press.
Smith, J. D. (2005). Studies of uncertainty monitoring and metacognition in animals and humans. In
H. S. Terrace & J. Metcalfe (Eds.), The missing link in cognition. Origins of self-reflective consciousness
(pp. 242–271). New York: Oxford University Press.
Smith, J. D., Beran, M. J., Redford, J. S., & Washburn, D. A. (2006). Dissociating uncertainty responses
and reinforcement signals in the comparative study of uncertainty monitoring. Journal of Experimental
Psychology: General, 135(2), 282–297.
Smith, J. D., Schull, J., Strote, J., McGee, K., Egnor, R., & Erb, L. (1995). The uncertain response in the
bottlenosed dolphin. Tursiops truncatus Journal of Experimental Psychology: General, 124, 391–408.
Smith, J. D., Shields, W. E., & Washburn, D. A. (2003). The comparative psychology of uncertainty moni-
toring and metacognition. Behavioral and Brain Sciences, 26(3), 317–373.
Sperber, D. (1997). Intuitive and reflective beliefs. Mind and Language, 12(1), 67–83.
Stich, S. (1978). Beliefs and subdoxastic states. Philosophy of Science, 45, 499–518.
Sugrue, L. P., Corrado, G. S., & Newsome, W. T. (2005). Choosing the greater of two goods: Neural
currencies for valuation and decision making. Nature Reviews Neuroscience, 6, 363–375.
Tager-Flusberg, H. (2000). Language and understanding minds: connections in autism. In S. Baron-Cohen,
H. Tager-Flusberg & D. J. Cohen (Eds.), Understanding Other Minds (pp. 124–149). Oxford: Oxford
University Press.
Williamson, T. (2000). Knowledge and its limits. Oxford: Oxford University Press.
Zaitchik, D. (1990). When representations conflict with reality: The pre-schooler’s problem with false belief
and ‘false’ photographs. Cognition, 35, 41–68.
123