Active Inference As A Theory of Sentient Behavior - ScienceDirect
Active Inference As A Theory of Sentient Behavior - ScienceDirect
Review
Show more
https://doi.org/10.1016/j.biopsycho.2023.108741
Get rights and content
Highlights
Abstract
This review paper offers an overview of the history and future of active inference—a
unifying perspective on action and perception. Active inference is based upon the
idea that sentient behavior depends upon our brains’ implicit use of internal models
to predict, infer, and direct action. Our focus is upon the conceptual roots and
development of this theory of (basic) sentience and does not follow a rigid
chronological narrative. We trace the evolution from Helmholtzian ideas on
unconscious inference, through to a contemporary understanding of action and
perception. In doing so, we touch upon related perspectives, the neural
underpinnings of active inference, and the opportunities for future development. Key
steps in this development include the formulation of predictive coding models and
related theories of neuronal message passing, the use of sequential models for
planning and policy optimization, and the importance of hierarchical (temporally)
deep internal (i.e., generative or world) models. Active inference has been used to
account for aspects of anatomy and neurophysiology, to offer theories of
psychopathology in terms of aberrant precision control, and to unify extant
psychological theories. We anticipate further development in all these areas and note
the exciting early work applying active inference beyond neuroscience. This suggests
a future not just in biology, but in robotics, machine learning, and artificial
intelligence.
Previous Next
Keywords
Active inference; Predictive coding; Generative model
:
1. Introduction
Psychologists and neuroscientists are increasingly entertaining the idea of the brain
as a “prediction machine”, which learns an internal (i.e., generative) model of the
lived world – and of the consequences of its actions – to make sense of sensations,
predict how the current situation will unfold (i.e., learning and perception), and to act
in a purposeful manner (i.e., action selection, exploration-exploitation, planning, et
cetera). This idea appears in several guises, including the Bayesian brain, the predictive
brain, predictive processing, predictive coding, active inference and the free energy
principle, to name a few.
Here, we critically review the origins, scope and impact of this idea, in fields like
psychology and neuroscience. For conceptual clarity, we focus specifically on active
inference: a normative theory of sentient behavior that formalizes the “predictive
brain” idea and provides a first-principle account of its computational and neuronal
processes (Parr et al., 2022).
While active inference is still relatively young, it has a growing impact across various
disciplines. It is increasingly used by (for example) neuroscientists interested in the
neural circuits supporting predictions and prediction errors (Bastos et al., 2012, Parr
et al., 2021, Parr and Friston, 2018, Walsh et al., 2020); psychologists interested in
how we deal with uncertainty and cognitive effort during decision-making (Parr et
al., 2023, Rens et al., 2023), modelers interested in the mechanisms of action-
perception, exploration-exploitation and higher cognition (Friston et al., 2017, Friston
et al., 2017, Pezzulo et al., 2015, Pezzulo et al., 2018), clinicians interested in
understanding aberrant behavior in psychopathology (Maisto et al., 2021, Van den
Bergh et al., 2017), roboticists interested in self-supervised learning of world models
and goal-directed behavior (Ahmadi and Tani, 2019, Taniguchi et al., 2023), and
neurophilosophers (Clark, 2015, Hohwy, 2013).
This breadth of application is appealing, but risks creating a fragmented picture and
some uncertainty about its original commitments and conceptual implications. The
aim of this brief manuscript is to help researchers using (or interested in) predictive
coding and active inference to “connect the dots” and orient themselves within a
growing literature. Despite distinct lines of work — that emphasize different aspects
:
of active inference — these applications all rest on the same core principles. To
foreground these core principles, we will look at the historical and conceptual origins
of active inference—to illustrate how its core principles were introduced; then
consider briefly how the scope of active inference has expanded into several
disciplines—and finally look to future developments. Given the brevity of this
treatment, we cannot provide a full introduction to active inference. Rather, we
provide an overview of the narrative in (Parr et al., 2022), which interested readers
can consult.
In the next section, we briefly discuss the conceptual (and historical) roots of active
inference in early views of prediction and action-based cognition. We then review
some key developments of active inference, by focusing on landmark papers that
explain how it stems from a single principle (namely, free energy minimization). We
next consider its scope across perception, action, planning, etc. This brief review
helps us make the point that active inference provides a unifying perspective on
several cognitive topics and theories and across levels of understanding, from
conceptual to neural. Finally, we briefly highlight some promising research directions
that could expand the scope of active inference – and potentially its impact on
psychology and neuroscience.
Yet another “root” is the idea of cyberneticists (Miller et al., 1960, Powers, 1973,
Wiener, 1948) that goal-directed action proceeds by firstly setting up a desired state
or observation (e.g., feeling warm), then monitoring the discrepancy – now referred
to as a “prediction error” – between the preferred and sensed state (e.g., feeling
excessively warm), and then selecting a course of action that reduces this discrepancy
– where “action” is a suitcase word and can include any means to exert control over
external stimuli; ranging from simple autonomic reflexes (e.g. thermoregulation) to
sophisticated plans (e.g., visiting one’s favorite ice cream shop). A key result in this
field – which coheres with the Helmholtzian perspective above – is the ‘Good
regulator theorem’ of (Conant & Ashby, 1970), which argues that effective regulatory
systems must [be a] model the environment they regulate. In a similar vein, in
psychology, ideomotor theory proposed that action control is essentially anticipatory
and that action are selected and controlled by their anticipated consequences or
outcomes, not through stimulus-response (Hoffmann, 2003, Hommel, 2003, James,
1890).
Besides cybernetics, there are other influential views that highlight the centrality of
adaptive regulation for behavior and life itself. One example is the idea that living
organisms are autopoietic systems, which create the conditions for their own
existence. More recently, this idea has been framed as ‘self-evidencing’ (Hohwy 2016)
– i.e., creatures seek out sensations that provide evidence for their continued
existence. Intuitively, sensing our body temperature to be around 37 °C offers more
evidence that we are still alive than body temperatures far from this value. The
concept of autopoiesis gave birth to enactive approaches in philosophy (Maturana &
Varela, 1980). From another angle, it has been postulated that a central imperative for
:
living organisms is maintenance of physiological homeostasis (i.e., correction of
deviations from preferred physiological states through reflexive actions) and the
regulation of basic imperatives (Cannon, 1929) – but more modern theories
emphasize that physiological regulation is fundamentally anticipatory (i.e., allostatic)
(Sterling, 2012). Various researchers have proposed that closed-loop adaptive
regulation (and not stimulus-response) is key to understanding not just physiology
but (potentially) all cognitive processing (Cisek, 1999, Pezzulo and Cisek, 2016).
Finally, another root is the idea that cognitive processes, such as learning, perception
and decision-making, require an active engagement of organisms with the
environment. One early example of this action-oriented perspective is the view of
Gibson that perceiving things consists in seeing what to do or not to do with them,
i.e., perceiving affordances (Gibson, 1979). More recently, various researchers
proposed the necessity of a “pragmatic turn” in cognitive science and neuroscience –
and the need to recognize the importance of action as part and parcel of our
cognition (Buzsaki, 2019, Cisek and Kalaska, 2010, Cisek and Pastor-Bernier, 2014,
Engel et al., 2016, Lepora and Pezzulo, 2015, O’Regan and Noe, 2001), rather than just
a way to report “central” decisions, as assumed in conventional (serial) theories.
Interestingly, each of these ideas implies a shift from reactive to predictive, enactive
views of the brain. While a reactive brain waits for incoming stimuli, a predictive and
active brain predicts external events (e.g., predictive coding) and actively gathers
evidence (i.e., active sensing and active learning) to make sense of the world. While a
reactive brain selects actions based on the past and present (e.g., the history of
reinforcement and the current cue), a predictive brain actively imagines its preferred
future and then makes this happen by acting (e.g., acts in a goal-directed manner).
While a reactive brain maintains homeostasis, a predictive brain acts to anticipate
needs and performs anticipatory regulatory (or allostatic) actions.
All these (and other) views contributed to raising the importance of predictive and
enactive views of the brain and of cognition. However, each of these perspectives
were somewhat disconnected from one another and linked to different research
traditions, which are sometimes seen as conflicting with one another (e.g., the
Helmholtzian and the Gibsonian traditions). One benefit of active inference is that it
:
helps unify and thereby advance these traditions, as we explain in the following
Sections.
The mathematics of variational free energy minimization is beyond the scope of this
article; we suggest to the interested readers to consult (Parr et al., 2022). Here,
instead, we introduce the key concepts of the theory, by briefly reviewing (non-
chronologically) selected landmark papers and linking them to the early theories.
Active inference starts from a simple consideration: that to maintain their existence
and integrity, all living organisms need to remain in a bounded set of characteristic
states that basically define their place within an ecological niche; for example, a fish
cannot live out of water. Using the lexicon of Bayesian inference, being out of water
for a fish is a “surprising” state. Clearly a fish should avoid this surprise, and the idea
generalizes to suggest that living organisms must avoid surprising states (Friston et
al., 2010). If they did not, they would not be living organisms for long. Another way of
looking at this is that everything (including me) is defined by being in some
characteristic (attracting) set of states. Conversely, I am defined by the kinds of states
I cannot be in. These are surprising states.
Fig. 1. Generative model and generative process in active inference. This Figure—
reproduced from (Parr et al., 2022)—illustrates the structure of active inferential
theories of brain function. Our worlds evolve according to some dynamical process
that generates observations (y) from hidden states (x * ). Our internal models account
for observations in terms of hypothetical hidden states (x). Our inferences about these
states based upon our observations then drive actions (u) that intervene on the
processes generating our sensations.
Put simply, an organism can minimize variational free energy by aligning the
predictions of its generative model and the data it observes. In different settings, this
minimization has been described in various ways, such as the minimization of
surprise, of prediction errors, or of the discrepancy between the model and the world.
All of these are equivalent to the minimization of variational free energy under
specific sets of assumptions.
Interestingly, aligning the predictions derived from a generative model and data can
:
be achieved in two ways: by changing the model predictions and by changing the
observed data. The former corresponds to revising the agent’s beliefs (used in the
technical sense of probability distributions over hidden variables) if they do not
explain the data well. This is exactly the inferential view of perception of (Helmholtz,
1866). The latter corresponds to acting in the world to change the data that will be
sampled next – to render them more like the organism’s prior predictions. This latter
perspective on action – and on its dependence on expected outcomes – is highly
congruent with cybernetics (Miller et al., 1960, Powers, 1973, Wiener, 1948) and
ideomotor theory (Hoffmann, 2003, Hommel, 2003, James, 1890).
In sum, changing beliefs about the causes of data (i.e., perception) and changing the
data (i.e., action) are two aspects of free energy minimization. In formal terms, they
map to its two components: the minimization of divergence and the maximization of
evidence, see Fig. 2. Recognizing that action and perception can be unified within a
single formal imperative – the minimization of free energy – is one of the key
innovations of active inference, which helps integrate and extend the early theories
reviewed above.
Subsequently, this theory was extended to also model autonomic control (Barrett and
Simmons, 2015, Pezzulo, 2014, Seth et al., 2012). The general idea is that autonomic
control might work by generating interoceptive predictions (i.e., homeostatic
setpoints) and then fulfilling them through autonomic reflexes, in much the same
way motor control might work by generating proprioceptive predictions and then
fulfilling them through motor reflexes. This development of active inference helps
connect it with theories of allostatic control (Sterling, 2012) and paves the way to a
better understanding of our ability to model and control the internal milieu, not just
the external environment. This stream of research underwrote novel approaches to
psychopathology – as deficits of interoceptive processing (Paulus et al., 2019).
So far, we have discussed active inference using generative models that characterize
processes that unfold in continuous time (e.g., predictive coding networks) and use
continuous variables (i.e., the formal framework of dynamical systems and state-
space models). However, many cognitive problems can be characterized at a distinct
level: as (sequences of) discrete decisions. These include problems that require the
selection of discrete responses during psychology experiments, the targets for
saccades, or navigational trajectories in discretized environments (Friston et al., 2017,
Friston et al., 2017). These problems can be modeled in active inference, using
generative models that use discrete variables (and the formal framework of Partially
Observable Markov Decision Processes).
:
In addition to the two aforementioned components (priors and likelihood function),
the generative models for active inference in discrete time often include a third
component: the transition function, which describes the way in which hidden states
change depending upon the agent’s actions (or sequences of actions, called policies).
Crucially, these generative models have temporal depth and afford a novel capability
that was not available in simpler models: namely, planning. In simple terms, planning
involves using the generative model to predict the consequences of different policies,
scoring the policies according to how much they are expected to minimize free energy
in the future and then (with some simplifications) select the best policy.
This planning process induces a novel quantity – expected free energy – that is the
functional that active inference uses to evaluate (and assign a prior to) policies and it
is distinct from the notion of variational free energy discussed so far (Friston et al.,
2017). The notion of expected free energy has been very useful in the development of
active inference models of things like (bounded) decision-making, planning,
exploration-exploitation and curiosity (Friston et al., 2017, Parr and Pezzulo, 2021,
Schwartenbeck et al., 2019). This is because this notion is richer than the common
optimization objectives used in other formal frameworks (e.g., economic theory and
reinforcement learning). This is because expected free energy considers jointly a
pragmatic imperative (utility maximization) and an epistemic imperative (information
gain, or the resolution of the uncertainty). Indeed, as Fig. 4 illustrates, it is possible to
map expected free energy to various other formal notions (e.g., Bayesian surprise,
Risk-sensitive control, Expected utility theory), by removing one or more of its terms.
:
Download: Download high-res image (119KB)
Download: Download full-size image
Fig. 4. Expected free energy and the way it can be mapped to different formal
notions (e.g., Bayesian surprise, Risk-sensitive control, Expected utility theory) by
removing one or more terms, denoted with numbers. This Figure—reproduced
from (Parr et al., 2022)—expresses expected free energy in terms of beliefs about
trajectories (indicated by the tilde ∼). The additional symbols here, not in previous
figures, are the π for policies and the C for preferences. Note that some terms
(including term 1) are expressed in terms of expectations—i.e., averages under the
subscripted distribution.
Active inference is a general scheme that can be applied to address various cognitive
processes. Crucially, the functioning of active inference is the same across all
problems: what differs is the generative model, which is task specific. This implies
that by designing the appropriate generative models, it is possible to address a variety
of cognitive tasks with the same approach – and to pass from the normative
perspective of active inference to specific implementations that have biological
plausibility (Friston et al., 2017, Parr and Friston, 2018).
Here, a worked example may be helpful. To illustrate some of the principles we have
outlined so far, we will consider how we might go about developing a model for a
ubiquitous task in cognitive neuroscience—a delay period oculomotor task. This is a
relatively simple task that can be performed by humans—and some animals—and that
is designed to probe working memory function (Funahashi et al., 1989). The task
sequence is as follows. First, a cross is presented on screen and our subject maintains
fixation on this cross. A target then appears at one of several possible locations
towards the periphery of the screen, but our subject still maintains fixation. The
target then disappears and, after a ‘delay period’, a stimulus appears to signify that
the subject should make a saccadic eye movement to the location of the target.
Successful performance of this task relies upon retaining a memory of the target
location during the delay and response phases.
To model this task, we must consider the data available to the subject. In this case,
:
these are the visual stimuli and proprioceptive inputs, and whether the correct action
was chosen. To do so, we need to take account of the causes of these data. The causes
of proprioceptive data are simply the direction in which our subject’s eyes are
pointing. Visual outcomes, depend upon a combination of (1) gaze direction, (2) the
intended target location, and (3) the current stage of the task (i.e., the fixation, target
presentation, delay, or response stage). For each of these three variables, we must
then specify how we expect them to evolve throughout the task. The gaze direction
will transition from one step to the next based upon the decisions our subject makes.
The intended target location will be fixed (although initially unknown) throughout
the task. The task stage evolves predictably through a sequence of steps. Together,
these beliefs about the way in which data are generated and the dynamics of the
causes allow our subject to predict what will be observed next, and to update these
beliefs when these predictions are violated.
As outlined above, active inference equips models with prior beliefs about the relative
plausibility of different choices based upon their relative expected free energies. In
this model, the key part of the expected free energy is a preference for receiving the
‘correct’ feedback outcome which is only available during the response phase of the
task (see (Mirza et al., 2016) for a similar setup in the context of scene categorisation,
in which the main role of the expected free energy is to promote information
seeking). It is this that prompts a saccade to the remembered target location. Finally,
the predicted action must be executed. This depends upon resolving the error
between the anticipated proprioceptive information given the inferred saccade and
current proprioceptive input. The result is the sequence of steps shown in Fig. 5.
:
Download: Download high-res image (146KB)
Download: Download full-size image
Fig. 5. A simulated oculomotor delay period task. This figure, taken from (Parr &
Friston, 2019b) (published under a CC BY 4.0 license), shows simulated performance
of a simple working memory task under active inference. Although simple, this task
calls for planning (of our next saccade), recall (of the target location), and movement
execution. The upper left images show a series of frames taken from the simulation,
as if we were observing our participant’s eyes. The black arrows link these
behavioural responses to the view of the stimulus screen from the time of the target
(red) presentation to the response phase. A series of black dots show the (cumulative)
:
trajectory of gaze direction. Because this model is formulated to have both
continuous (prediction-error minimising) and discrete (sequential planning) parts,
we can plot the trajectory both in terms of position and velocity (lower left) and in
terms of the sequence of actions taken.
The oculomotor control example illustrates how active inference can be concretely
applied to study cognitive tasks, by designing (or learning) the appropriate generative
models. Generative models represent formal hypotheses about how cognitive tasks
are accomplished – hypotheses that can be validated with empirical data. A useful
illustration of the design principles to realize (or train) generative models for
different cognitive problems is provided in (Parr et al., 2022). This treatment makes a
distinction between generative models in continuous time (that are useful to address
motor control tasks) and discrete time (that are useful to address decision and
planning tasks) and explains how these two types can be combined to form so-called
hybrid or mixed generative models, in which discrete-time models are placed on top
of continuous-time models. Furthermore, the generative models of active inference
can be extended hierarchically, to model processes that unfold at different timescales.
One example is the model of active listening processes, in which (for example) lower
hierarchical levels deal with words and higher levels deal with sentences (Friston et
al., 2021). Another example is a model of hierarchical action recognition that
recognizes actions at different levels, from low level kinematics to higher level goals
and intentions (Proietti et al., 2023). It is also possible to use hierarchical models to
model hierarchies of control, in which lower-to-higher levels deal with autonomic
imperatives (e.g., ensure a correct basic temperature) in increasingly complex ways
(e.g., from thermoregulation to the goal-directed plan to buy water before a long run)
(Pezzulo et al., 2015, Tschantz et al., 2021). These developments – from simple to
more sophisticated (e.g., hierarchically and temporally deep) generative models has
extended the range of cognitive models that have been addressed using active
inference over the years.
Another important development of active inference regards precision control and its
role in psychopathologies. In predictive coding, variables are encoded as Gaussian
distributions and precision simply refers to the inverse of the variance of a
distribution (Friston, 2005). Precision control refers to a mechanism that optimizes
the precision of (the distribution of) each variable of an organism’s generative model.
It is important since it regulates the relative importance of top-down predictions and
bottom-up prediction errors across the hierarchy. This is because prediction errors
that are assigned greater (lower) precision have greater (lower) impact on the belief
updating and the ensuing inference. Veridical inference requires the precision of (the
distribution of) each variable to be optimized, to reflect the signal-to-noise ratio of
sensory signals – therefore highlighting a link between precision control and
attention as gain control (Parr & Friston, 2019a) – or the importance of an organism’s
prior preferences – reflecting the fact that an organism’s innate drives or goal states
can be encoded as highly precise priors (Pezzulo et al., 2015). Interestingly, when
precision control fails, it can produce excessively rigid forms of inference (when
priors fail to be updated in the light of novel evidence) or excessive sensitivity to
stimuli (when belief revision follows the sensory input and its random fluctuations
too closely—i.e., it overfits). These forms of aberrant inference, which depend
sensitively on predicted precision, have been adopted to explain several
psychopathological conditions, such as delusions, depression, psychosis, and many
others (Barrett et al., 2016, Corlett and Fletcher, 2015, Edwards et al., 2012). In turn,
these theories also speak to aberrant neuromodulation, since the precision of (the
distribution of) different variables might be encoded by different neuromodulators,
e.g., acetylcholine for the precision of the likelihood, noradrenaline for the precision
of transitions, dopamine for the precision of policies, etc. (Parr & Friston, 2018).
Yet another development regards the analysis of generative models during sleep or
other ‘offline’ periods. It has long been hypothesized that learning generative models
benefits from alternating on-line and off-line periods (Hinton et al., 1995). While on-
:
line generative modelling maximises accuracy (under complexity constraints), during
off-line activity – in the absence of sensory data to “explain away” – model
optimisation can focus on minimising complexity; for example, by removing
redundant parameters (Friston et al., 2017, Pezzulo et al., 2021). From a neuronal
perspective, generative modelling during offline periods could be associated with
(generative) replay activity in the hippocampus, the prefrontal cortex and other brain
areas; but these links remain to be fully established (Foster, 2017, Schwartenbeck et
al., 2023, Stoianov et al., 2022).
Active inference proceeds the other way around: it starts from a single principle and
asks how far one can go with it. And to what extent it is possible to derive from that
principle empirically testable hypotheses about behavior and its cognitive and neural
mechanisms? This approach brings the benefits of unification, in at least six ways.
First, active inference assumes that everything, from perception to action selection
and learning ultimately serves to minimize variational free energy. A consequence of
this is that one can align the (sometimes vague) conceptual terms used in psychology
with crisp formal terms of free energy minimization. For example, one can assign
things like attention to precision control. At the neuronal level, the fast updates –
mediated by synaptic activity – might correspond to inferential processes that
minimize free energy at a fast time scale, whereas the slower updates – at the level of
synaptic efficacy – might correspond to learning processes that minimize free energy
at a slower timescale. Precision dynamics might correspond to the activity of
neuromodulators, which finesse the inference at multiple levels, for example, by
:
increasing the post-synaptic gain of sensory or prediction error-units (Feldman &
Friston, 2010). Oscillatory dynamics that are ubiquitous (and that often occur in
synchrony) both within and across brain area might be signatures of temporal
prediction and of the exchange of top-down and bottom-up information across
hierarchical levels of the brain’s generative model (Arnal & Giraud, 2012).
Third, active inference has the potential to unify different “levels of understanding” of
cognitive processes. Marr famously introduced a distinction between computational,
algorithmic and neural implementation levels and argued that progress can be made
within each level and by connecting different levels (Marr, 1982). Establishing links
between theories that operate at different levels is often challenging. Active inference
helps establish firm relations across levels of description. Rather than Marr’s
tripartite distinction, in active inference it is more common to appeal to a distinction
between normative theory and process theory (Friston et al., 2017). Free energy
minimization is the normative objective of living organisms, whereas predictive
coding and variational message passing are process-level theories that describe how
:
the brain might support free energy minimization. Importantly, as shown by (Friston,
2005), under certain assumptions predictive coding can be directly derived by the
minimization of variational free energy, connecting the two levels of explanation. A
similar case can be made for the variational message passing schemes proposed to
support discrete active inference in neural circuits (Friston et al., 2017).
Fourth, unification endows existing constructs with validity, via the application of
active inference across domains. One example is the development of theories of
interoceptive inference and autonomic control (Barrett and Simmons, 2015, Pezzulo,
2014, Seth et al., 2012) by analogy with the functioning of action control (Adams et
al., 2013). In this perspective, autonomic control works exactly like action control –
namely, it aims to minimize a discrepancy between a predicted and a sensed signal –
except that the “signal” refers to interoceptive streams rather than proprioceptive
streams. Another example can be found in computational psychiatry, where
numerous accounts of psychopathology appeal to a single mechanism: namely,
aberrant precision control.
Fifth, active inference has the potential to reconcile (or at least to contextualize)
theoretical perspectives that have long been considered at odds in psychology,
neuroscience and philosophy. One example is the Helmholtzian view that perception
constitutes an inference about the entities of the external world that cause our
sensations (Helmholtz, 1866) and the Gibsonian view that perceiving consists in
seeing action opportunities and affordances, not reconstructing a model of the
external reality within the brain (Gibson, 1979). This apparent dialectic could be
dissolved by considering that there are multiple ways to design generative models;
specifically, a relevant distinction is between generative models that explicitly model
the ways external states produce sensations (a.k.a., environmental models) or the
ways actions produce sensations (a.k.a., sensorimotor models) (Sims and Pezzulo,
2021, Pezzulo et al., 2023). Some active inference studies use generative models that
include explicit beliefs about entities in the external world that cause sensations,
such as one’s location in space (Friston et al., 2017). Other active inference studies use
generative models that only consider the sensory consequences of one’s action, such
as touch sensations that follow whisking at a given amplitude, but not explicit beliefs
about objects ‘out there’ (Mannella et al., 2021). The latter generative models adhere
:
more closely to the notions of affordance (Gibson, 1979) and of sensorimotor
contingency (O’Regan & Noe, 2001), despite the fact they still entail inferential
dynamics. Besides this specific topic, there is a vivid debate in philosophy that
concerns the most appropriate way to consider active inference, in relation to
internalist (Hohwy, 2013), externalist (Clark, 2013) or enactivist theories (Bruineberg
et al., 2018).
The compass of active inference is expanding rapidly, but the landscape of future
:
opportunities may be even ampler. Here, we focus on some of the developments that
we consider most promising and most likely in the near future.
The first and perhaps most obvious direction for the future regards a deeper empirical
scrutiny of active inference. A question that is sometimes asked of active inference is
whether any empirical findings could offer evidence for or against the framework.
This can be a vexed question to answer as it constitutes a category error. A framework
is not in itself a hypothesis. It is a way of formulating hypotheses. The relationship
between active inference and empirical psychology is that we can formalize
psychological theories in terms of the generative models that underwrite
neurophysiological and behavioural responses. Equipped with a proposed model, the
framework can be used to express a hypothesis, to predict the behaviour expected
under that hypothesis, and to fit to measured data to formally compare alternative
hypotheses. In other words, while active inference is an application of the free energy
principle – which is a principle (i.e., method) rather than a theory (Friston, 2010) –
theories tested under the active inference framework (e.g., those considered in this
article) make specific empirical predictions that can (and need to) be empirically
validated. One example of this is the oculomotor delay period model shown in Fig. 5,
which generate empirically testable predictions about oculomotor performance as a
function of varying delay periods (Parr & Friston, 2019b). Various empirical studies
are already addressing the empirical predictions of predictive coding, such as how
top-down and bottom-up dynamics support predictions and prediction errors,
respectively (Walsh et al., 2020). However, active inference makes a number of
specific predictions about (for example) the way the motor system works (Shipp et
al., 2013) and the way higher cognitive functions are implemented (Pezzulo et al.,
2018) that differ from mainstream theories and could be increasingly scrutinized by
future studies.
A second interesting direction for the future is assessing to what extent active
inference – and more broadly, the free energy principle – can help us understand the
evolution of complex neural circuits and life forms from simpler ones. Active
inference suggests a possible path from the simple mechanisms that supported
prediction and control in our earlier evolutionary ancestors to the more sophisticated
abilities of our species (Pezzulo et al., 2022), but a comprehensive account of the
:
evolution and “phylogenetic refinement” (Cisek, 2019) of living organisms remains to
be fully developed (Friston et al., 2023, Friston et al., 2023).
A third interesting direction for the future regards the realization of advanced
artefacts, such as AIs and robots, based on active inference. There have already been
several successful robotic implementations of active inference, but the full potential
of the theory has not yet been reached (Ahmadi and Tani, 2019, Lanillos et al., 2021,
Priorelli, Pezzulo, & Stoianov, 2023, Taniguchi et al., 2023). Interestingly, some of the
central concepts of active inference, such as the importance of generative models and
self-supervised, predictive learning, are becoming central in mainstream research in
AI, as testified by the recent successes in generative AIs such as large language
models. This creates an important opportunity, since (apart for their obvious
technological impact), state-of-the-art AI systems can be precious in advancing our
understanding of living organisms, providing that they incorporate appropriate
(design) principles (Pezzulo et al., 2023).
Box 1
Glossary of technical terms.
Expected Free Energy: This is the quantity that is used in active inference to
score action sequences or policies (and then to select between them). It takes
into consideration both the pragmatic value of policies – or how close a
policy’s expected outcomes are to the preferred outcomes – and their
epistemic value (or information gain) – or how much the policy is expected
to reduce uncertainty.
The authors did not use generative AI technologies for preparation of this work.
Data availability
No data was used for the research described in the article.
References
Adams et al., 2013 R.A. Adams, S. Shipp, K.J. Friston
Predictions not commands: active inference in the motor system
Brain Structure & Function, 218 (3) (2013), pp. 611-643, 10.1007/s00429-012-0475-5
View in Scopus Google Scholar
Bastos et al., 2012 A.M. Bastos, W.M. Usrey, R.A. Adams, G.R. Mangun, P. Fries, K.J. Friston
Canonical microcircuits for predictive coding
Neuron, 76 (4) (2012), pp. 695-711
View PDF View article View in Scopus Google Scholar
Clark, 2015 Clark, A. (2015). Surfing Uncertainty: Prediction, Action, and the Embodied
Mind. Oxford University Press, Incorporated.
Google Scholar
Clark, 2023 Clark, A. (2023). The experience machine: How our minds predict and shape
reality.
Google Scholar
Edwards et al., 2012 M.J. Edwards, R.A. Adams, H. Brown, I. Pareés, K.J. Friston
A Bayesian account of “hysteria
Brain: A Journal of Neurology, 135 (Pt 11) (2012), pp. 3495-3512, 10.1093/brain/aws129
View in Scopus Google Scholar
Friston et al., 2017 K. Friston, M. Lin, C.D. Frith, G. Pezzulo, J.A. Hobson, S. Ondobaka
:
Active Inference, Curiosity and Insight
Neural Computation, 29 (10) (2017), pp. 2633-2683, 10.1162/neco_a_00999
View in Scopus Google Scholar
Friston et al., 2021 K. Friston, N. Sajid, D.R. Quiroga-Martinez, T. Parr, C.J. Price, E. Holmes
Active listening
Hearing Research, 399 (2021), Article 107998, 10.1016/j.heares.2020.107998
View PDF View article View in Scopus Google Scholar
Friston et al., 2023 Friston, K., Friedman, D.A., Constant, A., Knight, V.B., Parr, T., &
Campbell, J.O. (2023). A variational synthesis of evolutionary and developmental
dynamics. https://doi.org/10.3390/e25070964 .
Google Scholar
Friston et al., 2023 K. Friston, L. Da Costa, D.A.R. Sakthivadivel, C. Heins, G.A. Pavliotis, M.
Ramstead, T. Parr
Path integrals, particular kinds, and strange things (arXiv:2210.12761)
arXiv (2023), 10.48550/arXiv.2210.12761
Google Scholar
Hinton et al., 1995 G.E. Hinton, P. Dayan, B.J. Frey, R.M. Neal
The “wake-sleep” algorithm for unsupervised neural networks
Science, 268 (5214) (1995), pp. 1158-1161
Crossref View in Scopus Google Scholar
Lanillos et al., 2021 P. Lanillos, C. Meo, C. Pezzato, A.A. Meera, M. Baioumy, W. Ohata, A.
Tschantz, B. Millidge, M. Wisse, C.L. Buckley, J. Tani
Active Inference in Robotics and Artificial Agents: Survey and Challenges
(arXiv:2112.01871)
arXiv (2021), 10.48550/arXiv.2112.01871
Google Scholar
Maselli et al., 2023 Maselli, A., Gordon, J.R., Eluchans, M., Lancia, G.L., Thiery, T.,
Moretti, R., Cisek, P., & Pezzulo, G. (2023). Beyond simple laboratory studies:
Developing sophisticated models to study rich behavior.
Google Scholar
Mirza et al., 2016 M.B. Mirza, R.A. Adams, C.D. Mathys, K.J. Friston
Scene Construction, Visual Foraging, and Active Inference
Frontiers in Computational Neuroscience, 10 (2016), p. 56, 10.3389/fncom.2016.00056
Google Scholar
Pezzulo et al., 2023 Pezzulo, G., Parr, T., Cisek, P., Clark, A., & Friston, K. (2023).
Generating Meaning: Active Inference and the Scope and Limits of Passive AI.
Google Scholar
Pezzulo et al., 2023 Pezzulo, G., D'Amato, L., Mannella, F., Priorelli, M., Van de Maele, T.,
Stoianov, I. P., & Friston, K. (2023). Neural representation in active inference:
using generative models to interact with--and understand--the lived world.
arXiv preprint arXiv:2310.14810.
Google Scholar
Tschantz et al., 2021 Tschantz, A., Barca, L., Maisto, D., Buckley, C.L., Seth, A.K., &
Pezzulo, G. (2021). Simulating homeostatic, allostatic and goal-directed forms
of interoceptive control using Active Inference. bioRxiv, 2021.02.16.431365.
https://doi.org/10.1101/2021.02.16.431365 .
Google Scholar
Van den Bergh et al., 2017 O. Van den Bergh, M. Witthöft, S. Petersen, R.J. Brown
Symptoms and the body: Taking the inferential leap
Neuroscience and Biobehavioral Reviews, 74 (Pt A) (2017), pp. 185-203,
10.1016/j.neubiorev.2017.01.015
View PDF View article View in Scopus Google Scholar
Walsh et al., 2020 K.S. Walsh, D.P. McGovern, A. Clark, R.G. O’Connell
Evaluating the neurophysiological evidence for predictive processing
as a model of perception
Annals of the New York Academy of Sciences, 1464 (1) (2020), pp. 242-268,
10.1111/nyas.14321
View in Scopus Google Scholar
Cited by (5)
Show abstract
Show abstract
1 Technically, surprise here refers to self-information (a.k.a., surprisal); namely, the implausibility of
some (sensory) outcome under a (generative) model of how that outcome was generated.