Dynamic Systems and Causal Structures in Psychology, Connecting Data and Theory 2020
Dynamic Systems and Causal Structures in Psychology, Connecting Data and Theory 2020
Structures in Psychology
Connecting Data and Theory
Proefschrift
door
Oisı́n Ryan
The studies in this thesis were funded by the Netherlands Organization for Sci-
entific Research (Onderzoekstalent Grant 406-15-128)
Beoordelingscommissie:
Prof. dr. E. Ceulemans Katholieke Universiteit Leuven
Prof. dr. P. de Jonge Rijksuniversiteit Groningen
Prof. dr. I. Klugkist Universiteit Utrecht
Prof. dr. H. L. J. van der Maas Universiteit van Amsterdam
Prof. dr. M. C. Völkle Humboldt Universität zu Berlin
ISBN: 978-94-6416-145-8
Cover design: Luke Keeshan
Print: Ridderprint | www.ridderprint.nl
© Oisı́n Ryan 2020. All rights reserved.
Contents
1 Introduction 1
1.1 Current Methodological Practice and Problems . . . . . . . . . . . 2
1.2 Alternative Methodological Frameworks . . . . . . . . . . . . . . . 3
1.3 Outline and Summary . . . . . . . . . . . . . . . . . . . . . . . . . 4
Appendices
2.A Moral-Equivalent DAGs: Violations of Sufficiency and Faithfulness 34
2.B The SE-set Algorithm: A Tool to Aid Causal Hypothesis Generation 35
2.C Empirical Illustration Details . . . . . . . . . . . . . . . . . . . . . 40
Appendices
3.A Matrix Exponential . . . . . . . . . . . . . . . . . . . . . . . . . . . 66
3.B Empirical Example Data Analysis . . . . . . . . . . . . . . . . . . . 67
i
CONTENTS
Appendices
4.A Centrality Measures as Summaries of Path-specific Effects . . . . . 101
4.B Centrality Values DT Stress-Discomfort System . . . . . . . . . . . 103
4.C The Matrix Exponential as Path-Tracing . . . . . . . . . . . . . . . 103
4.D Path-Tracing in CT models . . . . . . . . . . . . . . . . . . . . . . . 105
4.E Interventions and Path-Tracing in CT models . . . . . . . . . . . . 107
Appendices
5.A Determining Fixed Points . . . . . . . . . . . . . . . . . . . . . . . . 161
5.B Mean-Switching Hidden Markov Model . . . . . . . . . . . . . . . 163
5.C Data Generated from Estimated Models . . . . . . . . . . . . . . . 164
5.D Residual Partial Correlations TVAR(1) . . . . . . . . . . . . . . . . 166
5.E Differential Equation Model Building . . . . . . . . . . . . . . . . . 166
5.F Additional Results ESM Time Series . . . . . . . . . . . . . . . . . . 170
Appendices
6.A Simulated Data from the Panic Model . . . . . . . . . . . . . . . . . 205
6.B Additional Details: The Panic Model and Statistical Dependencies 205
6.C Details Empirical vs Simulated Ising Model . . . . . . . . . . . . . 207
References 213
Acknowledgements 245
ii
Chapter 1
Introduction
Psychological phenomena are best understood as dynamic processes: Political be-
liefs become more or less conservative as we age, the mathematics abilities of
children develop over the school year, and the duration and frequency of extreme
moods from day-to-day and week-to-week distinguish healthy from unhealthy
emotion regulation. As such, the key to understanding psychological phenomena
lies in understanding how behaviors, cognition, perceptions, emotions, disposi-
tions, abilities, and all other relevant facets of the mental world evolve, vary and
interact with each other over time, within an individual.
This process perspective has witnessed a tremendous growth in popularity,
bordering on consensus, in the psychological science literature in the past two
decades (Boker, 2002; Van Der Maas et al., 2006; Hamaker, 2012; Molenaar,
2004). In clinical psychology and psychiatry this idea has been particularly trans-
formative, with the traditionally static disease-based conceptualization of mental
disorder supplanted in recent years by the view that psychopathologies are in-
herently complex, multi-dimensional, and dynamic entities (Borsboom, Cramer,
Schmittmann, Epskamp, & Waldorp, 2011; Borsboom, 2017; Kendler, 2019; Nel-
son, McGorry, Wichers, Wigman, & Hartmann, 2017; Wichers, 2014). From this
perspective, to understand and eventually control and treat mental disorders we
must uncover the mechanistic relationships between psychological processes that
underlie psychopathology.
Researchers who subscribe to this theoretical perspective have collected a va-
riety of different types of empirical data with which they hope to study psycho-
logical processes. Due to the subject matter of clinical psychology, much of this
empirical data comes from non-experimental studies. Two popular categories of
non-experimental data can be distinguished. The first of these is cross-sectional
data, consisting of measurements of psychological processes across many indi-
viduals at a single point in time. The second is intensive longitudinal data, con-
sisting of repeated measurements of psychological processes over time, for one
or more individuals, typically in natural everyday settings. This latter type of
data has witnessed a surge in popularity in recent years, in part due to the ad-
vent of smartphone technology, and in part motivated by concerns over the diffi-
culty of making inferences from between-person data to within-person processes
(Molenaar, 2008; Hamaker & Wichers, 2017).
The core goal of this dissertation is to investigate how researchers can best
use non-experimental data on psychological processes to investigate the dynamic
mechanisms that give rise to psychopathology. Doing so will depend critically on
the methodology and more broadly the methodological framework that is used to
analyze these data.
1
1. Introduction
2
1.2. Alternative Methodological Frameworks
current practice, that is, the difficulty of using statistical models based on obser-
vational data to infer causal relationships. Ideally, we would study the aspirin-
headache relationship with an experiment. If we randomly assign headache-
suffering participants to take either an aspirin or a placebo, and we observe that
aspirin takers have lower headache levels two hours later, we can be reasonably
certain that this reduction was caused by the aspirin. However, if we are unable
to conduct such an experiment, we must instead rely on observations of aspirin
intake and headache levels in everyday life. Suppose that the analysis of this data
returns a negative statistical dependence between aspirin and headaches. When
(if ever) can we draw the same causal conclusion from this information as we
would have based on the experimental study? Unfortunately, the traditional sta-
tistical modeling literature provides few if any reasonable answers to this ques-
tion. Intuitive approaches which equate the fit of statistical models with their
causal veracity have variously been described as “radically suboptimal”, yield-
ing “garbage-can” models, and the very intuition behind this notion has been
labeled an urban-myth (Spirtes et al., 2000; Achen, 2005; Spector & Brannick,
2011; Rohrer, 2018). Without a principled way to link statistical with causal in-
formation, the utility of standard approaches for uncovering causal mechanisms
is left on somewhat uncertain footing.
3
1. Introduction
from observational data. In these fields, the different approaches which have
been developed to tackle this issue can broadly be described as the intervention-
ist causal inference framework, so called due to the definition of causal effects
in terms of hypothetical experimental interventions (Rubin, 1974; Greenland &
Robins, 1986; Angrist, Imbens, & Rubin, 1996; Pearl & Verma, 1991; Pearl, 2009).
This framework provides a mathematical language to describe causal structure,
formalizing the notion that statistical relationships (which describe patterns in
data) exist on a different level of explanation than causal relationships (which
describe the processes responsible for producing data). This in turn allows us
to answer the question of how and when statistical dependencies can be used to
infer causal structure, but moreover provides a new way to approach statistical
analyses when causal relationships are the target of inference.
Although these two approaches are distinct from one another, and indeed
tackle distinct shortcomings present in current practice, both the dynamical sys-
tems and interventionist causal inference frameworks offer promising avenues
by which we might improve our understanding of the mechanisms underlying
mental disorder. The chapters in my dissertation form an exploration of what we
can learn from these different approaches, and how we can use these lessons to
improve current practice.
4
1.3. Outline and Summary
5
1. Introduction
struction. The dissertation ends with a clear message for the fields of clinical
psychology, psychiatry and the methodologists who work within these fields: If
we hope to understand the dynamic processes that give rise to psychopathology,
a radical reorientation of current research practice towards formal theory devel-
opment is urgently needed.
6
Chapter 2
This chapter has been adapted from: Ryan, O., Bringmann, L. F. & Schuurman, N. K. (un-
der review). The Challenge of Generating Causal Hypotheses Using Network Models. Pre-print:
https://psyarxiv.com/ryg69/. Author contributions: OR conceptualized the initial project, wrote
the paper and R code and ran the analyses. LFB and NKS helped further develop the ideas in the
project, discussed progress and provided textual feedback.
7
2. The Challenge of Generating Causal Hypotheses
2.1 Introduction
The network approach to psychopathology is a theoretical framework in which
mental disorders are viewed as arising from direct causal interactions between
symptoms (Borsboom & Cramer, 2013; Borsboom, 2017). In practice, researchers
often aim to uncover aspects of the underlying network structure by estimating
network (i.e. graphical) models, generally from cross-sectional data (Van Borkulo
et al., 2014; Epskamp, Waldorp, et al., 2018; Epskamp, Borsboom, & Fried, 2018).
In these instances, researchers typically estimate a Pairwise Markov Random Field
(PMRF); for normally distributed variables this takes the form of a Gaussian
Graphical Model (GGM). These are network models with undirected connections
between variables, representing their conditional relationship (i.e., partial corre-
lation) controlling for all other variables in the network.
The PMRF is often promoted as an exploratory method of generating causal
hypotheses about the underlying network structure (Borsboom & Cramer, 2013;
Epskamp, van Borkulo, et al., 2018). In many instances, this data-generating
structure is conceptualized as consisting of directed causal relationships, and the
PMRF is taken to reflect the causal skeleton, identifying the presence, but not the
direction, of direct causal links (e.g., van Borkulo et al., 2015; Boschloo, Scho-
evers, van Borkulo, Borsboom, & Oldehinkel, 2016; Haslbeck & Waldorp, 2018).
Typically researchers are agnostic regarding what specific form the underlying
causal structure takes, for instance, whether it consists of uni-directional or bi-
directional relationships, and whether the structure is cyclic or acyclic (Cramer,
Waldorp, van der Maas, & Borsboom, 2010; Epskamp, Waldorp, et al., 2018; Isvo-
ranu et al., 2016; McNally et al., 2015; Robinaugh, Millner, & McNally, 2016;
Van Borkulo et al., 2014; Costantini et al., 2015).
However, this agnostic approach to causal structure means that the task of
causal hypothesis generation is fundamentally intractable: If the class of casual
structure we wish to make inferences about is not clearly defined, it is impossible
to know how to go about generating hypotheses about that structure. In other
words, the causal information conveyed by the absence or presence of connec-
tions in a PMRF depends entirely on the precise mapping from that network to
the underlying directed causal structure.
The uncertainty resulting from this causal-agnosticism is typified by the dis-
cussion surrounding Directed Acyclic Graphs (DAGs) in the network analysis lit-
erature. DAGs are a popular approach to conceptualizing directed causal struc-
tures in the causal inference literature, and one for which the relationship with
PMRFs is both relatively simple and well-known (Spirtes et al., 2000; Pearl, 2009;
Lauritzen, 1996). Typically, the hypothesis generation heuristics suggested by
users of PMRFs in psychology are consistent with, and even seem to be derived
from, the DAG as an underlying structure (see Epskamp, Waldorp, et al., 2018
p.457-458; Epskamp, van Borkulo, et al., 2018 p.420, and Borsboom & Cramer,
2013 p.105). Simultaneously, however, the DAG is rejected as a plausible target
structure, for example due to the absence of cyclic effects. This represents a fun-
damental problem for researchers wishing to generate causal hypotheses: It is
unclear whether these same heuristics still apply when mapping PMRFs to some
8
2.2. Background
2.2 Background
In this section we will give an overview of network models, also known as graph-
ical models, two terms which we will use interchangeably in the current article.
Specifically we will review two instances of graphical models under considera-
tion in the current paper: The Pairwise Markov Random Field (PMRF), which
is used in the network analysis literature to generate causal hypotheses, and the
Directed Acyclic Graph (DAG), which we will use as an example target causal
structure. The remainder of this paper will make use of this background to allow
for an informed evaluation of PMRFs as hypothesis-generating tools, under two
scenarios: First, when the target casual structure is unspecified (i.e., the causally-
agnostic approach); Second, when the target causal structure is a DAG.
For both the PMRF and DAG we describe a special case which can be ob-
tained when the variables in question have a joint Gaussian (normal) distribu-
tion: the Gaussian Graphical Model (GGM) and a weighted DAG based on linear
regression. Finally, we describe the relationship between PMRFs and DAGs, as
described in the graphical modeling literature (Lauritzen, 1996).
9
2. The Challenge of Generating Causal Hypotheses
Unweighted Weighted
Support Support
Undirected
Pressure Pressure
(a) Pairwise Markov Random Field (PMRF) (b) Gaussian Graphical Model (GGM)
Support Support
Directed
Pressure Pressure
(c) Directed Acyclic Graph (DAG) (d) Regression path model (weighted DAG)
Figure 2.1: Four different types of network (graphical model), arranged by their edge-characteristics.
10
2.2. Background
11
2. The Challenge of Generating Causal Hypotheses
is estimated directly from the data using regularization techniques such as the
graphical lasso (Friedman, Hastie, & Tibshirani, 2008; Epskamp & Fried, 2018).
These techniques introduce bias in the parameter estimates in order to avoid
over-fitting. The graphical lasso technique returns an estimate of the precision
matrix Ω̂ in which small values are set exactly to zero (Friedman et al., 2008). The
presence of exact zero values in this matrix means that the estimated precision
matrix, and thus the corresponding GGM, is often simpler (more sparse) than the
true population precision matrix.
12
2.2. Background
⊥ X −de(i) |X pa(i) ,
Xi ⊥ (2.3)
which can be used to read off conditional (in)dependencies between pairs of vari-
ables from the DAG. For example, in Figure 2.1(c), we can derive that Support
and Pressure are marginally independent (Support ⊥ ⊥ Pressure |∅), because Pres-
sure is a non-descendant of Support, and Support has no parents in this graph
(pa(Support) = ∅). Any two nodes connected by an edge are dependent condi-
tional on any subset of other variables.
Further conditional (in)dependency relationships between any pair of vari-
ables in the graph can be derived from the structure of a DAG using so called
d-separation rules (Pearl, 2009). These rules allow us to relate DAG structures to
other graphical models such as the PMRF. The most important of the d-separation
rules for the current paper relates to situations in which two variables share a
common child, also known as a common effect or collider structure Xi → Xk ← Xj .
According to d-separation rules, the parent variables Xi and Xj are dependent
conditional on the collider variable Xk . For example in Figure 2.1(c), although
Support and Pressure are marginally independent, they are dependent when
conditioning on Stress (Support 6⊥ ⊥ Pressure | Stress). For substantive exam-
ples applying d-separation rules in social science settings, readers are referred
to Glymour (2006).
A weighted DAG can be obtained from a set of Gaussian variables X, assum-
ing linear relationships between these variables, using a linear regression path
model in which child nodes are predicted by their parents
X = a + BX + e (2.4)
13
2. The Challenge of Generating Causal Hypotheses
14
2.3. Using PMRFs to Generate Causal Hypotheses
Moral Graph
DAG Skeleton
(PMRF)
A B A B A B
C C C
D D D
A B A B A B
C C C
D D D
A B A B A B
C C C
D D D
Figure 2.2: Three examples of DAGs which all have the same skeleton, but result in different moral
graphs depending on the orientation of the edges in the DAG.
umn of Figure 2.2 shows the moral graph for each DAG. We see that although all
three DAGs share the same skeleton, each moral graph is distinct. Furthermore,
the moral graph contains additional edges connecting variables which are not con-
nected in the DAG, apart from the DAG in the second row of Figure 2.2, in which
there are no collider structures.
15
2. The Challenge of Generating Causal Hypotheses
can be used to generate causal hypotheses once the target structure is specified.
Specifically, we review what the mapping between DAGs and PMRFs described
above means for hypothesis generation, address the particular challenges which
arise when using the PMRF to infer multivariate patterns of relationships in an
underlying DAG, and describe some additional assumptions which are needed to
make causal hypothesis generation feasible in practice.
Notably, these heuristics are consistent with treating the PMRF as the moral graph
of an underlying DAG, as described in the previous section. Furthermore, these
heuristics are often derived explicitly with reference to relationship between PM-
RFs and DAGs (Borsboom & Cramer, 2013; Epskamp, Waldorp, et al., 2018).
However, these same heuristics are typically described and applied by re-
searchers who simultaneously reject the possibility that the underlying structure
is a DAG, for instance due to the hypothetical presence of causal “loops” Xi Xj
(e.g., Cramer et al., 2010; Isvoranu et al., 2016; McNally et al., 2015; Robinaugh
et al., 2016; Van Borkulo et al., 2014; Costantini et al., 2015). Moreover, beyond
the presence of such hypothetical loops, an agnostic approach to the underly-
ing causal structure is typically taken, in that the precise form of this alternative
structure is left unspecified. This represents a fundamental contradiction: With-
out specifying the form of the underlying causal structure, it is impossible to
verify whether these heuristics apply outside of the case used to derive them in
the first place.
In fact, we know for certain that for some types of causal structure, the heuris-
tics relating conditional dependencies (in the form of PMRF edges) to causal de-
pendencies described above do not hold. For instance, take it that the target
16
2.3. Using PMRFs to Generate Causal Hypotheses
causal structure takes the form of directed relationships between dynamic, time-
varying processes, such as described by a Local Independence Graph (Schweder,
1970; Aalen, 1987; Didelez, 2000). This type of structure allows us to specify
causal loops in the form of time-forward relationships between processes linked
by a system of differential equations. For this type of causal structure, it is well
known that both heuristics described above can fail even when there are no un-
observed common causes: 1) observations of two causally independent processes
can be conditionally dependent (Aalen, Røysland, Gran, Kouyos, & Lange, 2016;
Maxwell & Cole, 2007) and 2) observations of two causally dependent processes
can be conditionally independent under certain conditions (Kuiper & Ryan, 2018).
This counter-example shows the critical necessity of specifying a target causal
structure. First, it shows that we cannot simply assume that the heuristic rules
described above hold for any and all types of causal structure. Second, it shows
that for an intuitive time-forward interpretation of causal loops, these heuris-
tics cannot be applied in an out-of-the-box fashion. This means that if a re-
searcher wishes to simultaneously apply these hypothesis generation heuristics,
while rejecting the notion of an underlying DAG, the burden of proof is on that
researcher to provide an alternative formalism for the underlying structure, and
to prove that the heuristics described above hold for that formalism. Finally, the
counter-example highlights the profound advantage of using the DAG as a target
causal structure, as we know for certain that these heuristics can be applied to
learn about an underlying DAG structure. Given that the DAG is, to our knowl-
edge, the only formalized causal structure for which we can currently be certain
these heuristics apply, the best case scenario for using these heuristics to generate
causal hypotheses is if the underlying causal structure is a DAG.4
In the following, we will examine the difficulties which remain in this hy-
pothesis generation task, even when the underlying structure is a (faithful) DAG
consisting only of observed variables (with these additional conditions for now
left implied, and discussed at the end of the section).
17
2. The Challenge of Generating Causal Hypotheses
the PMRF cannot be taken to imply the presence of a directed edge between those
two variables in the underlying DAG: This latter statement is a property of the
DAG skeleton, and not the moral graph.
It is pertinent to note here an overlap in terminology used in the DAG and net-
work analysis literature. In the network analysis literature, multiple researchers
refer to PMRF models explicitly as representing the causal skeleton of some di-
rected causal structure, encoding the presence but not the direction of a causal
relationship (van Borkulo et al., 2015; Borsboom & Cramer, 2013; Isvoranu et al.,
2016; Haslbeck & Waldorp, 2018; Fried, Epskamp, Nesse, Tuerlinckx, & Bors-
boom, 2016; Armour, Fried, Deserno, Tsai, & Pietrzak, 2017). We reiterate that,
if the underlying causal structure is a DAG, then in general the PMRF will not be
equivalent to the skeleton of that DAG.
This point bears repeating because it lies at the heart of the difficulties which
arise when using PMRFs to hypothesize about DAGs. The value of PMRFs for
discovering causal structure is often justified by a comparison to networks where
edges represent marginal dependencies (for instance, full rather than partial cor-
relations). In contrast to these marginal dependency graphs, by conditioning
on all variables, the PMRF omits what Costantini et al. (2015) and Heeren and
McNally (2016) (amongst others) refer to as “spurious” edges: That is, connec-
tions between variables which do not reflect the presence of a direct connection
in the causal data-generating structure (but which reflect either a common cause
or indirect relationship). However, what is perhaps under-appreciated by users
less familiar with the graphical modeling literature is that, when the underlying
causal structure is a DAG, conditioning on all other variables in the network also
has the effect of inducing additional “spurious” connections between variables
which are unconnected in the DAG, resulting from conditioning on collider vari-
ables (as discussed by, amongst others, Epskamp, Waldorp, et al., 2018; Borsboom
& Cramer, 2013).
18
2.3. Using PMRFs to Generate Causal Hypotheses
19
2. The Challenge of Generating Causal Hypotheses
PMRF DAGs
1) 2) 3) 4) 5)
A B A B A B A B A B
C C C C C
A B D D D D D
6) 7) 8) 9)
C
A B A B A B A B
C C C C
D D D D
D 10)
A B
11)
A B
12)
A B
13)
A B
C C C C
D D D D
Figure 2.3: A Pairwise Markov Random Field (PMRF) and each DAG structure which generates that
PMRF.
20
2.3. Using PMRFs to Generate Causal Hypotheses
the form of linear SEM models. We illustrate the generation of causal hypotheses
from GGMs in the next section.
2.3.4.1 Sufficiency
The first of these assumptions is known as sufficiency, which essentially entails
that the underlying DAG consists of only the observed variables X (in the exam-
ples above, A, B, C and D) in the sense that there are no unobserved common
causes of two or more observed variables (Lauritzen, 1996; Pearl, 2009; Spirtes et
al., 2000).7 This means that the conditional dependencies captured by the PMRF
of X are assumed to come about due to directed causal relationships between the
observed variables X, and not due to relationships the observed variables share
with unobserved variables (such as an unobserved common cause, or an unob-
served collider variable on which we have unwittingly conditioned).
Relaxing the sufficiency assumption means that the task of generating causal
hypotheses becomes much more difficult, as, depending on the number and type
of missing variables we are willing to consider, there are many more DAGs which
may have generated the given PMRF. For example, if we consider there to be a
single unobserved variable, E, which acts as a common cause of A and B, then
we must consider the DAG in panel (a) of Figure 2.4 as a plausible underlying
causal structure: This DAG produces the same conditional dependencies present
between the observed variables A − D as the PMRF in Figure 2.3. If we consider
only this exact type of unobserved variable, we must now consider 9 additional
causal structures in which at least one of the connections in the PMRF is induced
by this common cause (see Appendix 2.A).
If we relax the sufficiency assumption even further, for example allowing E
to be a common cause of other variables, as in panel (b) of Figure 2.4 or allowing
more than one unobserved common cause as in panel (c) of Figure 2.4, deriving
a complete set of possible underlying DAGs quickly becomes infeasible. In gen-
eral, without the sufficiency assumption, the causal information conveyed by the
presence of an edge (as stated in Heuristic 1) becomes less certain, as there are
7 More precisely we could say that there are no unblocked back-door paths, based on d-separation
rules, passing through unobserved variables, that connect any pair of observed variables (Pearl, 2009).
21
2. The Challenge of Generating Causal Hypotheses
2.3.4.2 Faithfulness
The second major assumption needed to generate causal hypotheses on the ba-
sis of the PMRF is called faithfulness (Spirtes et al., 2000; Pearl, 2009). A DAG
and associated probability distribution P meet the faithfulness condition if every
conditional (in)dependence relation in P is entailed by the causal Markov condi-
tion in Equation 2.3 (Spirtes et al., 2000). For example, if two variables Xi and Xj
are marginally independent, then by faithfulness the corresponding DAG should
have no directed paths which can be traced from Xi to Xj , e.g. Xi → Xk → Xj .
This means that we assume away the possibility that Xi and Xj are connected by
two different directed pathways, which when combined in the marginal relation-
ship between Xi and Xj , exactly cancel one another out. This would happen if,
for example, there was a negative direct pathway Xi → Xj as well as a positive
indirect pathway of equal size through Xk .
In panel (d) of Figure 2.4 we show a weighted DAG which would result in a
violation of faithfulness. In this DAG, both B and D have a positive direct effect
on C, making it a collider between them. Conditioning on this collider induces
a negative conditional relationship between B and D. However, simultaneously,
B has a positive direct effect on D: When combined, this negative and positive
relationship cancel one another out, and so B and D appear to be conditionally
independent given C - the partial correlation of B and D given C is zero, so they
are unconnected in the PMRF (see Appendix 2.A for details). In general, with-
out the faithfulness condition, the causal information conveyed by the absence
of an edge (as in Heuristic 2) becomes less certain. In theory, the assumption of
faithfulness is often justified in the context of relating the true probability dis-
tribution to a DAG; the probability of the true DAG having two pathways which
exactly cancel out is said to be negligible (Spirtes et al., 2000).
22
2.4. Empirical Illustration
23
2. The Challenge of Generating Causal Hypotheses
detail the type of inferences which researchers typically aim to make based on
an estimated PMRF, and how specifying a target structure influences these infer-
ences.
In this illustration, we will make use of a novel tool which takes a precision
matrix, as estimated in the GGM method, and produces the set of statistically
equivalent weighted DAGs (hereby referred to as the SE-set) under ideal condi-
tions. This can be considered the weighted equivalent of the moral-equivalent
set, described in the previous section. In deriving the SE-set, we make use of the
sufficiency assumption, outlined above, but we do not need to invoke the faith-
fulness assumption. Since we are deriving a weighted DAG equivalent to a linear
path model, we also assume linear relationships between variables (as captured
by the GGM), and uncorrelated error terms. Full details on the SE-set algorithm
can be found in Appendix 2.B and it can be downloaded as an R-package from
the github page of the first author.8
Hoorelbeke et al. (2016) analyzed the network structure of cognitive risk
and resilience factors in a cross-sectional sample of 69 remitted depression pa-
tients. This network consisted of six variables: self-reported cognitive con-
trol (BRIEF WM); performance on a behavioral cognitive control measures
(PASAT ACC); adaptive emotion regulation strategies (Adapt ER), maladaptive
emotion regulation strategies (Maladapt ER); resilience levels (Resilience); and
residual depression symptoms (Resid Depres).
In the original article, the authors estimated a GGM, and additionally a di-
rected relative importance network. The relative importance network is another
way of encoding how well one variable is predicted by another, based on ex-
plained variance, and is often used in conjunction with an estimated PMRF
model in order to further inform the generation of causal hypotheses (Robinaugh,
LeBlanc, Vuletich, & McNally, 2014; McNally et al., 2015; Heeren & McNally,
2016). We begin by first reproducing their analysis and examining the conclu-
sions they make on the basis of this analysis. Full details of the re-analysis is
given in Appendix 2.C. Following this, we use the SE-set algorithm to re-examine
and expand upon the causal hypotheses generated from the original analysis.9
24
2.4. Empirical Illustration
Maladapt ER Adapt ER
Adapt ER
Resilience BRIEF_WM
Maladapt ER Resilience BRIEF_WM
PASAT_ACC
Figure 2.5: Networks of cognitive risk and resilience factors based on re-analysis of Hoorelbeke et al.
(2016)
25
2. The Challenge of Generating Causal Hypotheses
shows the highest level of betweenness, closeness and strength centrality, and
is the only node with higher out-strength values (0.77) than in-strength values
(0.64). This means that Resilience appears to have a greater value in predicting
other variables than vice versa. However, the predictability of each node in Fig-
ure 2.5(b) shows that Resilience is the most predictable node in absolute terms,
followed by working memory and residual depressive symptoms. In compari-
son, the variables relating to emotion regulation strategies have relatively lower
predictability.
26
2.4. Empirical Illustration
Maladapt ER Maladapt ER
Adapt ER Adapt ER
PASAT_ACC PASAT_ACC
Maladapt ER Maladapt ER
Adapt ER Adapt ER
PASAT_ACC PASAT_ACC
Figure 2.6: Networks representing frequency of occurrence of edges in the DAG equivalence set.
maximum of 6! = 720 different weighted DAGs, one for every possible ordering of
the six variables from cause to effect. After rounding and the removal of duplicate
weights matrices, we are left with only nDAG = 58 unique weighted linear DAG
structures. We can explore the SE-set by inspecting some examples of DAGs in
the set and summarizing different characteristics of members of the set.
27
2. The Challenge of Generating Causal Hypotheses
Maladapt ER Maladapt ER
Adapt ER Adapt ER
PASAT_ACC PASAT_ACC
(a) Any edge relative frequency (b) Directed edge relative frequency
of the other variables in the network, as may have been hypothesized on the basis
of the relative importance network. Finally, Figure 2.6(d) represents an example
where Resilience is the common effect, i.e. caused by all other variables in the
network. Note that in this example, there are a number of edges present which
are not present in the estimated GGM, even though this weighted DAG does re-
produce the input precision matrix (for a discussion on how this may arise, see
the discussion of the faithfulness assumption in the previous section and Ap-
pendix 2.A).
By examining different elements of the SE-set we can see that there is a variety
of distinct directed structures that may have resulted in the estimated GGM. Re-
searchers can use substantive expertise to rule out some elements of the SE-set:
For instance, based on previous research we may know that depressive symp-
toms are a cause of working memory problems, and not vice versa, which would
rule out DAGs 52 and 20 (Figure 2.6(b) and 2.6(d)) respectively as plausible data-
generating structures. In the absence of such substantive knowledge, in the cur-
rent paper we treat each of the 58 DAGs in the SE-set as equally plausible data-
generating structures.
28
2.4. Empirical Illustration
is present. In this particular instance, the thick arrows denote edges which are
always present in the SE-set, with the less thick edges indicating that these edges
were present in only 52 percent of the SE-set.
From Figure 2.7(a) we see that every edge which is present in the estimated
GGM is present in every member of the SE-set. We can further see that none
of the weighted DAGs contain edges between Adapt ER and Maladapt ER, or
between PASAT ACC and any other variable. From this we can say that, for this
particular example, an edge in the GGM appears to correspond to the presence of
an edge in an underlying weighted DAG; none of the edges in the GGM appear to
be induced due to conditioning on a collider in this case. Conversely, the absence
of an edge in the GGM does not always correspond to the absence of an edge
in the underlying weighted DAG. In half the members of the SE-set there is at
least one directed relationship present which is absent in the GGM: These are
weighted DAGs which are unfaithful, in the sense that two or more conditional
dependencies approximately cancel one another out (as discussed in the previous
section).
We can further explore the equivalence set by seeing in what proportion of
the SE-set edges of a particular direction occur. This is represented as a weighted,
directed network in Figure 2.7(b), where the edges A → B describe how often a
corresponding directed relationship A → B is present in the SE-set.11 We can
see from Figure 2.7b that for most pairs of variables, edges of opposite directions
occur with equal frequency. However, this is not the case for edges relating to
the Resilience variable, which was more often the child (or effect) than the par-
ent (cause). Resilience was the child of the working memory in 71 percent of the
equivalence-set DAGs, and a parent in 29 percent of the cases. A similar split
is present in the connections between Resilience and Resid Depress (71/29), Re-
silience and Adapt ER (69/31) and Resilience and Maladapt ER (69/31).
It is interesting to note the different ideas we get about a possible underlying
directed causal structure from Figure 2.7(b) than the relative importance network
in Figure 2.5(b). From the latter researchers may be tempted to conclude, that,
because Resilience has a greater predictive value for other variables than vice
versa, that Resilience is more likely to cause those variables than vice versa. From
the SE-set analysis we can say that, if we think the data-generating mechanism is
a linear weighted DAG, and each member of the SE-set is equally plausible, then
it is more than twice as likely that Resilience is caused by the other variables in
our graph than vice versa.
29
2. The Challenge of Generating Causal Hypotheses
Resilience
Resid Depress
Variable
Maladapt ER
BRIEF_WM
Adapt ER
Figure 2.8: Distribution of controllability values for each variable, over DAGs in the SE-set. Control-
lability is here defined as the explained variance R2 of each variable when predicted by its parents in
a given DAG.
ity. However, the true controllability will differ depending on which variables
are causes and effects of others. For example, Resilience will have the highest
controllability in Figure 2.6(d), where it is a common effect of all other variables,
and a controllability of zero in Figure 2.6(c), where it is a common cause of all
others.
Using the SE-set algorithm, we can try to quantify our uncertainty about con-
trollability, by calculating a distribution of this value over the different possible
weighted DAG structures. This is shown in Figure 2.8. From this figure we can
see that, in most elements of the SE-set, the controllability of Resilience is high,
as we would expect given that Resilience is more often an effect than a cause of
other variables. We can also see that the controllability of both maladaptive and
adaptive emotion regulation strategies is quite low across different DAGs, with
parents of both strategies explaining a maximum of 20 percent of the variance
in each. However there is much more uncertainty about the controllability of
residual depressive symptoms and working memory. The distribution of con-
trollability for each is quite wide, ranging from zero to sixty percent, and with
peaks of similar height all across this range. This analysis shows us that, while
predictability in a standard GGM analysis captures the upper bound of these con-
trollability distributions, the SE-set analysis can supplement this information by
showing the variance and peaks of these distributions.
2.5 Discussion
In this paper we have critically evaluated a widespread practice in the current lit-
erature on psychological network analysis: Using estimates of PMRF-based mod-
els to generate hypotheses about an underlying directed causal structure. We
30
2.5. Discussion
have shown that the utility of PMRF models for causal hypothesis generation is
critically dependent on what target causal structure is specified. When taking a
causally-agnostic approach, hypothesis generation is a fundamentally intractable
problem: The heuristics used by researchers to generate these hypotheses cannot
be assumed to apply to any and all causal structures. Defining a target structure,
however, allows us to generate causal hypotheses in a principled way. We illus-
trate this procedure using the DAG and a linear path model (weighted DAG) as
those target structures. In each case, we have shown that the specification of a
target structure allows for detailed and novel causal hypotheses to be generated,
based on the global structure of the estimated PMRF. In contrast, we have shown
how the isolated application of heuristics, even in combination with other tools
which encode predictive relationships, can lead to at best incomplete and at worst
incorrect inferences regarding the underlying pattern of causal relationships.
Throughout the paper we have made use of DAGs and weighted DAGs, in the
form of linear path models, as candidate target causal structures. D-separation
rules allowed us to derive the moral-equivalent set of DAGs from a PMRF, and the
SE-set algorithm introduced here allowed us to derive the statistically-equivalent
set of weighted DAGs from a GGM. It should be noted that the SE-set is best
considered an illustrative or exploratory tool. The benefit of using the SE-set al-
gorithm in the current paper is that it directly relates GGMs to potential under-
lying causal structures. As a stand-alone DAG estimation method, however, the
SE-set algorithm is limited for that same reason - inferences are entirely reliant
on the GGM estimate, and uncertainty in those estimates is not accounted for. Re-
searchers who wish primarily to estimate a (weighted) DAG structure from data
would be better served in using algorithms specifically designed for that purpose,
such as the PC, LiNGAM and FCI algorithms, amongst others (Spirtes et al., 2000;
Shimizu, Hoyer, Hyvärinen, & Kerminen, 2006; Spirtes, Meek, & Richardson,
1995), many of which are implemented in R packages such as pcalg (M. Kalisch,
Mächler, Colombo, Maathuis, & Bühlmann, 2012) and bnlearn (Scutari, 2010).
Note that typically these algorithms are unable to uniquely identify a DAG from
data, instead estimating a Markov-equivalent class, that is, a set of DAGs with fully
equivalent conditional dependency relationships (cf. Andersson et al., 1997).12
Numerous tutorials on the use and advantages of different DAG estimation algo-
rithms, and the assumptions necessary for each, are available for the interested
reader (e.g., Malinsky & Danks, 2018; Spirtes & Zhang, 2016).
However, the possibility of directly estimating DAG structures from data
should not be taken as a panacea to the problem of inferring multivariate pat-
terns of causal relationships from data. The focus on DAGs in the current paper
is motivated by our focus on the heuristic mapping rules which are applied by
researchers to generate causal hypotheses: If the underlying structure is a DAG,
this represents the best case scenario for this practice, in the sense that we are
guaranteed those heuristics are correct. A focus on DAGs as target causal struc-
tures also confers some additional benefits, as they have been extensively studied
12 Furthermore, since these algorithms take disparate approaches to estimating DAG structures, we
are not guaranteed that the moral graph implied by the estimated Markov-equivalent DAGs will be
equal to the PMRF estimate, even when the underlying structure is a DAG.
31
2. The Challenge of Generating Causal Hypotheses
in the causal inference literature, and can be easily used in conjunction with, for
example, interventionist theories of causality (Spirtes et al., 2000; Pearl, 2009;
Dawid, 2002; Richardson & Robins, 2013). That being said, it is possible that
the DAG is not the optimal formalism for causal structure in psychological set-
tings, and that some alternative causal structure may be more appropriate (for
various discussions on this point, see McNally, 2016; Dawid, 2010; Cartwright,
1999, 2007). The problem however, is that it appears as though no such alter-
native formalism is readily available, or at least well-known in the psychology
literature.
Cyclic causal models (i.e., graphs which allow causal “loops” Xi Xj ) are
in general underdeveloped, and the interpretation of cyclic causal effects with-
out invoking a notion of time-forward dependency is notoriously difficult (e.g.,
Spirtes, 1995; Hayduk, 2009). As noted by, amongst many others, Borsboom
et al. (2012) and Epskamp, Waldorp, et al. (2018), cyclic effects between dy-
namic processes can easily be represented as acyclic time-forward relationships
Xi,t → Xj,t+1 → Xi,t+1 , either in a DAG or other related structure, as in time
series analysis (e.g., Bringmann et al., 2013; Hamaker & Dolan, 2009). For-
malized approaches to defining these time-forward causal relationships exist
both in discrete-time (Eichler & Didelez, 2010; Dahlhaus & Eichler, 2003) and
continuous-time (Didelez, 2000; Aalen, 1987). Alternatively, undirected graphs
such as the Ising model can potentially be given a dynamic interpretation, reflect-
ing symmetric time-forward directed relationships, by invoking Glauber dynam-
ics (Glauber, 1963; Haslbeck, Epskamp, Marsman, & Waldorp, 2018; Marsman
et al., 2018), as applied in recent theoretical models in psychology (Cramer et
al., 2016; Dalege et al., 2016; Borsboom, 2017). However, if we take these types
of dynamic systems to be the target structure(s), many additional assumptions
must be invoked in order to ensure that PMRF edges estimated from a given data
source reflect causal relationships therein. For example, assumptions about the
frequency of the underlying process are needed to recover the structure of a dy-
namic system from time-series data (Papoulis & Pillai, 2002; Marks, 2012) and
the oft-criticized ergodicity assumption is needed to ensure recovery of dynamic
relationships from cross-sectional data (Molenaar, 2004; Hamaker, 2012; Mole-
naar, 2008). There may be yet other approaches to cyclic causal effects which
are more promising for psychological applications: For example, cyclic effects
which represent dynamic systems in an equilibrium state (Mooij, Janzing, Hes-
kes, & Schölkopf, 2011; Forré & Mooij, 2018, 2019). However, more work is
needed to assess their suitability and applicability in practice, and to establish
any potential links between those types of causal structures and the conditional
dependency patterns estimated by methods such as the PMRF.
Outside of their use for causal hypothesis generation, there are many attrac-
tive reasons to use PMRF-based models in practice. Amongst other things, they
allow for the identification of predictive relationships, sparse descriptions of sta-
tistical dependency relationships in a multivariate density, and may be used as
a variable clustering or latent variable identification method (e.g., Golino & Ep-
skamp, 2017). However, we have shown that, when the aim is to uncover some
aspects of the underlying causal structure, it is generally unclear what conclu-
32
2.5. Discussion
sions we can draw directly from an estimated network structure alone (that is,
without further specification and exploration of a target structure). Ours is not
the first paper to come to such a conclusion: Dablander and Hinne (2019) showed
that node centrality in a GGM is a poor indicator of causal influence in an un-
derlying DAG; Bos et al. (2017) showed that conclusions about causal structure
based on a cross-sectional GGM do not generalize to those made if a time-forward
causal structure is assumed; Haslbeck and Ryan (2019) showed that it is typically
unclear how to use various statistical methods common in the network approach
literature, including the GGM, to draw conclusions about an underlying dynamic
system from time series data. This pattern of results is concerning, as our read-
ing of the literature suggests that discovering patterns of causal relationships
is the primary motivation behind most of the applications of these methods in
empirical research: These statistical methods appear to hold the promise of di-
rectly uncovering the causal interactions which are a cornerstone of the theoret-
ical framework which motivated their development (Borsboom & Cramer, 2013;
Borsboom, 2017).
In this paper we suggest one route by which the search for causal relationships
in psychological networks can move forward: The investigation and specification
of target causal structures. Specifying a target causal structure establishes the
rules and boundaries through which we can make some inductive or abductive
inference from an estimated network model to the underlying system of causal
relationships. The feasibility of this approach relies on first the availability of
suitable formalized causal structures and second the establishment of how infor-
mation captured by different statistical methods relates to those structures. An
alternative to this route would be to use methods such as the PMRF as purely
deductive or confirmatory tools, to test the implications and predictions of pre-
specified causal theories: This in turn relies on our ability to specify these the-
ories a-priori. In either case, the critical observation is that to use PMRFs, or in
essence, any statistical method, to discover potential causal relationships, it is
necessary to move beyond the causally-agnostic approach. Only when we make
clear what it is we wish to make some inference about, can we hope to use any
tool to make those inferences in a principled way. Only then can the power of this
and other statistical approaches be fully realized: By understanding what types
of inferences we can make, when we can make them, and how to go about doing
this in practice.
33
2. The Challenge of Generating Causal Hypotheses
D D D D
(e) (f) (g) (h)
E E E E
A B A B A B A B
C C C C
D D D D
Figure 2.9: Eight additional DAGs which generate the same PMRF between the variables A − D (as
shown in Figure 2.3). For each DAG, the set of variables A − D violates the sufficiency assumption,
with respect to the variable E, a common cause of both A and B.
34
2.B. The SE-set Algorithm: A Tool to Aid Causal Hypothesis Generation
35
2. The Challenge of Generating Causal Hypotheses
36
2.B. The SE-set Algorithm: A Tool to Aid Causal Hypothesis Generation
Support, Stress, Worry}. If the rows and columns of the covariance matrix Σ̃ are
sorted according to the topological ordering, then Equation (2.7) gives a unique
solution. In that case, B will be a lower triangular matrix with zero’s on the
diagonal, and Equation (2.7) will be equivalent to an LDLT matrix decomposition
(Abadir & Magnus, 2005).
37
2. The Challenge of Generating Causal Hypotheses
the variables V , of which there are p!. The output of the algorithm is thus a set of
matrices
B = {B1 , B2 , . . . Bp! } (2.8)
of size p!. To generate B we first generate V , the set containing each p! topological
ordering of the variables V . The elements of V are thus ordered sequences of p
variable names
Ω−1 = Σ̃
(I − Bq )−1 = I + B + B 2 + B 3 + . . .
X∞
= Bz
z=0
where B z will also be lower-triangular with zero-diagonal for all z > 0. This
means that (I − Bq )−1 will be a lower triangular matrix with diagonal elements
equal to one (Abadir & Magnus, 2005).
In practice the LDLT decomposition of a matrix can be done by first calculat-
ing the cholesky decomposition
Σ˜q = GG T (2.12)
38
2.B. The SE-set Algorithm: A Tool to Aid Causal Hypothesis Generation
39
2. The Challenge of Generating Causal Hypotheses
will describe identical linear path models: for a given weighted DAG, there may
be more than one equivalent topological ordering. The intended use of this tool is
for researchers interested in the different distinct (causal) structures which may
generate Σ̃, where we consider a-priori all distinct structures to be equally plau-
sible. As such, the implementation of this algorithm by default removes any
duplicate matrices in B.
One disadvantage to rounding and thresholding matrix elements is that el-
ements of B may no longer be statistically equivalent, that is, they may imply
different variance-covariance matrices. However these differences are likely to
be relatively small, and this disadvantage is offset by the increased simplification
and interpretability of the different path models.
40
2.C. Empirical Illustration Details
Table 2.1: Reported correlations and estimated partial correlations for the empirical example. The
upper triangle shows the marginal correlations reported by Hoorelbeke et al. (2016). The lower tri-
angle shows the partial correlations estimated using the reported correlation matrix as input to the
EBICglasso function from the qgraph package.
Table 2.2: Average error (“bias”) of elements of B in reproducing the input precision matrix Ω̂. Zero
elements of Ω̂ are highlighted in red.
We can see from Table 2.2 that positive elements of Ω̂, for example the diag-
onal elements as well as ω51 , ω54 and ω65 appear to have a slight positive bias,
and negative elements (ω53 , ω61 ) a slight negative bias. The true zero-elements,
highlighted in red in the table below, have a mix of positive, negative and zero
bias. All parameters relating to PASAT ACC, which is unconnected to all other
variables, are recovered as exactly zero. However other non-zero elements are
recovered as positive or negative non-zero elements, but the size of these ele-
ments is very small. This is not necessarily problematic, as the input precision
matrix makes use of thresholding in any case, where such parameters would be
set exactly to zero.
This performance analysis is conducted on the un-rounded equivalence set,
also including duplicates. Removing duplicates does not make a substantial dif-
ference to the results due to the small variances in individual errors. Rounding
41
2. The Challenge of Generating Causal Hypotheses
makes some difference to the performance, but not substantially: the signs of
some biases change, and the absolute values of the largest biases are now larger,
of the order 10−3 .
42
Chapter 3
A Continuous-Time Approach
to Intensive Longitudinal
Data: What, Why and How?
Abstract
This chapter has been adapted from: Ryan, O., Kuiper, R. M. & Hamaker, E. L. (2018). A
Continuous-Time Approach to Intensive Longitudinal Data: What, Why and How? In K. v. Montfort,
J. H. L. Oud, & M. C. Voelkle (Eds.) Continuous Time Modeling in the Behavioral and Related Sciences.
New York, NY: Springer. Author contributions: OR and ELH conceptualized the initial project. OR
and RMK wrote the R code and generated figures. OR wrote the paper. RMK and ELH helped further
develop the ideas in the project, discussed progress and provided textual feedback.
43
3. A Continuous-Time Approach to Intensive Longitudinal Data
3.1 Introduction
The increased availability of intensive longitudinal data – such as obtained with
ambulatory assessments, experience sampling, ecological momentary assess-
ments and electronic diaries – has opened up new opportunities for researchers
to investigate the dynamics of psychological processes, that is, the way psycho-
logical variables evolve, vary and relate to one another over time (cf. Bolger &
Laurenceau, 2013; Hamaker et al., 2005; Chow, Ferrer, & Hsieh, 2011). A useful
concept in this respect is that of people being dynamic systems whose current
state depends on their preceding states. For instance, we may be interested in the
relationship between momentary stress and anxiety. We can think of stress and
anxiety as each defining an axis in a two-dimensional space, and let the values of
stress and anxiety at each moment in time define a position in this space. Over
time, the point that represents a person’s momentary stress and anxiety moves
through this two-dimensional space, and our goal is to understand the lawful-
ness that underlie these movements.
There are two frameworks that can be used to describe such movements: 1)
the discrete-time (DT) framework, in which the passage of time is treated in dis-
crete steps; and 2) the continuous-time (CT) framework, in which time is viewed
as a continuous variable. Most psychological researchers are at least somewhat
familiar with the DT approach, as it is the basis of the vast majority of longi-
tudinal models used in the social sciences. In contrast, CT models have gained
relatively little attention in fields such as psychology: This is despite the fact
that many psychological researchers have been advocating their use for a long
time, claiming that the CT approach overcomes practical and conceptual prob-
lems associated with the DT approach (e.g., Boker, 2002; Chow et al., 2005; Oud
& Delsing, 2010; Voelkle, Oud, Davidov, & Schmidt, 2012). We believe there are
two major hurdles that hamper the adoption of the CT approach in psychological
research. First, the estimation of CT models typically requires the use of special-
ized software (cf. Driver, Oud, & Voelkle, 2017; Chow, Ferrer, & Nesselroade,
2007; Oravecz, Tuerlinckx, & Vandekerckhove, 2016) or unconventional use of
more common software (cf. Boker, Neale, & Rausch, 2004; Boker, Deboeck, Edler,
& Keel, 2010; J. S. Steele & Ferrer, 2011). Second, the results from CT models are
not easily understood, and researchers may not know how to interpret and rep-
resent their findings.
Our goal in this chapter is twofold. First, we introduce readers to the per-
spective of psychological processes as CT processes; we focus on the conceptual
reasons for which the CT perspective is extremely valuable in moving our under-
standing of processes in the right direction. Second, we provide a didactic de-
scription of how to interpret the results of a CT model, based on our analysis of
an empirical dataset. We examine the direct interpretation of model parameters,
examine different ways in which the dynamics described by the parameters can
be understood and visualized, and explain how these are related to one another
throughout. We will restrict our primary focus to the simplest DT and CT mod-
els, that is, first-order (vector) autoregressive models and first-order differential
equations.
44
3.2. Two Frameworks
yτ = c + Φyτ−1 + τ , (3.1)
45
3. A Continuous-Time Approach to Intensive Longitudinal Data
46
3.2. Two Frameworks
dy(t) dW (t)
= A(y(t) − µ) + G . (3.2)
dt dt
The model representation in Equation (3.2) is referred to as the differential
dy(t)
form as it includes the derivative dt . The same model can be represented in the
integral form, in which the derivatives are integrated out, sometimes referred to
as the solution of the derivative model. The integral form of this particular first-
order differential equation is known as the CT-VAR(1) or Ornstein-Uhlenbeck
model (Oravecz, Tuerlinckx, & Vandekerckhove, 2011). In this form, we can de-
scribe the same system, but now in terms of the positions of the system (i.e., the
values the variables take on) at different points in time. For notational simplic-
ity, we can represent y(t) − µ as y c (t), denoting the position of the process as a
deviation from its equilibrium.
The CT-VAR(1) model can be written as
where A has the same meaning as above, the V × 1 vector y c (t − ∆t) represents the
position as a deviation from equilibrium some time-interval ∆t earlier, e repre-
sents the matrix exponential function, and the V × 1 column vector w(∆t) repre-
sents the stochastic innovations, the integral form of the Wiener process in Equa-
tion (3.2). These innovations are normally distributed with a variance-covariance
matrix that is a function of the time-interval between measurements ∆t, the drift
matrix A, and the diffusion matrix Γ (cf. Voelkle et al., 2012).2 As the variables
in the model have been centered around their equilibrium, we omit any intercept
term. The relationship between lagged variables, that is, the relationships be-
tween the positions of the centered variables in the multivariate space, separated
by some time-interval ∆t, is an (exponential) function of the drift matrix A and
the length of that time-interval.
2 Readers should note that there are multiple different possible ways to parameterize the CT
stochastic process in integral form, and also multiple different notations used (e.g., Oravecz et al.,
2011; Voelkle et al., 2012).
47
3. A Continuous-Time Approach to Intensive Longitudinal Data
Despite this relatively simple relationship, it should be noted that taking the ex-
ponential of a matrix is not equivalent to taking the exponential of each of the
elements of the matrix. That is, any lagged effect parameter φij (∆t), relating vari-
able i and variable j across time-points, is not only dependent on the correspond-
ing CT cross-effect aij , but is a non-linear function of the interval and every other
element of the matrix A. For example, in the bivariate case the DT cross-lagged
effect of y1 (t − ∆t) on y2 (t), denoted φ21 (∆t), is given by
1
√2 2 1
√2 2
a21 (e 2 (a11 +a22 + a11 +4a12 a21 −2a11 a22 +a22 )∆t − e 2 (a11 +a22 − a11 +4a12 a21 −2a11 a22 +a22 )∆t )
q
a211 + 4a12 a21 − 2a11 a22 + a222
(3.5)
48
3.2. Two Frameworks
Equation 3.4 for Φ < 0. As such, we can refer to DT-VAR(1) models with a CT-
VAR(1) equivalent as those which exhibit ‘positive autoregression’. We will focus
throughout on the CT-VAR(1) as the data-generating model.3
49
3. A Continuous-Time Approach to Intensive Longitudinal Data
tor (e.g., Boker, Montpetit, Hunter, & Bergeman, 2010), in which the eigenvalues
of A are complex, with a negative real part. The fixed point of such a system is
described as a stable spiral. Further detail on the interpretation of these two types
of systems is given in the corresponding sections.
50
3.4. Making Sense of CT Models
lation studies indicate that this largely reduces the bias that results from using
DT estimation of unequally spaced data (De Haan-Rietdijk, Voelkle, Keijsers, &
Hamaker, 2017).
Furthermore, the issue of comparability between studies that use different
time-intervals can be solved, in certain circumstances, by a simple transforma-
tion of the estimated Φ matrix, described in more detail by Kuiper and Ryan
(2018). Given an estimate of Φ(∆t) we can solve for the underlying A using
Equation 3.4. This is known as the “indirect method” of CT model estimation
(Oud, van Leeuwe, & Jansen, 1993). However this approach cannot be applied
in all circumstances, as it involves using the matrix logarithm, the inverse of the
matrix exponential function. As the matrix logarithm function in the general
case does not give a unique solution, this method is only appropriate if both the
estimated Φ(∆t) and true underlying A matrices have real eigenvalues only (for
further discussion of this issue see Hamerle, Nagl, & Singer, 1991).
However, the CT perspective has added value above and beyond the poten-
tial practical benefits discussed above. Multiple authors have argued that psy-
chological phenomena, such as stress, affect and anxiety, do not vary in discrete
steps over time, but likely vary and evolve in a continuous and smooth manner
(Boker, 2002; Gollob & Reichardt, 1987). Viewing psychological processes as CT
dynamic systems has important implications for the way we conceptualize the
influence of psychological variables on each other. Gollob and Reichardt (1987)
give the example of a researcher who is interested in the effect of taking aspirin
on headaches: This effect may be zero shortly after taking the painkiller, substan-
tial an hour or so later, and near zero again after twenty-four hours. All of these
results may be considered as accurately portraying the effect of painkillers on
headaches for a specific time-interval, although each of these intervals considered
separately represent only a snapshot of the process of interest.
It is only through examining the underlying dynamic trajectories, and ex-
ploring how the cross-lagged relationships evolve and vary as a function of the
time-interval, that we can come to a more complete picture of the dynamic sys-
tem of study. We believe that – while the practical benefits of CT modeling are
substantial – the conceptual framework of viewing psychological variables as CT
processes has the potential to transform longitudinal research in this field.
51
3. A Continuous-Time Approach to Intensive Longitudinal Data
52
3.4. Making Sense of CT Models
example only, we will throughout interpret the estimated drift matrix parameter
as though they are true population parameters.
such that
" #
dDo(t)
E = −0.995Do(t) + 0.573T i(t) (3.8)
dt
" #
dT i(t)
E = 0.375Do(t) − 2.416T i(t) (3.9)
dt
where the rates of change of Down and Tired at any point in time are both de-
pendent on the positions of both Down and Tired at that time.
Before interpreting any particular parameter in the drift matrix, we can deter-
mine the type of dynamic process under consideration by inspecting the eigen-
values of A. The eigenvalues of A are λ1 = −2.554 and λ2 = −0.857; since both
eigenvalues are negative, the process under consideration is stable. This means
that if the system takes on a position away from equilibrium (e.g., due to a ran-
dom shock from the stochastic part of the model on either Down or Tired), the
53
3. A Continuous-Time Approach to Intensive Longitudinal Data
system will inevitably return to its equilibrium position over time. It is for this
reason that the equilibrium position or fixed point in stable systems is also de-
scribed as the attractor point, and stable systems are described as equilibrium-
reverting. As the eigenvalues of the system are real-valued as well as negative,
the system returns to equilibrium with an exponential decay; when the process is
far away from the equilibrium, it takes on a greater velocity, that is, moves faster
towards equilibrium. We can refer to the type of fixed point in this system as a
stable node (Strogatz, 2015).
Typical of such an equilibrium-reverting process, we see negative CT auto-
effects a11 = −0.995 and a22 = −2.416. This reflects that, if either variable in
the system takes on a position away from the equilibrium, they will take on a
velocity of opposite sign to this deviation, that is, a velocity which returns the
process to equilibrium. For higher values of Do(t), the rate of change of Do(t) is
of greater (negative) magnitude, that is, the velocity towards the equilibrium is
higher. In addition, the auto-effect of T i(t) is more than twice as strong (in an
absolute sense) as the auto-effect of Do(t). If there were no cross-effects present,
this would imply that T i(t) returns to equilibrium faster than Do(t); however,
as there are cross-effects present, such statements cannot be made in the general
case from inspecting the auto-effects alone.
In this case the cross-effects of Do(t) and T i(t) on each others rates of change
are positive rather than negative. Moreover, the cross-effect of T i(t) on the rate
of change of Do(t) (a12 = 0.573) is slightly stronger than the corresponding cross-
effect of Do(t) on the rate of change of T i(t) (a21 = 0.375). These cross-effects
quantify the force that each component of the system exerts on the other. How-
ever, depending on what values each variable takes on at a particular point in
time t, that is, the position of the system in each of the Do(t) and T i(t) dimen-
sions, this may translate to Do(t) pushing T i(t) to return faster to its equilibrium
or to deviate away from its equilibrium position, and vice versa. To better un-
derstand both the cross-effects and auto-effects described by A, it is helpful to
visualize the possible trajectories of our two-dimensional system.
54
3.4. Making Sense of CT Models
55
3. A Continuous-Time Approach to Intensive Longitudinal Data
1.0
1.0
Do(t) Do(t)
Ti(t) Ti(t)
0.5
0.5
Variable Values Y(t)
0.0
−0.5
−0.5
−1.0
−1.0
0 1 2 3 4 0 1 2 3 4
Time t Time t
1.0
Do(t) Do(t)
Ti(t) Ti(t)
0.5
0.5
Variable Values Y(t)
0.0
−0.5
−0.5
−1.0
−1.0
0 1 2 3 4 0 1 2 3 4
Time t Time t
Figure 3.1: Impulse response function for the model in Equation (3.9) for four different sets of im-
pulses; red solid line = Do(t) and blue dashed line = T i(t).
positive value of T i(0). We can further see that the deviation of Do(t) in Fig-
ure 3.1(b) is greater than the deviation of T i(t) in Figure 3.1(a): a positive value
of T i(t) exerts a greater push on Do(t) than vice versa, because of the greater
cross-effect of T i(t) on Do(t). In this case this strong cross-effect, combined with
the relatively weaker auto-effect of Do(t), results in Do(t) taking on a higher value
than T i(t) at around t = 1, one hour after the initial impulse. Substantively, when
our participant is feeling physically tired at a particular moment (Figure 3.1(b)),
he begins to feel a down over the next couple of hours, before eventually these
feelings return to normal (again in this case, around 4 hours later).
Figures 3.1(c) further demonstrates the role of the negative auto-effects and
positive cross-effects in different scenarios. In Figure 3.1(c), both processes take
on positive values at t = 0; the positive cross-effects result in both processes re-
turning to equilibrium at a slower rate than in Figures 3.1(a) and 3.1(b). In sub-
stantive terms this means that, when the participant is feeling very down, and
very tired, it takes longer for the participant to return to feeling normal. Here
56
3.4. Making Sense of CT Models
also the stronger auto-effect of T i(t) than Do(t) is evident: although both pro-
cesses start at the same value, an hour later T i(1) is much closer to zero than
Do(1), that is, T i(t) decays faster to equilibrium than Do(t). In substantive terms,
this tells us that when the participant is feeling down and physically tired, he re-
covers much quicker from the physical tiredness than he does from feeling down.
In Figure 3.1(d), we see that Do(0) and T i(0) taking on values of opposite signs
results in a speeding-up of the rate at which each variable decays to equilibrium.
The auto-effect of Do(t) is negative, which is added to by the positive cross-effect
of T i(t) multiplied by the negative value of T i(0). This means that Do(0) in Fig-
ure 3.1(d) takes on a stronger negative velocity, in comparison to Figures 3.1(a)
or 3.1(c). A positive value for Do(0) has a corresponding effect of making T i(0)
take on an even stronger positive velocity. Substantively, this means that when
the participant feels down, but feels less tired (i.e. more energetic) than usual,
both of these feelings wear off and return to normal quicker than in the other
scenarios we examined. The stronger auto-effect of T i(t), in combination with
the positive cross-effect of Do(t) on T i(t), actually results in T i(t) shooting past
the equilibrium position in the T i(t) dimension (T i(t) = 0) and taking on positive
values around t = 1.5, before the system as a whole returns to equilibrium. Sub-
stantively, when the participant is feeling initially down but quite energetic, we
expect that he feels a little bit more tired than usual about an hour and half later,
before both feelings return to normal.
57
3. A Continuous-Time Approach to Intensive Longitudinal Data
1.0
0.5 b) c)
Ti(t)
0.0
a)
−0.5
−1.0
d)
Do(t)
Figure 3.2: Vector field for Do(t) and T i(t), including blue and red nullclines.
position, taking one small step forward in time would result in a change in the
system’s position along the T i(t) axis (i.e., a change in the value of Tired), but
not along the Do(t) axis (that is, dDo(t)/dt = 0 and dT i(t)/dt , 0). The converse
is true for a horizontal arrow (that is, dDo(t)/dt , 0 and dT i(t)/dt = 0). The
two lines in Figure 3.2, blue and red, identify at which positions dDo(t)/dt = 0
and dT i(t)/dt = 0, respectively; these are often referred to as nullclines or equiv-
alently, solution lines. If the nullclines are not perfectly perpendicular to one
another, this is due to the presence of at least one cross-effect. The point at which
these nullclines cross represents the equilibrium position in this two-dimensional
space, here located at Do(t) = 0, T i(t) = 0. The crossing of these nullclines splits
the vector field in four quadrants, each of which is characterized by a different
combination of negative and positive values for dDo(t)/dt and dT i(t)/dt. The top
left and bottom right quadrants represent areas in which the derivatives are of
opposite sign, dDo(t)/dt > 0 & dT i(t)/dt < 0 and dDo(t)/dt < 0 & dT i(t)/dt > 0,
respectively. The top right and bottom left quadrants represent areas where the
derivatives are of the same sign, dDo(t)/dt < 0 & dT i(t) < 0 and dDo(t)/dt > 0 &
dT i(t) > 0, respectively.
By tracing a path through the arrows, we can see the trajectory of the sys-
tem of interest from any point in the possible space of values. In Figure 3.2, we
58
3.4. Making Sense of CT Models
include the same four bi-variate trajectories as we examined with the IRFs. In-
stead of the IRF representation of two variables whose values are changing, the
vector field represents this as the movement of one process in a two-dimensional
space. For instance, the trajectory starting at Do(t) = 0 and T i(t) = 1 begins in
the top-left quadrant, where dDo(t)/dt is positive and dT i(t)/dt is negative; this
implies that the value of Down will increase, and the value of Tired will decrease
(as can be seen in Figure 3.1(b)). Instead of moving directly to the equilibrium
along the T i(t) dimension, the system moves away from equilibrium along the
Do(t) dimension, due to the cross-effect of T i(t) on Do(t), until it moves into
the top-right quadrant. In this quadrant, dDo(t)/dt and dT i(t)/dt are both nega-
tive; once in this quadrant the process moves towards equilibrium, tangent to the
dDo(t)/dt nullcline. The other trajectories in Figure 3.2 analogously describe the
same trajectories as in Figure 3.1(a), 3.1(c) and 3.1(d).
In general, the trajectories in this vector field first decay quickest along the
T i(t) dimension, and slowest along the Do(t) dimension. This can be clearly seen
in trajectories b), c), and d). Each of these trajectories first change steeply in the
dDo(t)
T i(t) dimension, before moving to equilibrium at a tangent to the red ( dt )
nullcline. This general property of the bi-dimensional system is again related to
the much stronger auto-effect of T i(t), and the relatively small cross-effects. In a
technical sense we can say that that Do(t) represents the ‘slowest eigen-direction’
(see Strogatz, 2015, Chapter 5).
For this given time-interval, the cross-lagged effect of Down on Tired (φ21 (∆t =
1) = 0.077) is smaller than the cross-lagged effect of Tired on Down (φ12 (∆t = 1) =
0.117). However, as shown in Equation (3.5) the value of each of these lagged ef-
fects changes in a non-linear way depending on the time-interval chosen. To
visualize this, we can calculate Φ(∆t) for a range of ∆t, and represent this infor-
mation graphically in a lagged parameter plot, as in Figure 3.3.
From Figure 3.3, we can see that both cross-lagged effects reach their maxi-
mum (and have their maximum difference) at a time-interval of ∆t = 0.65; fur-
thermore, we can see that the greater cross-effect (a12 ) results in a stronger cross-
lagged effect φ12 (∆t) for a range of ∆t. Moreover, we can visually inspect how
the size of each of the effects of interest, as well as the difference between these
effects, varies according to the time-interval. From a substantive viewpoint, we
could say that the effect of feeling physically tired has the strongest effect on
feelings of being down around 40 minutes later.
59
3. A Continuous-Time Approach to Intensive Longitudinal Data
1.0
φ11(∆t)
φ22(∆t)
φ12(∆t)
φ21(∆t)
0.8
0.6
Φ(∆t)
0.4
0.2
0.0
0 1 2 3 4
Time−Interval ∆t
Figure 3.3: The elements of Φ(∆t) for the bivariate example (i.e., φ11 (∆t), φ12 (∆t), φ21 (∆t), φ22 (∆t))
plotted for a range of values for ∆t.
While the shape of the lagged parameters may appear similar to the shapes
of the trajectories plotted in the IRFs, lagged parameter plots and IRFs represent
substantively different information. IRFs plot the positions of each variable in
the system as they change over time, given some impulse (y(t) vs t given some
y(0)). In contrast, lagged parameter plots show how the lagged relationships
change depending on the length of the time-interval between them, independent
of impulse values (eA∆t vs ∆t). The lagged relationships can be thought of as the
components which go into determining any specific trajectories.
60
3.5. Discussion
interpreting regions in which there is little or no observed data (cf. Boker &
McArdle, 1995). The same logic applies for the interpretation of IRFs for im-
pulses that do not match observed values. Moreover, we should also be aware that
interpreting lagged parameter plots for time-intervals much shorter than those
we observe data at is a form of extrapolation: It relies on strong model-based
assumptions, such as ruling out the possibility of a high-frequency higher-order
process (Voelkle & Oud, 2013; Voelkle et al., 2012).
3.5 Discussion
In this chapter we have set out to clarify the connection between DT- and CT-
VAR(1) models, and how we can interpret and represent the results from these
models. So far we have focused on single-subject, two-dimensional, first-order
systems with a stable node equilibrium. However, there are many ways in which
these models can be extended, to match more complicated data and/or dynamic
behavior. Below we consider three such extensions: a) systems with more than
two dimensions (i.e., variables); b) different types of fixed points resulting from
non-real eigenvalues of the drift matrix; and c) moving from single-subject to
multi-level datasets.
61
3. A Continuous-Time Approach to Intensive Longitudinal Data
with more than two variables also becomes less straightforward. For example,
Deboeck and Preacher (2016), Aalen et al. (2012, 2016) and Aalen et al. (2018) ar-
gue that the interpretation of Φ(∆t) parameters in mediation models (with three
variables and a triangular A matrix) as direct effects may be misleading: Deboeck
and Preacher argue that instead they should be interpreted as total effects. This
has major consequences for the practice of DT analyses and the interpretation of
its results.
which is equivalent to our empirical example above, but with the value of a21
altered from 0.375 to −2.000. The eigenvalues of this matrix are λ1 = −1.706 +
0.800i and λ2 = −1.706 − 0.800i. In contrast to our empirical example, we can
see that the trajectories follow a spiral pattern; the trajectory which starts at
y1 (t) = 1, y2 (t) = 1 actually overshoots the equilibrium in the T i(t) dimension
before spiraling back once in the bottom quadrant. There are numerous ex-
amples of psychological systems that are modeled as damped linear oscillators
using second-order differential equations, which include the first- and second-
order derivatives (cf. Boker & Nesselroade, 2002; Bisconti et al., 2004; Boker,
Montpetit, et al., 2010; Horn, Strachan, & Turkheimer, 2015). However, as shown
here, such behavior may also result from a first-order model.
Stable nodes and spirals can be considered the two major types of stable fixed
points, as they occur whenever the real part of the eigenvalues of A are negative,
that is α < 0. Many other types of stable fixed points can be considered as spe-
cial cases: when we have real, negative eigenvalues that are exactly equal, the
62
3.5. Discussion
b) c)
1.0
0.5
y2(t)
0.0
a)
−0.5
−1.0
d)
y1(t)
Figure 3.4: Vector field for a stable spiral corresponding to a drift matrix with negative real part
complex eigenvalues.
fixed point is called a stable star node (if the eigenvectors are distinct), or a stable
degenerate node (if the eigenvectors are not distinct). In contrast, if the real-part
of the eigenvalues of A are positive then the system is unstable, also referred to
as non-stationary or a unit-root in the time series literature (Hamilton, 1994).
This implies that, given a deviation, the system will not return to equilibrium; in
contrast to stable systems, in which trajectories are attracted to the fixed point,
the trajectories of unstable systems are repelled by the fixed point. As such we
can also encounter unstable nodes, spirals, star nodes and degenerate nodes. The
estimation and interpretation of unstable systems in psychology may be fruitful
ground for further research.
Two further types of fixed points may be of interest to researchers; in the
special case where the eigenvalues of A have an imaginary part and no real part
(α = 0), the fixed point is called a center. In a system with a center fixed point,
trajectories spiral around the fixed point without ever reaching it. Such systems
exhibit oscillating behavior, but without any damping of oscillations; certain bi-
ological systems, such as the circadian rhythm, can be modeled as a dynamic
system with a center fixed point. Such systems are on the borderline between
stable and unstable systems, sometimes referred to as neutrally stable; trajectories
63
3. A Continuous-Time Approach to Intensive Longitudinal Data
are neither attracted to or repelled by the fixed point. Finally, a saddle point oc-
curs when the eigenvalues of A are real but of opposite sign (one negative, one
positive). Saddle points have one stable and one unstable component; only trajec-
tories which start exactly on the stable axis return to equilibrium, and all others
do not. Together spirals, nodes and saddle points cover the majority of the space
of possible eigenvalues for A. Strogatz (2015) describes the different dynamic be-
havior generated by different combinations of eigenvalues of A in greater detail.
64
3.5. Discussion
3.5.4 Conclusion
There is no doubt that the development of dynamical systems modeling in the
field of psychology has been hampered by the difficulty in obtaining suitable
data to model such systems. However this is a barrier that recent advances in
technology will shatter in the coming years. Along with this new source of psy-
chological data, new psychological theories are beginning to emerge, based on
the notion of psychological processes as dynamic systems. Although the statis-
tical models needed to investigate these theories may seem exotic or difficult to
interpret at first, they reflect the simple intuitive and empirical notions we have
about psychological processes: Human behavior, emotion and cognition fluctuate
continuously over time, and the models we use should reflect that. We hope that
our treatment of CT-VAR(1) models and their interpretation will help researchers
to overcome the knowledge-barrier to this approach, and can serve as a stepping
stone towards a broader adaptation of the CT dynamical system approach to psy-
chology.
65
3. A Continuous-Time Approach to Intensive Longitudinal Data
A = V DV −1
66
3.B. Empirical Example Data Analysis
Multiplying V eD V −1 gives us
b(−eλ1 ∆t +eλ2 ∆t )
R−a+d eλ1 ∆t + R+a−d eλ2 ∆t
A∆t
e = 2R c(−eλ1 ∆t +eλ2R2 ∆t )
R (3.13)
R+a−d λ1 ∆t R−a+d λ2 ∆t
R 2R e + 2R e
Note that we present here only a worked out example for a 2 × 2 square matrix.
For larger square matrices (representing models with more variables), the eigen-
value decomposition remains the same although the terms for the eigenvalues,
eigenvectors and determinants become much less feasible to present.
#Load Data#
rawdata <- read.csv("ESMdata.csv",
header=TRUE,
stringsAsFactors = FALSE)
67
3. A Continuous-Time Approach to Intensive Longitudinal Data
# Create an ID variable
data$id = rep(1,dim(data)[1])
## Data analysis
68
3.B. Empirical Example Data Analysis
meanIntervals = F,
plotOptimization = F,
nofit = F, discreteTime = F,
verbose = 0)
summary(fit)
69
Chapter 4
Time to Intervene: A
Continuous-Time Approach to
Network Analysis and
Centrality
Abstract
This chapter has been adapted from: Ryan, O. & Hamaker, E. L. (under review). Time to Inter-
vene: A Continuous-Time Approach to Network Analysis and Centrality. Author contributions: OR
and ELH conceptualized the initial project. OR wrote the paper and R code. ELH helped further
develop the ideas in the project, discussed progress and provided textual feedback.
71
4. A Continuous-Time Approach to Network Analysis and Centrality
4.1 Introduction
Dynamical network analysis, based on lagged regression models, has become a
popular approach for the analysis of experience sampling data in psychology
(Bringmann et al., 2013; Borsboom & Cramer, 2013). In clinical psychology in
particular, such analyses have been promoted as an aid in developing personal-
ized treatments for psychopathology. To facilitate this, centrality measures cal-
culated from parameter estimates are often used to identify which variable in
the network to target for an intervention (Bringmann et al., 2013; A. J. Fisher &
Boswell, 2016; Kroeze et al., 2017; Epskamp, van Borkulo, et al., 2018; Rubel,
Fisher, Husen, & Lutz, 2018; Bak, Drukker, Hasmi, & van Os, 2016; Bringmann
et al., 2015; Bastiaansen et al., 2019; A. J. Fisher, Reeves, Lawyer, Medaglia, &
Rubel, 2017; Christian et al., 2019).
Two developments in the methodological literature highlight problems with
this practice. First, the estimation of dynamical network structures is based on
path estimates from a discrete-time (DT) first-order Vector Auto-regressive (VAR)
model. The DT-VAR model suffers from the problem of time-interval dependency,
meaning that the parameters of this model can potentially lead to dramatically
different conclusions based on how the observations are spaced in time (Gollob &
Reichardt, 1987; Kuiper & Ryan, 2018). This is an issue which numerous authors
have argued could be resolved by modeling psychological processes as unfold-
ing continuously over time using continuous-time (CT) models (Boker, 2002; van
Montfort, Oud, & Voelkle, 2018; Ryan, Kuiper, & Hamaker, 2018). Second, the
use of out-of-the-box centrality measures for the identification of intervention
targets has been critiqued from various sides, both in terms of their applicability
to psychological networks (Bringmann et al., 2019) and in their actual ability to
detect optimal intervention targets (Borgatti, 2005; van Elteren & Quax, 2019;
Dablander & Hinne, 2019).
Although the consequences of taking a CT approach have been discussed in
general elsewhere (cf. Boker, 2002; Voelkle et al., 2012; Aalen et al., 2016, 2012;
Deboeck & Preacher, 2016), the specific implications for dynamical network
analysis have not yet been investigated. Taking a CT approach yields practical
benefits, such as an ability to deal with unequally spaced measurements, but also
leads to a new outlook on the meaning and interpretation of lagged regression
parameters. As a consequence, taking a CT perspective not only undermines the
typical interpretation of DT-VAR networks as well as their accompanying central-
ity measures, but it also creates opportunities for the development of new ways
to understand the underlying dynamic process and how best to intervene on it.
In this paper we introduce a CT approach to dynamical network analysis,
and develop new centrality measures which can be used to gain more insight
into what variables should be targeted for interventions, and what kinds of in-
terventions can be expected to lead to what kinds of outcome. In so doing we
tie together diverse strands of the psychological and methodological literature.
Specifically, we show how centrality measures are related to path-specific effects
from the SEM literature, which allows us to connect previous research on path-
specific effects in CT models (Aalen et al., 2012; Deboeck & Preacher, 2016) and
72
4.2. Current Practice: DT-VAR Networks
Yτ = c + ΦYτ−1 + τ (4.1)
73
4. A Continuous-Time Approach to Network Analysis and Centrality
í í
... Y1,ɒ Y1,ɒͳ Y1,ɒʹ
...
Y Y
í
... Y2,ɒ Y2,ɒͳ Y2,ɒʹ
... í
í í
í
Y4 Y
... Y4,ɒ Y4,ɒͳ Y4,ɒʹ ...
í
í
... Y2(t0) Y2(1ï») Y2(t1) Y2(2ï») Y2(t2)
...
í
... Y3(t0) Y3(1ï») Y3(t1) Y3(2ï») Y3(t2)
...
í
Y Y
... Y4(t0) Y4(1ï») Y4(t1) Y4(2ï») Y4(t2)
...
í
(b) DT-VAR model with latent waves and dynamical network representing Φ( 12 )
Figure 4.1: Path-model (left-hand side) and network (right-hand side) representations of two four-
variable DT-VAR models. In the path models, the presence of an arrow linking two variables denotes
some non-zero dependency between them, conditional on all variables at the previous wave. For the
networks solid arrows denote positive parameters and dashed arrows represent negative parameters.
which entails that the means, variance and covariances, and lagged regression
parameters remain the same over time.
The DT-VAR model describes a relatively simple dynamic system. Each vari-
able has some baseline level or equilibrium position defined by its mean value.
Typically, the variables themselves are centered around this mean value (e.g. As-
parouhov, Hamaker, & Muthén, 2018), and so the intercept term is often omitted
for notational simplicity (c = 0), a convention we will adopt throughout the re-
mainder of the paper. The random shocks τ push the system away from equi-
librium, and the lagged parameters Φ determine how the variables react to these
shocks, eventually returning to equilibrium over time.
The DT-VAR model can be represented as either a path-model, as shown in the
left-hand panel of Figure 4.1(a), or as a dynamical network structure, as shown
in the right-hand panel, where the nodes represent the random variables, and the
edges represent the values of the lagged parameters Φ (Bringmann et al., 2013;
Epskamp, van Borkulo, et al., 2018). The DT-VAR model, and its multi-level
extension, are popular for the analysis of experience sampling data in clinical
psychology (Bringmann et al., 2013; Pe et al., 2015; A. J. Fisher & Boswell, 2016;
74
4.2. Current Practice: DT-VAR Networks
Kroeze et al., 2017; Rubel et al., 2018; Bak et al., 2016).1 In dynamical network
approaches, each variable is typically used to represent some psychological con-
struct (such as an emotion or symptom of disorder) that varies over time, which
we refer to as a process. The lagged parameters in Φ are typically interpreted as
direct effects of these processes on each other over time.
For example, take it that the four variables in Figure 4.1(a) represent (re-
peated measurements of) Stress (Y1 ), Anxiety (Y2 ), feeling Self-Conscious (Y3 ) and
feelings of Physical Discomfort (Y4 ). We will refer to this throughout as the Stress-
Discomfort system. We can see from the parameter values in the dynamical net-
work that all variables share reciprocal cross-lagged relationships with all other
variables: The network is fully connected, with a mix of positive and negative
cross-lagged and auto-regressive parameters of different sizes. Typically, a cross-
lagged parameter such as φ41 = 0.05 would be interpreted as the direct effect of
current Stress (Y1,τ ) on Physical Discomfort at the next measurement occasion
(Y4,τ+1 ), conditional on (i.e., controlling for) current feelings of Anxiety, Self-
Conscious, and Physical Discomfort (Y2,τ , Y3,τ , Y4,τ ). This parameter is weakly
positive, leading to the interpretation that a high level of current Stress has a
small positive direct effect on feelings of Physical Discomfort at the next occa-
sion.
75
4. A Continuous-Time Approach to Network Analysis and Centrality
76
4.2. Current Practice: DT-VAR Networks
77
4. A Continuous-Time Approach to Network Analysis and Centrality
teren and Quax (2019) and Dablander and Hinne (2019), who show that central
nodes are often not the optimal target for an intervention, as well as Borgatti
(2005) and more recently Bringmann et al. (2019), who detail at length how the
utility of common centrality measures is highly dependent on the particular type
of network under consideration.
To solve this problem, we will make use of the interventionist causal inference
framework (Hernan & Robins, 2019; Pearl, 2009; Rubin, 1974). In this frame-
work, a causal relationship between X and Y is defined as the change we would
expect to see in Y if we were to intervene and change the value of X. The key
appeals of this approach are first, that it places an emphasis on an intervention-
effect as the target of inference, and second that it allows one to derive the con-
ditions under which the effects of such an intervention can be identified from
relationships between variables in observational data. Crucially, intervention-
based definitions have also been extended to path-specific effects: VanderWeele
and Tchetgen Tchetgen (2017) have shown that—under certain circumstances—
path-tracing total, direct and indirect effects in a lagged regression model (similar
to the DT-VAR model discussed above) can be interpreted as the expected change
in Y , given interventions to change the values of the cause variable X and the
mediating variable(s) M. Although, to our knowledge, intervention-based defi-
nitions have not been extended to centrality measures, the connection between
centrality and path-specific effects established above opens up a potential avenue
by which such a definition could be established. In Section 4.4.1 we will make
use of these connections to establish new centrality measures which are inspired
by this framework, ensuring an unambiguous connection between the value of
those measures and the choice of intervention target.
Of course, the identification of these intervention effects in practice will still
rely on a number of idealized assumptions. We must assume that there are no
unmeasured confounder variables, that we can intervene in the system without
changing how the variables in the system relate to one another (an assumption
known as modularity; Pearl, 2009), and that the first-order time series model we
fit to the data correctly characterizes the process. Assumptions regarding con-
founders and modularity are fundamental to the identification of intervention-
effects from data in any context (e.g. from experiments, cross-sectional data or
time-series). However, even if we are willing to make these strong and optimistic
assumptions, there is still a fundamental issue specific to lagged regression mod-
els which we must account for in order to come to sensible conclusions about
intervention targets. That fundamental issue is the time-interval problem.
78
4.2. Current Practice: DT-VAR Networks
problem.2 This problem arises when the processes under investigation vary at
a higher frequency than we observe them, which means that the processes take
on values at points in time in-between those particular occasions at which we
measure the process. In path-modeling terms, this means that there are unmea-
sured and un-modeled values of the observed variables Y in-between measure-
ment occasions, depicted as latent variables in Figure 4.1(b). The time-interval
problem has two consequences. First, the value of the regression parameters Φ
are a function of the time-interval between observations (denoted Φ(∆t)), and
so, except under restrictive circumstances, may change their sign, size and rela-
tive ordering if a different time-interval is used (Kuiper & Ryan, 2018). Second,
the lagged regression parameters at any interval should be interpreted as total,
rather than direct effects from a path-tracing perspective. The reason for this is
that cross-lagged parameters constitute both direct pathways and indirect path-
ways through latent values of the other processes in the model (Cole & Maxwell,
2003; Reichardt, 2011; Deboeck & Preacher, 2016).
To illustrate both issues, let us take it that the parameters that we introduced
in Figure 4.1(a) represent the lagged relationships at one-hour intervals; we de-
note these parameters as Φ(∆t = 1). In theory, we could also have obtained ob-
servations of the Stress-Discomfort system at half-hour intervals: The path model
representing this shorter-interval system is depicted in the left panel of Figure
4.1(b), where the potential half-hour measurements are depicted as latent vari-
ables Y (t = 1 12 ) and Y (t = 2 21 ). The effects matrix relating the half hour realiza-
tions of the process is denoted Φ(∆t = 12 ). A well-known result from the time-
series literature allows us to relate the parameters of these two models through
the matrix exponent (Hamilton, 1994). In this case, the matrix of lagged param-
eters of the half-hour system is related to the matrix of lagged parameters at the
longer one-hour time-interval by the expression
Φ( 12 )2 = Φ(1) (4.2)
that is, by squaring the matrix of parameters at the shorter interval, we obtain
the parameters at twice that interval. It is important to note here that squaring
a matrix is not equivalent to squaring the parameters of that matrix: Instead,
any given parameter in Φ(1) is a function of multiple parameters in Φ( 12 ). For
instance, the cross-lagged parameter which regresses Y4,τ+1 on Y1,τ can be re-
written in terms of the shorter-interval parameters as φ41 (1) = φ11 ( 12 )φ41 ( 12 ) +
φ21 ( 12 )φ42 ( 12 )+φ31 ( 12 )φ43 ( 12 )+φ41 ( 12 )φ44 ( 12 ). Clearly, this has major consequences
for the interpretation of lagged regression parameters.
First, we can see that the types of conclusions we would intuitively make
based on the parameter values in the half-hour network are entirely different
than those we made based on the parameters in the one-hour network. To be-
2 This is also sometimes referred to as the lag problem (e.g., Hamaker & Wichers, 2017). The lag of
a model also describes a notion of time-spacing, but more generally refers to the order of a lagged re-
gression model: Lag-one models relate current values to the exactly preceding measurement occasion,
and lag-two models relate current values to the variables two measurement occasions previously, and
so forth. Thus, the lag of a model and is distinct from the time-interval between observations, and so
for clarity we refer to this as the time-interval problem throughout.
79
4. A Continuous-Time Approach to Network Analysis and Centrality
gin with, the magnitude of each individual parameter, as well as their signs and
relative ordering, are different in the two networks. For example, in the one-
hour network, Stress and Anxiety both have positive lagged relationships with
Physical Discomfort (φ41 (1) = 0.047 and φ42 (1) = 0.077), with the effect of Anx-
iety being slightly larger. In the half-hour network, the corresponding lagged
relationships are both strongly negative, with the effect of Stress now the larger
(φ41 ( 12 ) = −0.275 and φ42 ( 12 ) = −0.151). If we calculate centrality measures for the
half-hour network, we may conclude that either Physical Discomfort is a promis-
ing intervention target, as it has the highest scores on the total and direct influ-
(2) (1)
ence measures (EI4 = 0.555 and EI4 = 0.557), or that Anxiety is a promising
target as it has the highest Betweenness Centrality (BC2 = 3; See Appendix 4.B
for a full table of centrality values).
Second, on a more fundamental level, the presence of the time-interval prob-
lem alters how we should interpret a set of lagged regression parameters. Specif-
ically, based on the relationship in Equation (4.2), the lagged parameters of the
one-hour network Φ(1) should be interpreted as total rather than direct effects
(Deboeck & Preacher, 2016; Aalen et al., 2016). Indeed, taking the power of a
matrix of direct effects, as in Equation (4.2), is suggested by Bollen (1987) exactly
as a method of calculating total effects in the SEM literature. This interpretation
also helps clarify how we come to seemingly contradictory conclusions based
on the time-interval used. For example, the parameter φ42 (1) = 0.077 would
typically be interpreted as a direct effect of Anxiety now (Y2,τ ) on Physical Dis-
comfort an hour from now (Y4,τ+1 ). However, when we examine how these vari-
ables are related to one another in Figure 4.1(b), we can see that the relationship
between them is made up of a number of different pathways, including direct
paths (Y2,τ → Y2 (1 12 ) → Y4,τ+1 and Y2,τ → Y4 (1 12 ) → Y4,τ+1 ) as well as paths that
pass through latent values of Stress (Y2,τ → Y1 (1 12 ) → Y4,τ+1 ) and Self-Conscious
(Y2,τ → Y3 (1 12 ) → Y4,τ+1 ). In this instance, a combination of strong positive indi-
rect paths and slightly weaker negative direct paths add up to the weak positive
total effect in the one-hour network: A high feeling of Anxiety has the effect of
lowering Physical Discomfort a half hour later, but also sets off a chain reaction of
Stress and Self-Conscious feelings that ultimately leads to an increase in Physical
Discomfort an hour later.
In summary, the time-interval problem is present whenever there are latent,
unobserved values of our processes of interest in-between measurement occa-
sions. The consequences of the time-interval problem for the identification of in-
tervention targets, using current practice, is twofold. First, any parameter, path-
specific effect, or centrality measure may lead to conclusions which are specific
to the time-interval used in data collection: The effect of X on Y should in fact
be interpreted as the effect at one specific time-interval (Gollob & Reichardt, 1987).
Second, the intuitive interpretation given to path-specific effects and centrality
measures based on a path-tracing logic are misleading: Direct effects are better
conceptualized as total effects (Deboeck & Preacher, 2016; Reichardt, 2011). This
leaves the usual path-tracing based notion of “direct”, “indirect” and “total” ef-
fects, and therefore also of centrality measures which summarize these effects,
80
4.3. CT Network Analysis: Accounting for Continuity
open to question. However, our presentation here also highlighted one poten-
tial solution to the time-interval problem: Decomposing lagged relationships be-
tween observations into truly direct effects operating over a shorter time-interval.
This decomposition opens up a new perspective on how lagged relationships
should be interpreted, a perspective which we can use to explore time-interval
dependency, and avoid coming to misleading or contradictory choices regarding
intervention targets.
4.2.4 Conclusion
We have now identified two problems with current DT-VAR network approaches,
in particular with the use of these models and associated centrality measures for
the identification of intervention targets. The first is that the link from inter-
ventions to model parameters and centrality measures is unclear. The second
problem is that the parameters and centrality measures themselves suffer from
time-interval dependency. In the next section we will present one potential solu-
tion to the latter problem: the use of continuous-time models. We will then use
this CT approach in the fourth section to address the other problem, describing
how CT models can in principle be used to identify intervention targets.
81
4. A Continuous-Time Approach to Network Analysis and Centrality
82
4.3. CT Network Analysis: Accounting for Continuity
í í
Y1(t0) Y1(t)Y1(t)Y1(s) Y1(sn) Y1(t1) Y1(t)Y1(t)Y1(s) Y1(sn) Y1(t2)
... ... ... ... 1.25
Y1 Y2
5.5
Y4 2.5 Y
Y4(t0) Y4(t)Y4(t)Y4(s) Y4(sn) Y4(t1) Y4(t)Y4(t)Y4(s) Y4(sn) Y4(t2)
... ... ... ...
í
í
Figure 4.2: CT-VAR path-model (left-hand side) and CT network in the form of a weighted local
independence graph representing A (right-hand side). In the path model, the latent variables and
ellipses represent an infinite number of latent unobserved process values in-between measurement
occasions, spaced an infinitesimally small time-interval apart; the presence of an arrow linking two
variables in the path model denotes some non-zero dependency between them, conditional on all
variables at the previous “wave”, that is, a local dependency. For the network, solid arrows denote
positive parameters and dashed arrows represent negative parameters.
change of the other, instead of the value of the other at the next occasion. This
type of network representation is known as a local dependence graph (Schweder,
1970; Aalen, Borgan, Keiding, & Thormann, 1980; Didelez, 2008), of which the
right-hand side of Figure 4.2 is a weighted variant. We will refer to this as a CT
network, where the edges depict direct moment-to-moment dependencies between
the time-varying processes.
The interpretation of drift matrix parameters is similar to that of a change-
score model from the time-series literature. The diagonal parameters are known
as auto-effects and are typically negative: If Stress takes on a high positive value,
its negative auto-effect (a11 = −6) ensures it will move back towards equilibrium
in response. The higher the absolute value of the auto-effect, the quicker the pro-
cess will move back towards equilibrium (ignoring all other parameters), similar
to an auto-regressive parameter close to zero. As such, the diagonal parameters
are also referred to as reflecting the “centralizing tendency” of a variable, for in-
stance by Oravecz et al. (2011). Off-diagonal elements known as cross-effects have
a similar interpretation to cross-lagged parameters: The negative cross-effect of
Anxiety on the rate of change of Physical Discomfort (a42 = −7.3) means that if
Anxiety takes on a positive value, this will result in a decrease in the value of
Physical Discomfort. The higher the absolute value of the parameter, the greater
the magnitude of the effect (for more details on the interpretation of these pa-
rameters, see also Ryan et al., 2018; Voelkle et al., 2012; Oravecz et al., 2011).
From the CT network, we can see that all of the Stress-Discomfort variables
have a negative auto-effect on themselves (a11 = a33 = a44 = −6); the auto-effect of
Anxiety is closer to zero, which roughly corresponds to a stronger auto-regressive
effect than the other variables (a22 = −2.5). Notably there are much fewer direct
dependencies (i.e. fewer connections in the network) between processes on a
moment-to-moment basis than we saw in the lagged DT-VAR networks in Fig-
ures 4.1(a) and (b): For instance, Stress has no direct moment-to-moment ef-
83
4. A Continuous-Time Approach to Network Analysis and Centrality
fect on Physical Discomfort (a41 = 0). The dependencies which are present are
mostly positive: Stress and Anxiety exert reciprocal positive effects on one an-
other (a12 = 1.25 and a21 = 5.5); Anxiety has a direct positive effect on the rate of
change of Self-Conscious; Anxiety has a strong direct negative effect (a42 = −7.3),
and Self-Conscious a positive direct effect (a43 = 2.5) on the rate of change of
Physical Discomfort; finally, Physical Discomfort has a positive direct effect on
the rate of change of Stress (a14 = 5). We will see why there is a discrepancy be-
tween the moment-to-moment CT network and the lagged relationships in the
DT-VAR network by connecting these two models through the integral form of
the differential equation.
where variables at the current measurement occasion Y (tτ ) are regressed on vari-
ables at the previous measurement occasion Y (tτ−1 ). Note that τ refers to the
measurement occasion, whereas t refers to the actual time when this measure-
ment took place. Hence, ∆tτ indicates the time interval between two consecutive
measurement occasions, which may differ across pairs of observations. The cur-
rent expression of the CT-VAR model is very similar to the DT-VAR model that
was presented in Equation (4.1). The c term represents a vector of intercepts,
and (∆tτ ) represents the residual vector, normally distributed with mean zero
and variance-covariance matrix that is also a function of the time-interval (for
more details see Oud & Jansen, 2000; Voelkle et al., 2012; Voelkle & Oud, 2013).
In place of the Φ matrix in Equation (4.1), these lagged variables are related
by eA∆tτ , the matrix exponential of the drift matrix A, multiplied by the time-
interval between those measurement occasions.
It follows that the CT- and DT-VAR lagged regression matrices are related to
each other by the expression
84
4.3. CT Network Analysis: Accounting for Continuity
which states that the lagged parameters for any particular time-interval Φ(∆t)
can be found by taking the matrix exponential of the moment-to-moment drift
matrix A to the power of the length of that time-interval ∆t (cf. Oud & Jansen,
2000; Voelkle et al., 2012). Notice the similarity between this relationship, and
the expression used to relate the half-hour and one-hour parameter matrices in
Equation (4.2): To find the lagged relationships at a longer time-interval, we
again take the appropriate matrix exponent of the lagged relationships at the
shorter interval.
Since there is a clear connection between the CT- and DT-VAR models, we
can now consider the two consequences of the time-interval problem from a CT
perspective. First, the conclusions we would intuitively make based on a DT-
VAR model using any one particular interval may differ from those we would
make at any other interval. Figure 4.3 depicts the lagged relationships of the
Stress-Discomfort system over a range of time-intervals, from zero to two hours,
showing clearly that observing this process at say, 15 minute intervals (∆t = 0.25)
instead of half-hour or one-hour intervals leads to a quantitatively and qualita-
tively different set of conclusions about the underlying process. In Figure 4.4 we
show how the DT network centrality measures also change according to the time-
interval, leading to different conclusions regarding optimal intervention targets
when applying current standard practice (an effect which is particularly pro-
nounced for the betweenness measure in panel (c)).
Second, the CT perspective shows that lagged regression parameters Φ(∆t)
at any interval should be interpreted as total rather than direct effects from a
path-tracing perspective (Aalen et al., 2016; Deboeck & Preacher, 2016). This
follows directly from the observation that the matrix exponential relationship in
Equation (4.5) is simply a generalization of the path-tracing operation described
above when relating the half-hour and one-hour networks (Equation (4.2)): The
lagged relationships at a very short time-interval are described by (a function of)
the A matrix, and to find the relationships between variables spaced further (i.e.,
∆t) apart, we apply a path-tracing operation through the latent values of Y (t) in-
between those occasions. See Appendix 4.C for an accessible derivation of this
term as a path-tracing operation.
This interpretation also allows us to explain the changing patterns of lagged
relationships we observe in Figure 4.3. For example, we know that Stress has
no direct moment-to-moment effect on the rate of change of Physical Discom-
fort (a41 = 0), but we see that the corresponding lagged relationship in the DT
model (φ41 (∆t) in Figure 4.3(c)) is strongly negative at short intervals and weakly
positive at longer intervals: This is because it is a total effect, made up of one
negative indirect pathway (through Anxiety) and one positive indirect pathway
(through Anxiety and Self-Conscious). In order to calculate lagged direct and in-
direct effects from the CT perspective, we can use path-tracing methods defined
by Deboeck and Preacher (2016) and Aalen et al. (2016) and described in detail in
Appendix 4.D. The direct lagged effect of Stress on Physical Discomfort is found
by first omitting the indirect pathways from A (in this case, a31 , a42 , a32 and a43 )
before applying the matrix exponential function in Equation (4.5). In so doing we
trace a path from Stress to Physical Discomfort which does not pass through any
85
4. A Continuous-Time Approach to Network Analysis and Centrality
1.0
Parameter Value
0.5
φ11(∆t)
φ22(∆t)
φ33(∆t)
φ44(∆t)
0.0
−0.5
φ12(∆t)
0.5
φ14(∆t)
φ21(∆t)
φ32(∆t)
φ42(∆t)
0.0
φ43(∆t)
−0.5
φ13(∆t)
0.5
φ23(∆t)
φ24(∆t)
φ31(∆t)
φ34(∆t)
0.0
φ41(∆t)
−0.5
Figure 4.3: Lagged regression parameters as a function of the time-interval for the Stress-Discomfort
system. Black dotted lines indicate the parameter values of the half-hour and one-hour networks in
Figure 4.1(a) and (b).
86
4.3. CT Network Analysis: Accounting for Continuity
1.0
Centrality Value
0.5
Y1( t)
Y2( t)
Y3( t)
0.0
Y4( t)
−0.5
Y1( t)
Y2( t)
Y3( t)
0.0
Y4( t)
−0.5
●
Centrality Value
● ● ● ●
2.0
Y1(t)
Y2(t)
Y3(t)
● ● ● Y4(t)
1.0
● ●
● ● ● ● ● ● ● ● ●
0.0
Figure 4.4: Centrality measures as a function of the time-interval for the Stress-Discomfort system.
Black dotted lines indicate the centrality values of the half-hour and one-hour networks in Figure
4.1(a) and (b).
87
4. A Continuous-Time Approach to Network Analysis and Centrality
4.3.3 Conclusion
To summarize, the CT-VAR model offers a promising alternative to DT-VAR net-
work approaches, allowing us to directly address the time interval problem. First,
it allows for an elegant treatment of unequal time-intervals between observa-
tions, which is a typical characteristic of ESM data. Second, the CT model offers
an alternative conceptualization of the dynamic network structure by breaking
down lagged relationships between measurements into their most fundamental
building blocks, that is, the moment-to-moment relationships encoded by the
drift matrix A. Given an estimate of this drift matrix (which can be obtained with
software packages such as ctsem (Driver et al., 2017) or dynr (Ou et al., 2019)), the
relationship in Equation (4.5) allows us to explore how the dynamic relations de-
pend on the time-interval, as we illustrated with the Stress-Discomfort system.
Third, we can use path-tracing logic suggested by previous authors to calculate
direct, indirect and total lagged relationships in a CT network.
The CT perspective also has consequences for the interpretation and practical
utility of DT centrality measures. Specifically, existing measures fail to account
for the CT nature of the underlying process, and —as a consequence— even their
intuitive interpretation as a summary of direct, indirect and total relationships is
not valid from a CT perspective. This implies that using DT centrality measures
to decide which variable should be intervened on is simply unfounded. Hence,
we need to define new centrality measures that adhere to the CT perspective, but
that also have a clear link with interventions.
88
4.4. Interventions and Centrality for CT Networks
89
4. A Continuous-Time Approach to Network Analysis and Centrality
90
4.4. Interventions and Centrality for CT Networks
1.0
1.0
0.5
0.5
Variable Values
Variable Values
● Y1(t)
Y2(t)
0.0
0.0
●
● Y3(t)
● Y4(t)
−0.5
−0.5
−1.0
−1.0
0.00 0.25 0.50 0.75 1.00 0.00 0.25 0.50 0.75 1.00
Time (t) Time (t)
Figure 4.5: Illustration of each total and direct effect of Anxiety (Y2 (t)) on Physical Discomfort (Y4 (t +
∆t)) in terms of the interventions they describe in the system over time. Acute interventions are
indicated by empty diamonds, and continuous interventions by filled diamonds.
to keep one or more mediating variables fixed (cf. Robins, 2003; Robins & Richard-
son, 2010; VanderWeele, 2015). For example, suppose that we can intervene in
our system to keep the variables Stress and Self-Conscious fixed to a normal value
at every moment in time over an interval. We will refer to this type of intervention
as a continuous intervention, denoting it do(Y1 (t + ∆t) = 0) and do(Y3 (t + ∆t) = 0)
for Stress and Self-Conscious respectively.5 For example, we can imagine that we
can prompt the participant to engage in a long-lasting mindfulness meditation
focused on reducing stress reactivity (Hoge et al., 2013), or, alternatively, that we
can administer a drug which keeps stress levels fixed to a low or normal level.
Now suppose that we are interested in evaluating the effect of momentarily
increasing Anxiety, on the value of Physical Discomfort, if we intervene to keep
Stress and Self-Conscious fixed. This is visualized by the trajectories in Figure
4.5(b). Just like the total effect intervention, Anxiety starts at a high level and
dissipates back to equilibrium. Here however, we keep Stress and Self-Conscious
fixed at all moments in time, so they stay at their equilibrium value. We can see
that the effect of increasing Anxiety on Physical Discomfort has changed: Phys-
ical Discomfort is pushed even further from equilibrium, taking on a stronger
negative value at short intervals. Since we keep the mediating variables fixed, we
no longer activate the positive feedback loop present in the total effect. Instead,
Physical Discomfort takes longer to return to baseline, still taking on a negative
value at t = 2.
Just as we did for the total effect, we can define the effect of this intervention
formally as the difference between two conditional expectations. Here we change
5 We can consider this to be a special case of an acute intervention defined by Eichler and Didelez
(2010), that is repeated continuously over time. See Appendix 4.E for details. Note that this type of
direct would be referred to as a controlled direct effect rather than a natural direct effect in the causal
inference literature (VanderWeele & Tchetgen Tchetgen, 2017). Since we assume a linear model, the
controlled and natural direct effect formulations lead to equivalent results.
91
4. A Continuous-Time Approach to Network Analysis and Centrality
Anxiety, but keep Stress and Self-Conscious fixed in both cases, and we can define
the effect of this combination of interventions as the direct effect. It can be written
formally as
h i
DE24·13 (∆t) = E Y4 (t + ∆t) do(Y2 (t) = 1), do(Y1 (t + ∆t), Y3 (t + ∆t) = 0)
h i
− E Y4 (t + ∆t) do(Y2 (t) = 0), do(Y1 (t + ∆t), Y3 (t + ∆t) = 0) (4.8)
where A(D[−1,−3]) denotes the drift matrix in which the indirect pathway parame-
ters linking Anxiety to Physical Discomfort through the mediating variables (that
is, a12 , a32 , a43 ) are set to zero; hence, in this drift matrix only the direct links be-
tween Anxiety and Physical Discomfort are retained. The proof that the effect of
this intervention can be expressed in this way is given in Appendix 4.E.
Again, we can see that this expression is exactly equivalent to the path-tracing
definition of a direct effect in a CT model given by Deboeck and Preacher (2016).
This means that the CT direct effect under ideal conditions describes the effect
of a very specific set of interventions: An acute intervention to change the value
of the cause variable, combined with a whole set of continuous interventions to
keep each mediating variable fixed in value. Note that we can also define direct
effects in which only one of the mediating variables is kept fixed with a continu-
ous intervention in an analogous way: For example, we could describe the direct
effect of Anxiety on Physical Discomfort relative to Self-Conscious as DE24·3 (∆t),
in which only Self-Conscious is kept fixed by a continuous intervention. Then we
would need to omit only the parameters a32 and a43 from the altered drift matrix
A(D[−3]) .
Since direct effects consist of combinations of at least two or more interven-
tions (depending on the number of mediators considered), they may not be the
most useful measure to inform interventions in practice. However, it is necessary
to define the direct effect in this way in order to define indirect effects, which we
discuss next, and which form the basis of the indirect effect centrality measure
that we present later.
92
4.4. Interventions and Centrality for CT Networks
In terms of the trajectories in Figure 4.5 the indirect effect is the difference in
value of Physical Discomfort at any point in time in panel (a)(representing the
total effect T E(∆t)), and the corresponding point in time in panel (b) (represent-
ing the direct effect DE(∆t)), and so this particular indirect effect describes the
mediating role of the variables Stress and Self-Conscious combined. We can ex-
press this indirect effect in terms of the CT-VAR parameters as
(D[−1,−3]) ∆t
IE24·13 (∆t) = eA∆t [42] − eA [42] (4.11)
that is, as a difference between the total and direct effect calculations given above.
It follows that the effect of this intervention is equivalent to the path-tracing def-
inition of the indirect effect (given in Appendix 4.D).
We may be interested in the indirect effect if we wish to answer questions like:
How does keeping Stress and Self-Conscious fixed alter the effect of a momentary
increase in Anxiety on Physical Discomfort? We can get an idea of the role of
these mediators by comparing the total effect and the direct effect trajectories in
Figure 4.5(a) and (b) respectively. In this instance we see that keeping both me-
diators fixed changes how Stress effects Self-Conscious, meaning Self-Conscious
reacts in a more strongly negative way to changes in Stress than if both mediators
are not kept fixed. We may also be interested in assessing indirect effects one
mediator at a time. For instance, if we want to decrease the effect that experi-
encing an acutely high level of Anxiety has on feelings of Physical Discomfort,
we may choose to either apply a continuous intervention on Stress (IE24·1 (∆t)) or
Self-Conscious (IE24·3 (∆t)), depending on whichever particular indirect effect is
largest. It is this kind of indirect effect that we will use later in defining a cen-
trality measure that describes the flow of information in the network through a
particular variable.
93
4. A Continuous-Time Approach to Network Analysis and Centrality
hence, we sum over all the total effects of Yi on other variables in the network
(excluding Yi itself). The T EC thus summarizes the effect of a single intervention,
an acute intervention to change Yi (t), on the system as a whole, that is, the cu-
mulative effect on the network, some time-interval ∆t later. Furthermore, since
we explicitly make this centrality measure a function of the time-interval, we can
examine how the cumulative effect of this intervention varies and evolves fol-
lowing the intervention moment. By calculating this measure for each variable,
we can directly inform the choice of optimal intervention target: Which variable
should I change the acute value of to achieve the biggest change in the cumulative
activation levels in the rest of the system later?
Figure 4.6(a) shows the T EC of each variable in the Stress-Discomfort system
over a range of intervals, from ∆t = 0 to ∆t = 1.5. From this we can see that
at short intervals, an acute intervention to increase Physical Discomfort has the
biggest cumulative effect on the network: Overall, this intervention on Physical
Discomfort results in the other variables increasing in value over the next half an
hour or so, before eventually the effect of this intervention fades away. Interven-
tions to increase Stress and Self-Conscious respectively have similar but weaker
effects. Notably, an intervention to increase Anxiety has a weak net negative ef-
fect on the system at shorter intervals, and a weak net positive effect at longer
intervals: We would expect this based on our visualization of that intervention in
94
4.4. Interventions and Centrality for CT Networks
Figure 4.5(a), where a pulse to Anxiety resulted in Stress and Physical Discomfort
taking on negative values at short intervals.
Based on this, if we want to pick the optimal intervention target for an acute
intervention, the T EC measure allows us to see that Physical Discomfort is the
optimal target for this type of intervention, assuming we can set Physical Dis-
comfort to a low or negative value (e.g. do(Y4 (t)) = −1).
where for each pair Yi (t) and Yj (t + ∆t) all possible mediating variables k ∈ p\(i, j)
are kept fixed. Note that unlike the T EC measure, DEC reflects a summary
measure of different interventions, as for each pair Yi (t) and Yj (t + ∆t), there is
a different set of mediators that must be intervened on to establish the direct
effect. For instance, in Figure 4.5(c) we showed the direct effect of Anxiety on
Physical Discomfort when intervening to keep Stress and Self-Conscious constant
(DE24·13 (∆t)). This represents only one of the three components which make up
the DEC of Anxiety: The other two terms consider interventions where either
Self-Conscious and Physical Discomfort are kept constant (DE21·34 (∆t)) or Phys-
ical Discomfort and Stress are kept constant (DE23·14 (∆t)). Although it is not
straightforward to use this measure to choose a specific intervention target, this
measure does reflect that part of one variables relationship with the network as
a whole which is truly a result of only direct relationships (from a path-tracing
perspective).
Figure 4.6(b) shows the direct effect centrality of each variable over a range
of intervals. From this we see that Stress has the highest positive DEC, Self-
Conscious and Physical Discomfort have weaker overall positive direct influ-
ences, that fade to zero at a quicker rate, and Anxiety has a net direct influence
closest to zero across a range of intervals.6 Although Stress has the strongest truly
direct influence on the network, the practical value of the DEC —or any measure
that is based on direct effects— for choosing intervention targets is limited, as it is
6 The high DEC of Stress is due to the positive feedback loop between Stress and Anxiety in the
underlying drift matrix (a21 = 5.5, a12 = 1.25). Since the direct effects involve interventions to control
the rest of the variables in the system, and no other pair of variables has this direct feedback loop
relationship, all other values of DEC are lower.
95
4. A Continuous-Time Approach to Network Analysis and Centrality
1
Centrality Value Y1( t)
Y2( t)
0
Y3( t)
Y4( t)
−1
Y1( t)
Y2( t)
0
Y3( t)
Y4( t)
−1
Y1( t)
Y2( t)
0
Y3( t)
Y4( t)
−1
Figure 4.6: Illustration of the new total, direct and indirect centrality measures for CT networks,
applied to the Stress-Discomfort system.
96
4.4. Interventions and Centrality for CT Networks
that is, the sum of all possible indirect effects between other pairs of variables
Yj (t) and Yk (t + ∆t), in which Yi serves as the only mediator. Note here that in
comparison to how we described the indirect effect above, we have switched the i,
j and k notation to reflect that the IEC(∆t) is defined as a property of a mediator,
instead of a property of one particular cause-effect relationship. The summation
denotes that we omit auto-regressive relationships (j , k) and pairs of variables
where the mediator is either the cause or effect variable (j , i and k , i). The IEC
therefore can be understood as quantifying how a continuous intervention on Yi
changes the effects of other variables on each other.
As such, this measure may be especially of interest in the case of networks
of psychopathology variables. For instance, in the Stress-Discomfort system, we
would like to avoid a high value on all four variables as much as possible. The
current measure can be used to determine which of these variables is most impor-
tant in terms of mediating the effects of one variable on another in the system,
such that by intervening on this variable, these indirect paths become blocked
and the flow of activation from one variable to another is interrupted.
Figure 4.6(c) shows the IEC of each variable over a range of intervals. From
this we can see that Physical Discomfort has the strongest indirect effect central-
ity in absolute terms. A strong negative value of IEC means that keeping Physical
Discomfort fixed actually increases the size the effects of other variables on each
other, since the component direct effects are greater than the corresponding total
effects. This happens because Physical Discomfort plays a key role in the only
negative feedback loop in the network: Since an increase in Anxiety actually de-
creases Physical Discomfort (a42 = −7.3), the total effect of Anxiety on Stress is
less strong than its direct effect. If, however, we intervene to keep Physical Dis-
comfort fixed, then this negative compensating effect is not activated, meaning
an increase to Anxiety in fact has a greater effect on the network as a whole. In
contrast, Stress has the largest positive IEC, meaning that keeping Stress fixed
decreases the effects of other variables on one another.
From this we would conclude that we should choose Stress as a target for a
continuous intervention, as it decreases the short-term impact of other variables
97
4. A Continuous-Time Approach to Network Analysis and Centrality
in the network on each other. However, we should avoid at all costs applying a
continuous intervention on Physical Discomfort: Such an intervention would in
fact increase the strength of positive relationships between the other variables.
4.5 Discussion
In this paper we have critically appraised current best practice in dynamical
network analysis. We have especially focused on the interpretation of DT-VAR
models as representing direct causal dependencies, and on the use of central-
ity measures predicated on that interpretation to identify optimal intervention
targets. We identified two major problems with current best practice and ad-
dressed each in turn. First, we addressed the time-interval problem by intro-
ducing a continuous-time approach to network analysis. CT models aim to cap-
ture the moment-to-moment dynamics operating between processes, which can
be represented in network form as a local dependence graph. The estimation of
a CT model overcomes both practical and conceptual shortcomings of the DT
approach, providing a new outlook on how parameters should be interpreted,
and the ability to explore how lagged relations vary and evolve as a function of
98
4.5. Discussion
the time-interval. Second, we address the lack of clarity regarding how current
centrality measures should be used to inform intervention targets. We do this by
introducing new centrality measures for CT models which have a clear link to
interventionist concepts of causal relationships and can be used to identify the
optimal target for either an acute or continuous intervention.
These developments represent a promising new approach to dynamical net-
work analysis. From a practical point of view, recent R packages ctsem (Driver
et al., 2017) and dynr (Ou et al., 2019), have made it relatively easy to apply the
CT modeling approach in empirical research (e.g., Voelkle et al., 2012; Voelkle
& Oud, 2013; Oud, Voelkle, & Driver, 2018; Ryan et al., 2018), and moreover,
in the supplementary materials of this paper we provide extensive R functions
which can be used to calculate CT path-specific effects and centrality measures
introduced in the current paper based on a given CT-VAR drift matrix.7
As this was a first step towards linking CT approaches, network analyses and
interventionist concepts of causal relationships, our approach was necessarily
simplified, both with respect to the interventions we consider, and our treat-
ment of the assumptions necessary for this type of causal inference. Eichler and
Didelez (2010) describe a variety of different interventions that could be applied
to dynamic systems, and a variety of assumptions that must be met for the effects
of these interventions to be identified from data. Driver and Voelkle (2018) de-
scribe how various hypothetical interventions can be simulated from a CT model,
however, they do not do so within an interventionist framework, and so the as-
sumptions needed to identify the effects of those interventions are unclear. More
research is needed to investigate and evaluate identifiability assumptions in a
psychological context, and to operationalize actual psychological treatments as
dynamic systems interventions. For instance, cognitive-behavioral interventions
may be better defined as interventions on moderators of symptom-symptom re-
lationships (i.e., lagged parameters of a network), rather than interventions on
the symptoms themselves.
Furthermore, the dynamic models we focused on here, that is the DT-VAR
and its CT-equivalent, are very simple models, describing a system that fluctu-
ates around a single fixed point. Although this reflects the majority of dynamical
network analyses in empirical research, it may be more beneficial —from a the-
oretical point of view— to investigate dynamic system models which can show a
qualitative change in behavior, for example bi-stable systems which can transi-
tion between different equilibria (Haslbeck & Ryan, 2019). For those systems we
may be interested in identifying which intervention should be applied in order
to move the system from one equilibrium position to another. We hope that the
current paper will serve as the groundwork for such future developments.
Finally, a further simplification we made in the current paper was to ignore
the role of uncertainty in parameter estimates. Previous research has shown that
existing centrality measure estimates can be somewhat unstable as they reflect
the sums of numerous parameter estimates (Epskamp, Borsboom, & Fried, 2018),
an issue which is likely shared by the CT centrality measures estimated here. In
7 https://osf.io/9sgdn/
99
4. A Continuous-Time Approach to Network Analysis and Centrality
100
4.A. Centrality Measures as Summaries of Path-specific Effects
101
4. A Continuous-Time Approach to Network Analysis and Centrality
Table 4.1: Relationship between different network metrics and path-tracing quantities, in the context
of a VAR(1) model with p variables, regression coefficient matrix Φ, and corresponding dynamical
network with weights matrix Φ T .
fect is strongest, and in turn, how often a specific variable acts as a mediator
of these strongest indirect effects. It seems that this is how psychological re-
searchers using the BCi measure typically interpret it (e.g., Bringmann et al.,
2013, 2015; David, Marshall, Evanovich, & Mumma, 2018). However, the ac-
tual calculation of this measure differs greatly from the mediator-based metric
described above. Specifically, instead of identifying the largest indirect effect,
Betweenness is based on the identification of the shortest network-path between
two variables (d(jk)). The length of this network-path is based on the inverse
of the sum rather than the product of the individual pathways: While large SEM
paths are those where multiplying each individual part leads to a high num-
ber, we say that short network-paths are those where the sum of each individual
part leads to a small number. Similar to standard path-tracing nomenclature,
these network-paths can be either direct (e.g. Yj,τ → Yk,τ+1 = φkj ) or indirect (e.g.
Yj,τ → Yi,τ+1 → Yk,τ+2 = φij + φki ) and each path may span a different number of
measurement occasions. The Betweenness Centrality of Yi is found by first cal-
culating all the shortest paths between all pairs of variables, and then counting
how often Yi lies on that shortest path. It is clear then that, despite how this mea-
sure is interpreted the relationship of Betweenness Centrality with path-specific
effects is much less direct than for the other measures considered above.
102
4.B. Centrality Values DT Stress-Discomfort System
∆t = 0.5 ∆t = 1
EI (2) EI (1) BC EI (2) EI (1) BC
Stress -0.025 -0.029 1 0.245 0.274 0
Anxiety 0.035 0.038 1 -0.071 -0.109 3
Self-Conscious -0.024 -0.027 0 0.125 0.152 0
Physical Discomfort 0.008 -0.002 1 0.555 0.557 0
Table 4.2: Two-Step Expected Influence (EI (2) ), One-Step Expected Influence (EI (1) ) and Betweenness
Centrality (BC) for each of the four variables in the half-hour (Φ(∆t = 0.5)) and one-hour (Φ(∆t = 1))
networks. These measures are based on the lagged parameter matrices displayed in Figure 4.1. In
each column, the largest centrality values are highlighted in bold.
dY (t) Y (t + s) − Y (t)
= lim
dt s→0 s
which means that the deterministic part of the first-order differential equation
103
4. A Continuous-Time Approach to Network Analysis and Centrality
104
4.D. Path-Tracing in CT models
This follows from the interpretation of the matrix exponential term eA∆t as a
path-tracing operation, relative to the moment-to-moment auto-regressive effects
matrix (I + A limn→∞ ∆t n ) (described in Appendix 4.C). In Figure 4.7(a) we show
a four-variable CT-VAR model with a full A matrix in path-model form, with
n → ∞ latent values of the processes in between measurement occasions, spaced
at intervals of s → 0. From this it is clear that tracing a path from, for instance,
Y1 (t) to Y4 (t + ∆t) includes paths through latent values of Y1 and Y4 , (e.g. Y1 (t) →
Y1 (t+1s) → Y1 (t+2s) → Y4 (t+3s) → · · · → Y4 (t+∆t)) as well as paths through latent
values of Y2 and Y3 (Y1 (t) → Y2 (t + 1s) → Y3 (t + 2s) → Y4 (t + 3s) → · · · → Y4 (t + ∆t)).
As such, we can interpret this total effect as constituted of all possible pathways
linking Y1 (t) and Y4 (t + ∆t), as is the standard interpretation of a total effect.
In order to find the path-tracing direct effect from Yi (t) to Yj (t + ∆t) relative
to some mediator variable(s) Yk , Deboeck and Preacher (2016) state that the drift
matrix should first be altered so that the parameters which make up the indirect
pathways are omitted. We can alter the drift matrix to achieve this by setting the
kth row and column elements of A to zero, yielding a drift matrix containing only
direct relationships between Yi and Yj , which we will denote A(D[−k]) . The path-
tracing direct effect is then found by applying the matrix exponential function to
the altered drift matrix.
(D[−k]) ∆t
DEij·k (∆t) = eA [ji] . (4.19)
For example, for a four-variable system, to define the path-tracing direct effect of
Y1 (t) to Y4 (t + ∆t) relative to the mediators Y2 and Y3 we would need to alter the
105
4. A Continuous-Time Approach to Network Analysis and Centrality
Figure 4.7: Path-model representation of the four-variable CT-VAR model with full drift matrix A. In
the top panel the red pathway highlights a path which is included in the difference-method calcula-
tion of the indirect effect but omitted from the path-method calculation. The bottom panel shows the
pathways which make up the direct effect, with indirect paths (removed from the altered drift matrix
A(D[−k]) ) shaded in gray
This altered drift matrix defines a new path model, absent of any lagged rela-
tionships linking Y1 to Y4 through the mediators Y2 and Y3 . This is displayed in
Figure 4.7(b). Applying the matrix exponential function to this new drift matrix,
it is clear that we only trace through direct pathways linking Y1 (t) to Y4 (t + ∆t)
(e.g. Y1 (t) → Y1 (t + 1s) → Y1 (t + 2s) → Y4 (t + 3s) → · · · → Y4 (t + ∆t)). This process is
exactly equivalent to how Bollen (1987) describes the calculation of a direct effect
106
4.E. Interventions and Path-Tracing in CT models
107
4. A Continuous-Time Approach to Network Analysis and Centrality
By assuming that the system is fully observed (i.e., there are no unobserved con-
founders), and by assuming modularity, we can substitute the expected value of
Yj (t + ∆t) following an intervention do(Yi (t) = yi ) with the expected value given
we observe Yi (t) = yi . This yields the expression
h i h i
T Eij (∆t) = E Yj (t + ∆t) Yi (t) = yi − E Yj (t + ∆t) Yi (t) = yi∗
Now we plug in the CT-VAR model for those expected values. Take it that Y (t)
represents a column vector of variable values with ith element Yi (t) = yi . Using
this, we can express the first expected value as
h i
E Yj (t + ∆t) Yi (t) = yi = {eA∆t Y (t)}[j]
that is, the jth element of the column vector obtained by multiplying the square
matrix eA∆t with the column vector Y (t). To obtain the second expectation, we
take it that Y ∗ (t) represents a column vector of variable values with ith element
Yi (t) = yi∗ but which is otherwise identical to Y (t).
h i
E Yj (t + ∆t) Yi (t) = yi∗ = {eA∆t Y ∗ (t)}[j]
Since the vectors differ only with respect to their ith element, we obtain
where eA∆t [ji] is the element in the jth row and ith column of the matrix eA∆t . If
we define the intervention as increasing the value of Yi (t) by one unit (yi − yi∗ = 1),
this yields an expression exactly equivalent to the path-tracing definition of a
total effect given in Appendix 4.D.
108
4.E. Interventions and Path-Tracing in CT models
for some mediator(s) k ∈ p. Intuitively, if we want to block the indirect effect that
acts through a mediator, we would need to ensure that either the mediator does
not react to changes in the cause variable, or that it does not transmit information
to the effect variable, or both. If we wish to achieve this by intervening on a
variable, it is straightforward to see that we must do so by intervening to set the
value of the mediator to a constant at every point in time between t and t + ∆t.
As with the total effect derived above, the next step consists of plugging in the
CT-VAR model for the expected values in this expression. However, note that due
to the need to define the continuous intervention on the mediator do(Yk (t + ∆t) =
yk this proof is a little more involved than that of the total effect above (and relies
on the strong assumption that applying such an intervention does not change
the dynamics of the underlying process). To derive an expression for the direct
effect, we first begin with the expression for the expected value of Y (t + ∆t) given
an acute intervention on the cause variable Yi (t), that is,
h i
E Y (t + ∆t) Yi (t) = yi = eA∆t Y (t)
one of the components of the total effect given above. Recall from the derivation
in Appendix 4.C that we can write the CT-VAR model as describing lagged rela-
tionships over an infinitesimally small time interval limn→∞ ∆tn , hereby referred
to as the moment-to-moment relationship. This gives us
h i
E Y (t + lim ∆t ∆t
n ) = (I + A lim n )Y (t).
n→∞ n→∞
where
h thei second and third line follow by substituting in the expression for
E Y (t + s) given above.
Now, to define the direct effect we need to express the expected value of Y (t +
2s) given that we have intervened to set the current value of the mediator (Yk (t +
2s)), the value of the mediator one “moment” previously (Yk (t + s)), and the initial
value of the mediator Yk (t) to zero. In order to derive such an expression, we
introduce two simplifications here. First, since we are focusing on a linear model,
and we are interested in the difference between two expected values in which
in both cases the mediator Yk is set to the same value yk , the specific value we
choose for yk is irrelevant. For ease of notation we will therefore consider only
an intervention by which yk is equal to zero (i.e., the equilibrium position of
Yk ). Second, to aid in our derivation, we will express the do operator in matrix
algebraic terms. That is, we will represent the operation do(Yk (t) = 0) using a
109
4. A Continuous-Time Approach to Network Analysis and Centrality
1 0 0
D[−2] = 0 0 0 .
0 0 1
Pre-multiplying a column vector by the matrix D[−k] reproduces the original col-
umn vector but with a zero as the kth element. That means that D[−k] Y (t) denotes
the acute intervention do(Yk (t) = 0). Again, for ease of notation we will drop the
[−k] notation and leave it implied, that is, in the proof below, D = D[−k] unless
otherwise specified.
Using this matrix-representation of the do operator, we can express the ex-
pected value of Y (t + 2s) given that we have intervened to set the current value of
the mediator (do(Yk (t + 2s) = 0)), the value of the mediator one “moment” previ-
ously (do(Yk (t+s) = 0)), and the initial value of the mediator (do(Yk (t) = 0)) to zero.
Subsequently, since we repeat this acute intervention at every “moment” in time
in an interval, we can describe it as a continuous intervention do(Yk (t + s) = 0)
that is, an intervention that is present for all possible time points in an interval.
The expected value of Y (t + 2s) given this continuous intervention can be written
as
h i
E Y (t + 2s) do(Yk (t + s)) = 0 = D(I + As)D(I + As)DY (t)
= (DI D + DADs)2 Y (t)
Now, using the same substitutions as described in Appendix 4.C, we can express
the expected value an arbitrary time interval ∆t later, given that we intervene to
set Yk to zero at each of the limn→∞ time points in that interval. This is given by
h i n o
n
E Y (t + ∆t) do(Yk (t + ∆t)) = 0 = limn→∞ (DI D + DAD ∆t n ) Y (t)
where DY (t) ensures that the initial value of Yk (t) is set to zero.
Pre- and post-multiplying A by D[−k] has the effect of setting the kth row
and column of A to zero. Hence, the expression eDAD∆t is exactly equivalent
to the path-tracing definition of the direct effect given in Appendix 4.D, that is,
110
4.E. Interventions and Path-Tracing in CT models
DAD = A(D[−k]) . This implies that by plugging the above expression in for the
expected values in the direct effect definition, we obtain
which shows that the effect on Yj (t + ∆t) of an acute intervention to change Yi (t)
combined with a continuous intervention to keep the mediator Yk fixed is identi-
cal to the path-tracing direct effect.
111
Chapter 5
This chapter has been adapted from: Haslbeck, J. M. B.* & Ryan, O.* (under review). Recovering
Bistable Systems from Psychological Time Series. Pre-print: https://psyarxiv.com/kcv3s/. Both
JMBH and OR are considered joint first authors, with both contributing equally to this project.
113
5. Recovering Bistable Systems from Psychological Time Series
5.1 Introduction
Conceptualizing psychopathologies as complex dynamical systems has become a
popular framework to study mental disorders (e.g., Wichers, Wigman, & Myin-
Germeys, 2015; Cramer et al., 2016; Borsboom, 2017). This framework is attrac-
tive because it acknowledges the fact that many mental disorders are massively
multifactorial (e.g., Kendler, 2019), and because it allows one to specify powerful
within-person dynamical systems models that capture many of the characteris-
tics hypothesized for mental disorders. The central goal of this framework is to
obtain such models to further our understanding of mental disorders, and allow
us to develop and test more successful interventions.
The class of dynamic systems that has received most attention in this emerg-
ing literature is the class of bistable systems (e.g., Wichers et al., 2015; Cramer et
al., 2016; Borsboom, 2017; Wichers, Schreuder, Goekoop, & Groen, 2019; van de
Leemput et al., 2014; Nelson et al., 2017; R. Kalisch et al., 2019). The reason is
that its behavior maps well on many phenomena observed in mental disorders:
Bistable systems describe variables that have two stable states, which can be in-
terpreted as different psychological states such as “healthy” or “unhealthy” (e.g.,
depressed). The stability landscape reflecting the dynamics of the system deter-
mines how easy it is to transition from one state to the other, and thereby offers
a possible formalization of properties of the mental disorder, such as vulnera-
bility or resilience to developing it (Scheffer et al., 2018). Bistable systems can
also show sudden transitions from one state to another, thereby mapping well
on, for example, bipolar disorder or the phenomenon of sudden gains and losses
in psychotherapy (Stiles et al., 2003; Lutz et al., 2013).
In parallel, the realization that inferences from between-subjects data to
within-person data are only possible under stringent assumptions (Molenaar,
2004; Hamaker, 2012) together with the increasing availability of psychologi-
cal time series collected from mobile devices has led to a surge in studies aiming
at recovering the within-person dynamics associated with mental disorders (e.g.,
Bringmann et al., 2013; Pe et al., 2015; A. J. Fisher et al., 2017; Snippe et al.,
2017; Groen et al., 2019). This is an exciting development, because within-person
time series potentially allow one to recover bistable systems and other dynamical
systems from empirical data. This would be a major step forward for studying
mental disorders as complex systems, because so far these models were only used
as stylized toy models.
However, so far there has been no systematic investigation into to what extent
dynamical systems models can in fact be recovered from psychological time se-
ries. To investigate this question for a given dynamical system and data analytic
method, it has to be broken down into two parts. The first question is whether
the method at hand can recover (some aspect of) a dynamical system in principle,
that is, with “ideal” data (long time series, extremely high sampling frequency).
If this is the case, the second question is whether the method also works with re-
alistic data (shorter time series, much lower sampling frequency). In the present
paper, we investigate both questions for a bistable dynamical system and a selec-
tion of the most popular (e.g., the Vector Autoregressive (VAR) model; Hamilton,
114
5.1. Introduction
1994) and some more advanced (e.g., differential equation estimation; Boker, De-
boeck, et al., 2010) methods. Specifically, we use a basic bistable dynamical sys-
tem for emotion dynamics to simulate both an ideal time series with extremely
high sampling frequency (measurement every six seconds) and a more realistic
time series with a sampling frequency common for Experience Sampling Method
(ESM) studies (measurement every 90 minutes). Using these time series, we eval-
uate how useful each method is for recovering bistable dynamical systems in
principle, and how useful it can be in practice when analyzing realistic ESM time
series.
We will show that the popular VAR model (and the Gaussian Graphical Model
fitted on its residuals; Epskamp, Waldorp, et al., 2018) is in principle unable
to recover the global dynamics (e.g., location and variance of stable states, fre-
quency of transitions) and succeeds only in recovering some of the microdynam-
ics (moment-to-moment interactions) of the true bistable system. However, de-
scriptive statistics, data visualization and more flexible statistical models are able
to capture the global dynamics. The only method that recovered the complete
bistable system is an iterative model building procedure that directly estimates
the system of differential equations (DEs). Reducing the sampling frequency
from every six seconds to every 90 minutes affects the considered methods differ-
ently. The VAR model and its extensions no longer recover any microdynamics,
and the DE-estimation procedure fails. However, descriptives, data visualization
and appropriate statistical models still recover the global dynamics. These re-
sults raise two fundamental issues involved in studying mental disorders from
a complex systems perspective. First, it is generally unclear what to conclude
from a statistical model about an underlying complex systems model. Second,
if the sampling frequency is too low, it is impossible to recover microdynamics.
In response to these findings, we outline a different research strategy to arrive at
dynamical systems models for mental disorders: Proposing initial formal models
which can subsequently be scrutinized and developed by deriving data implica-
tions that can be tested empirically. We will show that in this process many of the
presented methods are instrumental to testing predictions of the formal model
and thereby triangulating the formal model that captures the true dynamical sys-
tem best.
Our paper is structured as follows. In Section 5.2 we introduce a simple
bistable dynamical system for emotion dynamics, discuss its dynamics and char-
acteristics, and describe how we generate the ideal and the more realistic time
series from it. We use the ideal data (measurement every six seconds) in Section
5.3 to evaluate for each method to which extent it can recover a bistable dynam-
ical system. Next, in Section 5.4 we evaluate the same methods but using the
time series with a sampling frequency that matches typical ESM studies (mea-
surement every 90min). Finally, in Section 5.5 we discuss the implications of our
results for the framework of empirically studying mental disorders from a com-
plex systems perspective, and outline a new research strategy based on formal
modeling, which avoids shortcomings of a purely data analytic approach.
115
5. Recovering Bistable Systems from Psychological Time Series
where ri can be thought of as the main effect of an emotion on itself over time, that
is, the effect of xi on its own rate of change. This parameter is set to 1 for positive
emotions, and will be varied between r3 , r4 ∈ [0.9, 1.1] for negative emotions. We
interpret the variations in r3 , r4 as being related to stress: Higher stress means that
the effects of a high degree of negative emotion stays in the system longer. The
matrix C represents the dependencies between emotions in the form of interaction
effects
116
5.2. Bistable Emotion System as Data-Generating Model
The interactions in the matrix C show that emotions of the same valence re-
inforce each other, while emotions of different valence suppress each other. For
example, C12 = 0.04 indicates that the rate of change of x1 (Cheerful) depends
on the product of x1 and x2 (Content) weighted by 0.04. Similarly, C13 = −0.2
indicates that the rate of change of x1 depends on the product of x1 and x3 (Anx-
ious) weighted by −0.2. The diagonal elements are quadratic effects: For exam-
ple, C11 = −0.2 indicates that the rate of change of x1 depends on the product
x1 x1 = x12 . Note that we choose the matrix C to be symmetric purely for the sake
of simplicity. Since we aim to specify the simplest possible bistable system, we
specify that all within-valence effects (e.g., C12 and C34 ) are equal to 0.04 and all
between-valence interaction effects (e.g., C13 and C24 ) and quadratic effects (Cii )
are equal to −0.2.
We interpret xi = 0 as the absence of positive/negative emotion, and therefore
do no not allow emotions to become negative. We ensure this with high proba-
bility by setting the constant ai = 1.6 for all i. The Gaussian noise term i has a
mean of zero and a fixed standard deviation σ and represents short-term fluctu-
ations in emotions due to the environment the system interacts with. Note that
we used the same parameterization as van de Leemput et al. (2014), except that
in our model we use an additive noise term instead of a multiplicative noise term
for simplicity and set all ai = 1.6.
Due to the symmetries in C, r and a, emotions with the same valence are ex-
changeable. We can therefore describe the dynamics of the 4-dimensional system
using a 2-dimensional system consisting of one dimension for positive emotions
and one dimension for negative emotions (for details see Appendix 5.A). Figure
5.1 illustrates the dynamics of the deterministic part (i.e., with i = 0) of this
model: Panel (a) displays the stable (solid lines) and unstable (dashed lines) fixed
points for positive (green) and negative (red) emotions, as a function of stress. For
example, for a low stress level of 0.9 there is only a single fixed point: the positive
emotions (PE) have the value 5.28 and the negative emotions (NE) have the value
1.15. We therefore also refer to this fixed point as the healthy state. If the stress
level remains unchanged, the system will always end up at this fixed point, no
matter how one chooses the starting values. This dynamic is illustrated in the
corresponding vector field in panel (b). The arrows depict the partial derivatives
with respect to the two emotions and therefore describe the linearized dynamics
at a given point in the 2-d space. The vector field shows us that whichever ini-
tial values we choose, the system will always end up at the fixed point at (PE =
5.28, NE = 1.15). Thus, the system with stress = 0.9 describes a person whose
emotions can be changed by external influences, but eventually always returns to
the healthy state of having strong positive emotions and weak negative emotions.
The solid lines in panel (b) indicate the values of positive and negative emotions
for which the two differential equations are zero. At the intersections of those
117
5. Recovering Bistable Systems from Psychological Time Series
5
Emotion Strength
Positive Emotions
3
Negative Emotions
2
Negative Emotion
Negative Emotion
●
5 5 ● 5
● ●
0 0 0
0 5 10 0 5 10 0 5 10
Positive Emotion Positive Emotion Positive Emotion
Figure 5.1: The dynamics of the bistable system we will use as the data-generating model throughout
the paper. Panel (a) shows the fixed points of the deterministic part of the model as a function of
stress, operationalized by the rate of change of the negative emotions. Solid lines indicate stable fixed
points and dashed lines indicate unstable fixed points. Panels (b), (c) and (d) show the vector fields
of the system for the stress values r3 , r4 = 0.9, 1 and 1.1. Solid points indicate stable fixed points and
empty points indicate unstable fixed points. The solid lines indicate the values at which derivative of
positive emotion (orange) and negative emotion (light blue) is equal to zero. At the points at which
the two lines meet, both derivatives are equal to zero and the system remains in this (stable) state.
lines both differential equations are equal to zero, which means that the system
does not change anymore, which is the definition of a fixed point.
Panel (a) of Figure 5.1 shows that when increasing stress from 0.9 until around
0.95, the stable fixed point changes quantitatively: The value of positive emotion
value decreases, and the value of negative emotion value increases. However,
from around 0.95 on the dynamics of the system change qualitatively: the system
now has three fixed points. For example, at stress = 1, the fixed points are (PE =
4.89, NE = 1.36), (PE = 2.80, NE = 2.80), and (PE = 1.36, NE = 4.89). The first
fixed point is the stable healthy fixed point we also observed for values smaller
than 0.9. The second fixed point is an unstable fixed point. Specifically, it is a
saddle point, because the arrows in the vector field flow towards this fixed point
in one direction, but flow away in the other direction (Strogatz, 2015). The third
fixed point is again stable, however, now negative emotions have a high value and
positive emotions have a low value. We could call this fixed point the unhealthy
118
5.2. Bistable Emotion System as Data-Generating Model
fixed point.
The presence of these three fixed points means that, if the system is initialized
anywhere except on the diagonal, the system will end up at one of the two stable
fixed points. This behavior is illustrated in panel (c), which shows the vector
field of the system for stress = 1. We see that eventually all arrows point away
from the unstable fixed point (PE = 2.80, NE = 2.80) and towards one of the two
stable fixed points. Thus, the system will never converge to this point except if
it is initialized exactly on the diagonal. For all other starting values, the system
will converge to one of the two stable fixed points. For the particular case of
stress = 1, starting values above the diagonal line will converge to the unhealthy
fixed point (PE = 1.36, NE = 4.89), whereas starting values below the diagonal
line will converge to the healthy fixed point (PE = 4.89, NE = 1.36). This system
describes a person that starts out in the healthy (unhealthy) state, and always
returns to the healthy (unhealthy) state after small outside influences. However,
a large influence can push the person into the unhealthy (healthy) state, and now
the person remains there until a large enough influence pushes her back into the
healthy (unhealthy) state.
When increasing stress further until around 1.06, we observe again a quan-
titative change of the three fixed points: the negative emotions go up, and the
positive emotions go down. However, from around 1.06 on the system changes
again qualitatively. It now again exhibits only one fixed point, which is now the
unhealthy fixed point. Thus, when stress is larger than around 1.06, the system
will always converge to the unhealthy fixed point. This behavior is illustrated in
panel (d), which depicts the vector field for the system with stress = 1.1. We see
that there is only a single fixed point at (PE = 1.03, NE = 5.98) and the arrows
show that the system will always converge to this point. This system describes a
person that will always return to the unhealthy state, no matter how large of an
outside influence is applied.
So far, we only discussed the deterministic part of the model, that is, our
model with noise set to zero (i.e., with i = 0). Introducing noise changes the dy-
namics of the system, and how exactly it changes depends on the stress level. For
low stress (below 0.95), the system will fluctuate around the healthy fixed point.
For high stress (above 1.06), the system will fluctuate around the unhealthy fixed
point. The interesting behavior is observed for stress values between 0.95 and
1.06: then, the system will fluctuate around one of the two fixed points, but oc-
casionally the noise will be large enough to push the system to the other fixed
point. The frequency of switching is a function of the distance between the two
fixed points, the vector field between the two fixed points, and the variance of
the Gaussian noise process i . If the variance is low, the probability of a noise
draw that is large enough to “push” the system to the other fixed point is small,
and consequently the frequency of switching is low. In contrast, if the variance
is high, the probability of a large enough noise draw to switch to the other fixed
point is high, and consequently the switching frequency is high.
119
5. Recovering Bistable Systems from Psychological Time Series
120
5.2. Bistable Emotion System as Data-Generating Model
(a) (b)
8 Cheerful
Content
Anxious
Emotion Strength
Sad
6
0 5 10 14
Days
Figure 5.2: Panel (a) shows the ideal time series of the four emotion variables Cheerful, Content,
Anxious and Sad. We see that the system switches 17 times between healthy and unhealthy state.
Panel (b) displays the twelfth switch, which is a transition from the unhealthy to the healthy state,
which occurs on day 9.
or the sampling frequency too low. We therefore first study all methods with the
ideal time series in order to identify their fundamental limitations. In the second
part of the paper, we make the time series more realistic by taking measurements
at a sampling frequency that is typical for ESM studies. This will allow us to
investigate the impact of sampling frequency on all methods. In the following
section we describe how we generate this ESM time series.
121
5. Recovering Bistable Systems from Psychological Time Series
yields 224 measurements. This would mean that we would compare the “ideal”
time series with 201600 measurements which essentially implies the absence of
sampling variation to an ESM time series with 224 measurement which implies
quite a considerable degree of sampling variation. Thus, any comparison would
be confounded by the difference in the number of measurement points (i.e. sam-
ple size). To avoid this confound, we increase the measurement interval of the
ESM time series to 1800 weeks, which ensures that the new ESM time series has
exactly the same sample size as the ideal data ( 2242 × 1800 = 201600). Thereby,
we ensure that any drop in performance is a function of the lowered sampling
frequency and cannot be explained by lower sample size (and higher sampling
variation). Note that studying the performance of methods as a function of sam-
ple size (sampling variation) is of paramount importance to evaluate how useful a
given method is in a realistic application. However, here we study the more fun-
damental question of the impact of reducing the sampling frequency to a level
that is typical for ESM studies. We do this because if we find that a method is
ill-suited to recover (some aspect of) a bistable system with a realistic sampling
frequency on the population level (i.e., with infinite sample size), then it does
not make sense to investigate the performance of the method in the less ideal
scenario with realistic (small) sample sizes.
So far, we only discussed that we sample every 90 minutes. However, to em-
ulate ESM measurements, we also need to formalize how exactly ESM questions
measure the four emotion variables. This is far from trivial: questions in some
ESM studies refer to the very moment of measurement and are phrased along
the lines of “How cheerful do you feel right now?”. Such measurements could
be formalized by defining the measurement as the set of current values of the
system (a “snapshot” of the system) at the measurement time. In contrast, other
ESM studies refer to the time period since the last measurement. A question of
this type could be phrased “How cheerful did you feel in the time since the last
notification?”. Such measurements could be formalized by defining the measure-
ment as the average values of the system since the last measurement. However,
many other measurement functions are also possible. In this paper we analyze
the first type of ESM question, because its measurement function is the simplest.
However, we also performed all analyses with the second kind of ESM question,
and all our main conclusions also hold in this situation.
Figure 5.3 displays the two week long original time series (see also Figure 5.2
panel (a)) next to the ESM time series which was obtained by taking “snapshots”
of the process at 90 min intervals. The ESM time series in panel (b) appears
less dense, which is what we would expect since it contains only 1/900 of the
time points of the ideal time series in panel (a). However, we see that the system
is still bistable and that the location of and variance around the fixed points is
largely the same. In Section 5.4 we will use this emulated “snapshot” ESM time
series to try to recover the true bistable system using the same array of methods
as in Section 5.3, in which we analyze the ideal time series.
122
5.2. Bistable Emotion System as Data-Generating Model
Emotion Strength
6 6
4 4
2 2
0 0
0 5 10 14 0 5 10 14
Days Days
Figure 5.3: Panel (a) shows the original time series that was already shown in panel (a) of Figure 5.2.
Panel (b) shows the ESM time series which was obtained by taking snapshots every 90 minutes in the
series. Note that the ESM time series we analyze in Section 5.4 is much longer (1800 weeks) than the
14 day ESM time series shown here.
Global dynamics
1. Bistability (two stable fixed points)
123
5. Recovering Bistable Systems from Psychological Time Series
The first four characteristics describe the global dynamics of the dynamical
system. The first is bistability, which means that the data-generating mechanism
exhibits two stable fixed points. This is the case for the data-generating mecha-
nism with stress set to 1, which we use to generate data from and aim to recover
throughout the paper (see Figure 5.2, panels (a) and (c)). The second character-
istic is the position of the fixed points, which are at (PE = 4.89, NE= 1.36) for
the healthy fixed point, and (PE = 1.36, NE= 4.89) for the unhealthy fixed point.
Third, we consider the variability around the different fixed points. Figure 5.2
shows that, for both fixed points, the variability of the emotions with lower val-
ues is smaller than the variability of the emotions with larger values. The fourth
characteristic is the frequency of transitions between the area around the healthy
fixed point and the area around the unhealthy fixed point. In the time series
shown in Figure 5.2 we see that the system switches around 17 times.
The remaining three characteristics describe the microdynamics of the dy-
namical system. The fifth characteristic is that emotions of the same valence re-
inforce each other, while emotions of different valence suppress each other. The
sixth characteristic is the fact that the size (absolute value) of the reinforcing ef-
fects (0.04) are smaller than the suppressing effects (0.2). The last characteristic
is that all parameters in the system of differential equations are independent of
time and independent of variables outside the model.
124
5.3. Recovering the Bistable System from Ideal Data
switching behavior using time-varying parameters (Hamaker, Zhang, & van der
Maas, 2009; Hamaker & Grasman, 2012). While all models so far are misspeci-
fied, we include one final method that is capable of recovering the full bistable
system: a two-step model building approach based on direct estimation of differ-
ential equations from data, following the dynamic systems modeling approach
of Boker, Deboeck, et al. (2010) and Chow (2019).
Density
Density
Density
0 2 4 6 8 0 2 4 6 8 0 2 4 6 8 0 2 4 6 8
Figure 5.4: The histograms of the emotion intensity of the four modeled emotions Cheerful, Content,
Anxious and Sad, for the ideal data.
We see that at most time points in the time series, each emotion either takes
on values around 1 or around 5. This is what we would expect from inspecting
the time series plot, however, the histograms give a more precise picture of the
distributions and allow one to guess possible fixed points with greater precision.
For instance, we could separate the two distributions (using a fixed threshold, or
clustering algorithm) and take their means as estimates for the fixed points.
While eyeballing the data should be the first step in any time series analysis,
the conclusions are subjective and do not allow us to quantify how certain we are
about bistability and the switching frequency. We can quantify the observation
that there are two states and that the system is switching between them by fitting
a Hidden Markov Model (HMM) (e.g., Rabiner, 1989) to the data, which we will
125
5. Recovering Bistable Systems from Psychological Time Series
We can see from the estimate µ̂1 that in state 1 the means of positive emotions
are low, and the means of negative emotions are high. We can therefore identify
state 1 as the unhealthy state. We also see that the standard deviations of positive
emotions are lower than for negative emotions in the unhealthy state which is
what we already observed in the time series plot in Figure 5.2. Similarly, state 2
can be identified as the healthy state, with high means and standard deviations
for positive emotions, and low for negative emotions. The transition matrix A
indicates the probabilities of switching between states. We see that there is a very
high probability for remaining in the same state (Â11 = Â22 = 0.9996), and a cor-
respondingly low probability to switch states (Â12 = Â21 = 0.0004). This is what
we would expect, because we take one measurement every six seconds, but the
system changes states only a couple of times within the two week window. Multi-
plying the number of time points of the time series with the switching probability
we obtain 201600 × .0004 ≈ 81 switches, which is in the same order of magnitude
of the eyeballed number of switches (17) reported in Section 5.2.2.
In addition to obtaining estimates of means and standard deviations of the
two fixed points and the transition matrix, the HMM allows to predict the most
likely state for each time point. We show the predicted states for the entire time
series in Figure 5.5:
2 In principle, the distributions may also differ with respective to their covariances, but in this
analysis, we set all covariances to zero due to limitations of the software package used in estimation.
126
5.3. Recovering the Bistable System from Ideal Data
State 1 State 2
8
Emotion Strength
6
Cheerful
Content
Anxious
4 Sad
0 5 10 14
Days
Figure 5.5: Time series of the four emotion variables, also shown in panel (a) of Figure 5.2, with
background color indicating whether a given time point is assigned to the first or second component
of the mean-switching HMM.
When inspecting the predicted states visually, it seems that the HMM cap-
tured the switches well. Next to the larger blocks in which the system stays at the
same fixed point, it also identifies switches in which the system switches back
and forth within only a few time points. These switches might have been missed
when inspecting the time series visually alone.
Taking all results together, which characteristics of the bistable system did
we recover with the HMM (based on the list in Section 5.2.3)? We obtained an
estimate of the location (characteristic 2) and variance (characteristic 3) around
two fixed points, which are very close to the healthy and unhealthy fixed points
in the true bistable system. We also quantified the frequency of transitions in
the transition matrix A. Since the transition frequency (characteristic 4) is not
explicit in the true bistable model, there is no clear way to evaluate this estimate.
However, the number of predicted transitions (81) is at least of the same order of
magnitude as the number of transitions eyeballed from the entire time series (at
least 17). Note that while a bistable HMM seems to fit the data well, we provided
K = 2 as an input to the model, and therefore bistability (characteristic 1) cannot
be considered a characteristic we recovered with this model. Instead of fixing
a particular K, an optimal K can be obtained via model selection. However, in
Appendix 5.B.2 we show that at least the standard approach to selecting K in
mean-switching HMMs performs poorly since the data was not generated from a
mean-switching HMM.
One additional way to visualize or ascertain how much of the true systems
behavior a given model is able to capture is by generating new data from the
estimated model parameters. In Figure 5.16 in Appendix 5.C.1 we generate a two
week time series from the estimated mean-switching HMM and compare it to the
original time series. We find that the data generated from the HMM is similar
to the original data, except for two features: First, the system tends to switch
between states somewhat more frequently, and second, there are no observations
on the transitions between states.
127
5. Recovering Bistable Systems from Psychological Time Series
The remaining three characteristics (5-7) are about the microdynamics of the
true bistable system, that is, about how the components are related to each other.
Clearly, the mean-switching HMM we used here cannot elucidate these charac-
teristics since it does not model any dependencies between the four emotion vari-
ables. In the following sections we fit models that include such dependencies.
8 (a) 8 (b)
6 6
Anxious
Content
4 4
2 ρ = 0.98 2 ρ = −0.97
0 0
0 2 4 6 8 0 2 4 6 8
Cheerful Cheerful
−0.97 −0.24
−0.97 −0.24
Anxious 0.98 Sad Anxious 0.51 Sad
Figure 5.6: Panel (a) shows the relationship between Content and Cheerful, two emotions with the
same valence, at the same time point. The red line indicates the best fitting regression model. Sim-
ilarly, panel (b) shows the relationship between Anxious and Content, two emotions with different
valence. Panel (c) displays the correlation matrix as a network, and panel (d) displays the partial
correlation matrix as a network.
128
5.3. Recovering the Bistable System from Ideal Data
Panel (b) displays the relationship between Cheerful and Anxious, two emo-
tions of different valence. We see that that the observations cluster around two
points, one close to (1, 5) and the other one close to (5,1). The red line indicates
the best fitting regression line (correlation ρ = −0.97).
Panel (c) displays the correlation network for all four emotion variables. As
we have already seen in panel (a) and (b) there is a positive correlation (ρ = 0.98)
between Content and Cheerful, and a negative correlation (ρ = −0.97) between
Cheerful and Anxious. Due to the symmetry in the true bistable system, all cor-
relations between emotions with the same valence are equal to ρ = 0.98 and all
correlations between emotions with different valences are equal to ρ = −0.97.
Panel (d) shows the partial correlation network (i.e., GGM). We see that the par-
tial correlations between emotions with the same valence are equal to θ = 0.51,
and the partial correlations between emotions with different valences is equal to
θ = −0.24 or θ = −0.25.
What can we learn from these results about the underlying bistable system?
From inspecting the pairwise relationships of emotions with same and different
valence in panels (a) and (b) one could guess the location and variance of pos-
sible fixed points, similarly to inspecting the histograms in Section 5.3.1. How-
ever, the 2-dimensional representation offers additional information about the
stability landscape, for example the shape around the fixed points and the most
likely paths to transition between them. When interpreting the correlations in
panel (b) as “contemporaneous” relationships, we would conclude that there are
strong positive linear relationships between emotions with the same valence, and
similarly strong negative linear relationships between variables with different va-
lences at a relatively short time scale. The partial correlations in (d) are smaller
than the correlations, which is what one would expect since all correlations are
high.
Using our knowledge about the true bistable system, which characteristics
did we correctly recover? From inspecting the scatter plots in panels (a) and (b)
one sees that most observations fall in one of two clusters indicating bistability
(characteristic 1). Also, one can obtain rough estimates of the position of the
fixed points (characteristic 2) and sees that the variances around the fixed points
is different (characteristic 3). Note that the shape of the scatter plot in panel (b)
is determined by the vector field in Figure 5.1 (c). The two clusters are exactly at
the location of the two fixed points, and the observations between the clusters are
both due to variance around the fixed points and switches between fixed points.
From the correlation and partial correlation networks, we correctly find that
there are reinforcing effects within valences, and suppressing effects between va-
lences (characteristic 5). However, the correlation network suggests that their
relative size is equal, and the partial correlation network suggests that the rein-
forcing effects are stronger. In the true bistable system, however, the suppressing
effects between valences are larger than the reinforcing effects within valences.
Thus, judging the relative size of suppressing/reinforcing effects within/between
valences from (partial) correlation would lead to incorrect conclusions.
In sum, inspecting scatter plots of pairwise relationships indicated bistability,
and allowed us to obtain a rough estimate of the location of and variances around
129
5. Recovering Bistable Systems from Psychological Time Series
the fixed points. The scatter plots also allowed one to get a projection of the
stability landscape on two dimensions and thereby provided more information
than histograms. While inspecting the scatter plots allows one to recover global
dynamics of the true bistable system, one cannot infer the coupling between the
emotion variables in the true bistable system from (partial) correlations. This
is not too surprising since the Gaussian distribution is very restrictive in that
it only models pairwise linear relationships (opposed to e.g., 3-way, 4-way, etc.
interactions). In addition, it does not model any dependencies across time, which
are the types of dependencies that constitute the microdynamics (characteristics
5-7) of the true model. In the next section, we inspect those dependencies across
time and fit a Vector Autoregressive (VAR) model to the data, which captures
temporal linear dependencies.
Xt = b + ΦXt−1 + et (5.2)
where b is a vector of intercepts, Φ is a matrix containing the auto-regressive (φii )
and cross-lagged (φij , i , j) effects, that is, conditional linear dependencies, and et
is a vector of normally distributed residuals et ∼ N (0, Ψ ), which are independent
across time, with residual variance-covariance matrix Ψ .
The VAR(1) model has been used widely to analyze experience sampling data
in psychopathology research, particularly in the form of dynamic network analy-
sis, wherein the Φ and Ψ matrices are used to construct directed and undirected
network structures, respectively (e.g., Bringmann et al., 2013; Pe et al., 2015;
Epskamp, Waldorp, et al., 2018). The VAR(1) model describes a system which
fluctuates around a single stable fixed point: Stochastic input in the form of a
130
5.3. Recovering the Bistable System from Ideal Data
residual term pushes the system away from this fixed point, and the system re-
turns to the fixed point with an exponential decay (Hamilton, 1994). The location
of the stable fixed point is given by the mean vector µ, a function of the intercepts
and the lagged relationships µ = (I − Φ)−1 b, where I is the identity matrix.
8 (a) 8 (b)
6 6
Anxioust
Contentt
4 4
2 2
0 0
0 2 4 6 8 0 2 4 6 8
Cheerfult 1 Contentt 1
0.05
Cheerful Content Cheerful 0.04 Content
0.05
−0.04
−0.02 −0.02
−0.02 −0.02 −0.02 −0.02
−0.02 −0.03 −0.03
−0.02
−0.03
0.05
Anxious Sad Anxious 0.04 Sad
0.05
0.91 0.91
Figure 5.7: Panel (a) shows the relationship between Content and Cheerful, two emotions with the
same valence, spaced one time point apart (at a lag of one). The red line indicates the best fitting
regression model. Similarly, panel (b) shows the relationship between Anxious and Content, two
emotions with different valence, at a lag of one. Panel (c) displays the matrix of lagged regression pa-
rameters, estimated from a VAR(1) model, as a network, and panel (d) displays the partial correlation
matrix of the residuals of the VAR(1) model as a network. This latter network is often referred to as
the contemporaneous network.
Panel (c) of Figure 5.7 displays the network of estimated lagged regression co-
efficients (Φ̂) between Cheerful, Content, Anxious and Sad. We can see that the
auto-regressive parameters are large and positive for all four variables (φ̂ii = .91).
Furthermore, there are positive cross-lagged relationships between variables of
the same valence (φ̂12 = φ̂21 = φ̂43 = φ̂34 = .05) and weaker, negative cross-
lagged effects between variables of opposite valence (φ̂13 = φ̂31 = · · · = −.02).
All within-valence effects, and all between-valence effects, are of roughly equal
131
5. Recovering Bistable Systems from Psychological Time Series
132
5.3. Recovering the Bistable System from Ideal Data
(1)
Xt = b(1) + Φ (1) Xt−1 + et if zt ≤ τ
(2)
Xt = b(2) + Φ (2) Xt−1 + et if zt > τ
133
5. Recovering Bistable Systems from Psychological Time Series
for a two-regime model with a single threshold, where the VAR(1) parameters are
(r)
indexed by regime, with et ∼ N (0, Ψ (r) ), and mean vectors µ(r) = (I − Φ (r) )−1 b(r)
(Tong & Lim, 1980; Hamaker, Grasman, & Kamphuis, 2010). The threshold vari-
able zt may be an exogenous variable, or one of the variables in the VAR model.
Here we choose to use Cheerful (zt = x1,t−1 ) as the threshold variable. The thresh-
old value τ is a hyper-parameter that is estimated. Here, we estimate the TVAR(1)
model using the R-package tsDyn (Fabio Di Narzo, Aznarte, & Stigler, 2009),
which estimates τ using a grid search which selects the model with minimum
summed squared residuals.
Figure 5.8 displays the main results from the estimated TVAR(1) model, in
which the threshold is estimated as τ̂ = 2.811. In panel (a) of Figure 5.8 we
show the time-series with shading indicating which observations are below (grey)
or above (white) the threshold. We can see that the estimated threshold nicely
separates the time series into periods in which the system is in an unhealthy state
(based on Cheerful values below the threshold) and a healthy state (Cheerful
values above the threshold).
8
Emotion Strength
0 5 10 14
Days
(b) Healthy Regime (c) Unhealthy Regime
0.88 0.88 0.87 0.88
0.02 0.01
Cheerful Content Cheerful Content
0.02 0.01
0.01 0.02
Anxious Sad Anxious Sad
0.02 0.02
0.88 0.87 0.88 0.88
Figure 5.8: Panel (a) shows the two weeks of the time series, with observations shaded in either grey
or white as a function of whether x1,t−1 is above or below the threshold τ̂ = 2.811. Panels (b) and
(c) show the estimated VAR(1) parameters as lagged networks in the healthy (white) and unhealthy
(grey) regimes respectively.
134
5.3. Recovering the Bistable System from Ideal Data
Inspecting the lagged networks for each regime in panels (b) and (c) of Fig-
ure 5.8 we can see that the auto-regressive effects, and the within-valence cross-
lagged effects are pretty similar across both regimes. However, the cross-lagged
effects between variables of opposite valence are different. In the healthy regime,
negative valence emotions have much stronger cross-lagged effects on positive
(2) (2) (2) (2)
emotions (φ̂13 = φ̂14 = φ̂23 = φ̂24 = −.08), and vice versa for the unhealthy
(1) (1) (1) (1)
regime (φ̂31 = φ̂41 = φ̂32 = φ̂42 = −.08). Residual partial correlation networks
for both regimes are shown in Appendix 5.D, which display a similar pattern to
the regular VAR(1) model of weak positive residual partial correlation within-
valence and weak negative residual partial correlation between-valence. For the
TVAR, however, the residual covariance matrix is not symmetric across regimes:
In the healthy regime there is a slightly higher covariance between positive emo-
tions than negative emotions, and vice versa. The estimated means are given as
µ̂2 = {4.74, 4.75, 1.45, 1.46} for the healthy state and µ̂1 = {1.49, 1.48, 4.69, 4.69} for
the unhealthy state. Data generated by the TVAR(1) model estimates is shown in
Figure 5.18 in Appendix 5.C.3. From this figure we can see that most of the global
dynamics are well reproduced, although the system contains fewer switches be-
tween regimes than we would expect and there are fewer observations on the
switches between states compared to the original time series.
Which characteristics of the bistable system do we recover on the basis of the
TVAR(1) parameter estimates? First, the model picks up a number of character-
istics related to the bistability of the system. The estimated mean vectors capture
approximately the position of the two stable fixed points (characteristic 2), and
the estimated threshold correctly captures the position of the unstable fixed point
in the Cheerful dimension. However, note that bistability (characteristic 1) has
been specified a-priori and therefore cannot be considered to be recovered by the
model. Second, although the simulated data in Figure 5.18 (Appendix 5.C.3) ex-
hibits less frequent switches between states than we would expect, we can see
that the combination of state-dependent lagged parameters and residual vari-
ances does reproduce higher variability of positive emotion in the healthy state
in comparison to the unhealthy state, and vice versa for negative emotions (char-
acteristic 3). Finally, the lagged regression parameters in each regime correctly
capture that there are reinforcing effects within valence, and suppressing effects
between valence (characteristic 5).
The result that stands out in this analysis is the asymmetry in lagged regres-
sion coefficients across both regimes. This asymmetry would appear to indicate
that the parameters relating processes either change over time or are all explic-
itly a (step) function of the Cheerful variable. This last result is striking because
this intuitive interpretation does not correctly characterize the relationship be-
tween variables of different valences in the true bistable system. This is because
we know that the dependencies in C are invariant over time and fully symmetric.
However, the dependencies in C relate to pairwise interaction effects rather than
linear dependencies in the VAR(1) model. For example, the relationship between
Anxious, denoted x3 , and the rate of change of Content, dx 2
dt , depends both on the
value of C23 and on the current value of x2
135
5. Recovering Bistable Systems from Psychological Time Series
dx2
= r2 x2 + (C23 × x2 )x3 + . . . . (5.3)
dt
If we view x2 as a moderator, we can see that, when x2 is high, the effect of x3
on the rate of change, given by C23 × x2 , is relatively greater than when x2 is low.
In our system, separating the time-series into two regimes based on a threshold
of 2.811 for the Cheerful emotion essentially means we condition on high val-
ues of x1 and x2 in the healthy regime, and low values in the unhealthy regime.
This leads to the relatively stronger linear relationship from negative emotions to
positive emotions in the healthy regime, and vice versa in the unhealthy regime.
As such, we can see that the asymmetry in lagged relationships over time picked
up by the TVAR(1) model is a characteristic of the true bistable system. Notably,
however, the mechanism by which this asymmetry occurs is entirely due to non-
linear relationships between the observed variables and the similarity of vari-
ables that share the same valence, while the TVAR(1) modeler might be tempted
to ascribe this entirely to the effect of the level of Cheerful.
To summarize, the TVAR(1) model allows us to recover global dynamics, and
it recovers some aspects of the microdynamics. However, we saw that a naive in-
terpretation of the TVAR(1) parameter estimates may easily lead to the incorrect
conclusion that there is one time-varying variable which moderates the relation-
ships between all variables. In addition, we provided bistability as an input to
the model, and therefore cannot be considered a characteristic recovered from
data. In principle one could perform model selection between TVAR(1) models
with different numbers of components, however compared to the Mean switching
HMM in Section 5.3.2, the run time for such a model comparison was unfeasible
for the large data set used in our paper.
Furthermore, note that the threshold VAR(1) model does remarkably well for
this specific system for the following reason: While TVAR(1) models have fre-
quently been discussed in the literature (e.g. Warren, 2002; Hamaker et al., 2009,
2010; De Haan-Rietdijk, Gottman, Bergeman, & Hamaker, 2016) a major limi-
tation of this method is the difficulty in choosing a threshold variable. In our
data-generating mechanism, we know there to be an unstable fixed point defined
in multivariate space, x1 = x2 = x3 = x4 = 2.8. It just so happens that in this case,
almost always when we pass this position in one dimension (e.g., x1 > 2.8) we also
do so in all other dimensions (e.g., x2 > 2.8, x3 < 2.8, x4 < 2.8). This means that
the true mechanism of state-switching behavior is very well approximated by the
univariate mechanism in the TVAR(1) model, for this choice of parameter val-
ues. In more general situations, the choice of threshold variable(s), and number
of thresholds, is likely to be less trivial. While the TVAR(1) model does capture
that there are suppressing and reinforcing effects between and within valences,
it does not capture the relative size of these effects, and it may easily lead to the
incorrect conclusion that there is a single time-varying variable which moderates
all of the relationships between other variables in the system.
Finally, the TVAR(1) is only one of a variety of different regime-switching
dynamic models which could be fitted to the data at hand. Another alternative
would be the Markov-Switching (MS-)VAR model (Hamilton, 1989; Hamaker et
136
5.3. Recovering the Bistable System from Ideal Data
al., 2010; Hamaker & Grasman, 2012; Chow et al., 2018), a combination of the
HMM and VAR models, in which the regime-switching behavior is determined by
a random Markov process operating between latent categorical variables. While
this model is more flexible than the Threshold VAR model, we show here the
TVAR results for two reasons. First, in this instance the switching behavior will
be less well approximated by the MS-VAR model, leading to even less straightfor-
ward conclusions about the data-generating process, but otherwise likely highly
similar lagged parameter estimates. Second, while recent advances such as the
dynR package (Ou et al., 2019) have made this model easier to estimate, it is still
prohibitively difficult and time consuming to fit to data.3
Now that we have shown the capabilities and limitations of the TVAR model
in recovering the bistable system, there are a few different avenues we could pur-
sue to further increase our model complexity in the hope of recovering more and
more of the features of underlying system. For example, both the TVAR and
MS-VAR can be considered special cases of time-varying parameter models, that
assume the true time-varying model is a partition between a finite set of com-
ponents. Other types of time-varying VAR models assume the parameters are a
smooth function of time (e.g., Haslbeck, Bringmann, & Waldorp, 2017). How-
ever, we would not expect these models to outperform the threshold VAR in this
instance for two reasons. First, the threshold VAR model is already able to cap-
ture the major source of variation in parameters over time, that is, the step-like
switches between stable states. Second, since these models are still based on fit-
ting locally stationary VAR models, the fundamental limitations of approximat-
ing the dynamics with linear relationships remain. As such, in the next section
we examine an approach which aims to recover the exact system of differential
equations (DEs) from data, by allowing non-linear terms to enter into a step-wise
model building procedure.
137
5. Recovering Bistable Systems from Psychological Time Series
138
5.3. Recovering the Bistable System from Ideal Data
dxi,t
Model ∼ ai + ri xi + . . .
dt q R2
P
A j,i rj xj 5 0.04464
P Pp
B j,i Rij xj + j Cij xj xi 9 0.06874
P Pp
C j,i Rij xj + (j,k) βj xj xk 15 0.06870
Table 5.1: Model fit results for each of the four models described in text. The second column gives
the model equation for each variable, q denotes the number of parameters estimated per univariate
regression model, and the final column indicates the mean proportion of explained variance R2 ,
calculated on the hold-out sets of a 10-fold cross-validation scheme (for details see Appendix 5.E)
h iT
â = 1.40 1.37 1.25 1.27
h iT
σ̂ = 1.35 1.34 1.34 1.34
10
Positive Emotion
Figure 5.9: Left panel: the parameters estimated from the ideal data. Right panel: the vector field
defined by the estimated parameters in the left panel. Solid points indicate stable fixed points and
empty points indicate unstable fixed points. The solid lines indicate the values at which derivative of
positive emotion (orange) and negative emotion (light blue) is equal to zero. At the points at which
the two lines meet, both derivatives are equal to zero and the system remains in this (stable) state.
While we would not expect to recover the exact parameters of the true model
with a different functional form, we see that the signs, size and relative orderings
of parameters in the estimated Ĉ matrix are quite accurate. Based on these pa-
rameters, we recover that there are suppressing effects between valences and re-
139
5. Recovering Bistable Systems from Psychological Time Series
inforcing effects within valences (characteristic 5), that the reinforcing effects are
smaller in absolute value than the suppressing effects (characteristic 6), and by
capturing approximately the correct functional form, we capture that the micro-
dynamic parameters are dependent only on variables inside the model (charac-
teristic 7). Furthermore, we can see that false positive (i.e., off-diagonal) elements
of R̂ are of a much smaller size than the true positive diagonal elements. The full
parameter estimates, with standard errors and p-values are shown in Appendix
5.E.
Beyond inspecting the estimated parameters, we can judge how good of an
approximation of the true bistable system our estimated model represents by
comparing the dynamics implied by that model to that of the true system. The
dynamics of a differential equation model are described by its vector field, which
we depict for Model B in the right panel of Figure 5.9. To construct this vec-
tor field we use the same two-dimensional approximation (positive and nega-
tive emotion) as we did in Section 5.2.1 (see Appendix 5.A for details). The or-
ange and light blue lines are solution lines which indicate the locations where
the rate of change in one dimension (orange for no change in positive emotion,
light blue for no change in negative emotions) is zero. The points at which these
solutions line cross determine the fixed points. We can see that our model cor-
rectly identifies three fixed points in this range of values: one stable healthy
(x1 = x2 = 4.91, x3 = x4 = 1.34), one stable unhealthy (x1 = x2 = 1.39, x3 = x4 =
4.84), and one unstable fixed point approximately halfway between those two
(x1 = x2 = 2.79, x3 = x4 = 2.82). If we compare these global dynamics to the global
dynamics of the true bistable system depicted in Figure 5.1 in Section 5.2.1, we
see that Model B very accurately reproduces these dynamics, approximating the
position of the fixed points in the true system closely. From this we can say that
the estimated DE model captures characteristics 1 (bistability) and 2 (location of
the fixed points) in the true system.
An additional way to evaluate whether the dynamics of the estimated model
are similar to the dynamics of the true model is to generate data from the esti-
mated model and compare this data to the original data. Figure 5.10 shows a time
series generated from Model B using a step size of .1. We can see that the data
looks very similar to the original data generated from the true bistable system, in
that the fixed points are at roughly the same location, there is a difference in vari-
ance across the high and low emotion value fixed points (characteristic 3), and
there is a similar number transitions (around 14) between the healthy and un-
healthy state (characteristic 4). Thus, even though we did not exactly recover the
set of true parameters exactly, we seem to have recovered a model that is equal
to the true model in all relevant aspects, capturing all of the seven characteristics
listed in Section 5.2.3.
140
5.3. Recovering the Bistable System from Ideal Data
Emotion Strength 8
6
Cheerful
Content
Anxious
4 Sad
0 5 10 14
Days
Figure 5.10: Data generated from the estimated DE model, with the same initial values as the ob-
served data
141
5. Recovering Bistable Systems from Psychological Time Series
5) ) 7)
) .( (6 t(
1) 3) ( 4 n f e a n
y( 2) ( ns ei Siz onst
ili
t n( ce tio ./R tive
sta
b
sit io
ri an
an
si
pp r
la e-c
i o a r u e i m
B P V T S R T
Data Visualization X X X × × × ×
HMM X∗ X X X × × ×
Lag-0 / GGM × × × × X × X∗
Lag-1 / VAR(1) × × × × X × X∗
TVAR(1) X ∗
X X X X × ×
DE-Estimation X X X X X X X∗
Table 5.2: Summary of which method recovered which of the seven qualitative characteristics listed
in Section 5.2.3 from the ideal time series. The first four characteristics are global dynamics, thee
last three are microdynamics. The check marks with asterisk indicate that the method includes the
characteristic as a model assumption, and can therefore not be considered recovered from the time
series.
We showed that data visualization (Histograms and the pairwise marginal re-
lationships in Sections 5.3.3 and 5.3.4) revealed bistability and provided a rough
estimate of the position of and variances around the fixed points. However, when
comparing the eye-balled number of switches with the estimates of the HMM,
we saw that we missed instances in which the system quickly switched back and
forth. The Mean switching HMM recovered all global dynamics, however, we
provided bistability as a model assumption, which is why we mark the check
mark at the first characteristic with an asterisk.
Turning to methods that capture dependencies between variables, the analysis
of lag-0 relationships with the GGM and the analysis of lag-1 relationships with
the VAR(1) model (and a GGM on its residuals) fundamentally cannot recover
any global dynamics of the bistable system, but they recovered some micrody-
namics: the characteristic that within valence effect are reinforcing, and between
valence effects are suppressing; and that the parameters are constant across time,
however this is again an assumption of the model and therefore cannot be consid-
ered recovered from the data. The TVAR(1) model was able to recover all global
dynamics with the same caveat as in the HMM, that bistability is a model as-
sumption and not recovered from data. Similarly to the VAR model, it recovered
the reinforcing/suppressing characteristic. However, a naive interpretation of
the model parameters would lead one to conclude that the parameters are time-
varying. Finally, the DE-estimation method was able to recover all microdynam-
ics reasonably well, which implies that it also recovered all global dynamics.
The purpose of this section was to establish whether or not each method can
recover, in principle, some aspect of the bistable system. To do this we used
a highly idealized dataset, with an unrealistically high sampling frequency. As
such, the performance of each method described above can be considered an up-
per bound on its performance in any more realistic scenario. It remains to be
seen exactly how the performance of each method, and in general our ability to
142
5.4. Recovering the Bistable Systems from ESM Data
and the predicted states for two weeks of the time series are shown in Appendix
5.F Figure 5.22.
We see a very similar pattern of results as obtained from the HMM fit to the
ideal time series in Section 5.3.2, with the means and standard deviations of state
1 and state 2 reflecting the unhealthy and healthy states respectively. However,
the parameters of the estimated transition matrix A for this time series show sub-
stantially higher switching probabilities, Â12 = .085 and Â21 = .090. As we can
see from Figure 5.3, although the sub sampled ESM time series contains only
224 observations for a two-week period, rather than 201600, the sampling fre-
quency is still high enough to capture each of the 17 switches between states in
this period. That means that, although the amount of transitions that occur over a
period of time remains the same, the number of measurement occasions between
any two transitions is lower, which results in a higher transition probability. We
143
5. Recovering Bistable Systems from Psychological Time Series
can see that this higher transition probability captures the number of transitions
over two weeks quite accurately — the model predicts between 224 × .090 ≈ 20
and 224 × .085 ≈ 18 switches on average over a two week period. As such, the
HMM fitted on the ESM time series still allows us to estimate the location of and
variance around the two fixed points (characteristics 2 and 3), and approximate
the frequency of transitions between these two fixed points (characteristic 4). In
fact, the transition probabilities appear to be even more accurate than the ideal
case — most likely this numerical imprecision in the ideal case is because the
number of transitions relative to total time series was so low that slight changes
in the transition probability value lead to very different predictions regarding the
number of transitions over 201600 time points.
144
5.4. Recovering the Bistable Systems from ESM Data
8 (a) 8 (b)
6 6
Anxioust
Contentt
4 4
2 2
0 0
0 2 4 6 8 0 2 4 6 8
Cheerfult 1 Contentt 1
0.21
Cheerful Content Cheerful 0.51 Content
0.2
−0.23
−0.21 −0.19
−0.2 −0.2 −0.21 −0.19
−0.2 −0.23 −0.24
−0.19
−0.23
0.19
Anxious Sad Anxious 0.51 Sad
0.2
0.2 0.19
Figure 5.11: Panel (a) shows the relationship between Content and Cheerful, two emotions with
the same valence, spaced one measurement occasion apart (i.e. at a lag of one but with 90 minutes
between measurements) for the ESM dataset. The red line indicates the best fitting regression model.
Similarly, panel (b) shows the relationship between Anxious and Content, two emotions with different
valence, at a lag of one. Panel (c) displays the matrix of lagged regression parameters, estimated from
a VAR(1) model, as a network, and panel (d) displays the partial correlation matrix of the residuals
of the VAR(1) model as a network. This latter network is often referred to as the “contemporaneous”
network.
145
5. Recovering Bistable Systems from Psychological Time Series
146
5.4. Recovering the Bistable Systems from ESM Data
147
5. Recovering Bistable Systems from Psychological Time Series
8
Emotion Strength
0 5 10 14
Days
(b) Healthy Regime (c) Unhealthy Regime
0.08 0.14
−0.03 0.06
0.06 0.14
Cheerful Content Cheerful Content
−0.03 0.08
0.15 0.03
Anxious Sad Anxious Sad
0.16 0.04
Figure 5.12: Panel (a) shows the first two weeks of the time series, with observations shaded in either
grey or white as a function of whether x1,t−1 is above or below the threshold τ̂ = 2.796. Panels (b) and
(c) show the estimated VAR(1) parameters as lagged networks in the healthy (white) and unhealthy
(grey) regimes respectively.
in each state (i.e., high variance for positive emotions, low variance for negative
emotions in the healthy state, and vice versa) and those observations which jump
from one fixed point to the other across consecutive measurement occasions, as
discussed in the previous section.
As we did throughout Section 5.3, we could evaluate how well this model de-
scribes the bistable system by generating data from it. Notably, the dynamics
defined by Φ (1) and Φ (2) reflect an unstable system in both regimes: The eigenval-
ues of both contain a value outside the unit circle (i.e., with absolute value greater
than one) (Hamilton, 1994). This means that, if we were to generate data using
these parameters, the time series would always diverge towards infinity. This in-
stability also precludes us from making any statement regarding the variance of
positive and negative emotions in each regime (characteristic 3), as the long run
variances implied by the model are infinite. As such, we can say that overall, the
set of estimated parameters for the TVAR(1) based on the ESM time series are a
poor characterisation of the microdynamics of the model at any time-scale.
148
5.4. Recovering the Bistable Systems from ESM Data
In summary, the TVAR(1) model fitted on the ESM time series still picks up a
global characteristic of the system, but the recovery of microdynamic character-
istics fails. In fact, the relationship between the estimated lagged parameters and
the characteristics of the system was much more opaque than in the ideal data
case, and our ability to generalize from the estimated parameters to the behavior
of the system at any time scale was considerably worse than in the ideal case.
Again here, we should note that the only difference between the ideal and ESM
time series is the sampling frequency. Fundamentally, the results here indicate
that, if we do not have a sufficiently high sampling frequency, then fitting increas-
ingly complex models, or extensions to simpler models such as the TVAR(1), does
not aid us in recovering the characteristics we are interested in: Even when we
have an arbitrarily large number of observations, we fail to recover basic charac-
teristics of the microdynamics due to the spacing between measurements.
149
5. Recovering Bistable Systems from Psychological Time Series
dxi,t
Model dt ∼ a + ri xi + . . . q R2
P
A j,i rj xj 5 0.13991
P Pp
B j,i Rij xj + j Cij xj xi 9 0.16827
P Pp
C j,i Rij xj + (j,k) βj xj xk 15 0.16928
P Pp Pp 3
D j,i Rij xj + (j,k) βj xj xk + j γj xj 19 0.19455
P Pp Pp 3 P p
E j,i Rij xj + (j,k) βj xj xk + j γj xj + j,k,l ζj (xj xk xl ) 23 0.19801
P Pp Pp 3 Pp
F j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl ) 35 0.19940
P Pp Pp 3 Pp
j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl )
G Pp 70 0.20420
+ (j,k,l,m) ηj (xj xk xl xm )
Table 5.3: Model fit results for each of the seven models described in text, for the emulated snapshot
ESM data. The second column gives the model equation for each variable, q denotes the number
of parameters estimated per univariate regression model, and the final column indicates the mean
proportion of explained variance R2 , calculated on the hold-out sets of a 10-fold cross-validation
scheme (for details see Appendix 5.E)
150
5.4. Recovering the Bistable Systems from ESM Data
h iT
â = −0.07 −0.16 0.31 0.20
h iT 10
σ̂ = 0.01 0.01 0.01 0.01
●
Negative Emotion
−0.16 −0.18 −0.01 −0.04
R̂ =
−0.09 −0.29 −0.05 −0.01
5
●
−0.12 −0.22 −0.07 0.10
●
Positive Emotion
Figure 5.13: Left panel: the parameters estimated from the snapshot ESM time series. Right panel:
the vector field defined by the estimated parameters. Solid points indicate stable fixed points and
empty points indicate unstable fixed points. The solid lines indicate the values at which derivative of
positive emotion (orange) and negative emotion (light blue) is equal to zero. At the points at which
the two lines meet, both derivatives are equal to zero and the system remains in this (stable) state.
are further denoted by dots, with filled dots indicating a stable fixed point, and
empty dots indicating an unstable fixed point.
We can immediately see from Figure 5.13 that the stability landscape is much
more complex than the one of the true bistable system, with high-degree poly-
nomial solution lines, and with four rather than three fixed points. Interest-
ingly, the system correctly identifies that there are two stable fixed points re-
lating to the healthy state (x1 = x2 = 4.51, x3 = x4 = 1.67) and the unhealthy state
(x1 = x2 = 1.71, x3 = x4 = 4.47), and that there is an unstable fixed point ap-
proximately half-way between those two (x1 = x2 = 2.99, x3 = x4 = 3.19). Despite
having an entirely different functional form, the estimated model does capture
two stable fixed points (characteristic 1) and the approximate position of those
fixed points (characteristic 2). This shows that Model G performs well in captur-
ing the characteristics of the system for emotion values that were observed in the
time series, that is, near the two stable fixed points.
Crucially, however, we cannot say that this system recovers the global dy-
namics of the true system, not least because the system contains an additional
unstable fixed point at (x1 = x2 = 5.31, x3 = x4 = 7.94), which is not present in
the true bistable system. The presence of this unstable fixed point means that if,
for instance, both negative and positive emotions take on a high value simulta-
neously, then the system enters an unstable region and diverges to infinity. If we
examine the behavior of the system even further outside the range of observed
values (−∞ > X > 0 and 10 < X < ∞) even more fixed points and regions of stabil-
ity and instability can be found. We can further demonstrate these dynamics by
generating data from Model G. Figure 5.14 shows a time series generated from
151
5. Recovering Bistable Systems from Psychological Time Series
the difference-form of Model G (i.e., with a step size equal to that of the observed
data).4 We see that the process moves between the healthy and unhealthy fixed
point for the first ten days, exhibiting the bistable behavior we see in the true
system. However around the eleventh day, the stochastic input is large enough
to move the system to an unstable region in the vector field and which leads the
system to diverge.
(a) (b)
8 Cheerful
Content
Anxious
Sad
6
Emotion Strength
0 5 10 14
Days
Figure 5.14: Data generated from the estimated DE model, with the same initial values as the “ideal”
data
Note that the complexity of the final model here is not a result of over-fitting
the data, as we performed model selection based on the out-of-bag R2 , an ap-
proximation of the out-of-sample R2 . Rather, the complexity of this model can
be attributed to two factors. First, due to the low sampling frequency, our ap-
proximation of the derivative at each point in time is poor. The second, as we dis-
cussed in Section 5.3.6.3, is that given the spacing between observations, the best
one can hope for is to approximate the integral solution to the data-generating
equation, which is likely of a highly complex functional form. The ability of the
misspecified Model G to reproduce some characteristics in regions where we have
observed data can be attributed to the high flexibility afforded by the many non-
linear terms. In that sense, this behavior is highly comparable to the problem of
using a high-degree polynomial regression model to make predictions outside of
the range of observed values. The vector field in Figure 5.13 is constructed by
obtaining predicted values for the derivatives across a grid of input values and
as such, it is unsurprising that the vector field is accurate where the input values
are close to the observed data, and inaccurate elsewhere.
In summary, we do not at all recover the functional form or parameters of
the system; we do recover some of the global characteristics and behavior of the
4 This is obtained by re-fitting the differential equation using the unscaled difference x
i,t+1 − xi,t as
the outcome variable, leading to equivalent results with parameters approximately scaled by dt = 90.
The residual variance used is the estimated residual variance scaled down to .65 the magnitude, to
account for the non-normal residual distribution. Using the estimated residual standard deviation
results in shocks which immediately move the system into an unstable region.
152
5.4. Recovering the Bistable Systems from ESM Data
system in the region where we have observations, capturing that there are two
stable fixed points and one unstable fixed point, and their locations. However,
the estimated model also implies the presence of at least one extra unstable fixed
point, which has major implications for the dynamics of the model, implying
divergent behavior. Thus, the estimated model implies fundamentally different
microdynamic and global characteristics. Based on the simulated data in Figure
5.14, it does not seem that we correctly capture the variability around these fixed
points, or the frequency of transitions, as any reasonable simulation of data from
this model eventually leads the system to diverge. Crucially, we fail in recov-
ering an interpretable approximation of the data-generating model. As such, it
is not feasible to assess whether there are truly suppressing effects between va-
lences and reinforcing effects within valences, or the relative size of these effects
(characteristics 5 and 6).
(5
) ) 7)
) 4) f. (6 nt(
ty
(1 2) (3
) s( n ein
e
Siz onst
a
b ili i o n( n ce i tio r. /R ive - c
s t
sta sit ria an pp la m
e
Bi Po Va Tr Su Re Ti
Data Visualization X X X × × × ×
HMM X∗ X X X × × ×
Lag-0 / GGM × × × × × × X∗
Lag-1 / VAR(1) × × × × × × X∗
TVAR(1) X∗ X X X × × ×
DE-Estimation × × × × × × ×
Table 5.4: Summary of which method recovered which of the seven qualitative characteristics listed
in Section 5.2.3 from the ESM time series. The first four characteristics are global dynamics, the
last three are microdynamics. The check marks with asterisk indicate that the method includes the
characteristic as a model assumption, and can therefore not be considered recovered from the time
series.
Our main findings are that, in general, we remain able to recover global char-
acteristics of the system using simple methods, but that we are completely unable
to recover any of the microdynamics. We saw that each approach which aimed to
capture microdynamic characteristics either deteriorated dramatically in perfor-
mance (for the VAR and TVAR approaches) or broke down altogether (for the DE
model building approach) as soon as we applied them to a time series obtained
with a more realistic sampling frequency. This is despite the fact that the time
series we used in this section can be considered a highly idealized approxima-
tion to ESM time series, in terms of the number of observations and the quality
153
5. Recovering Bistable Systems from Psychological Time Series
5.5 Discussion
In this paper we explored to what extent dynamical systems models can be re-
covered from psychological time series by investigating two successive questions.
First, how well does a set of popular and more advanced methods recover (char-
acteristics of) a basic bistable system with an ideal data set sampled at extremely
high sampling frequency (every six seconds)? And second, how is the perfor-
mance of each method affected when reducing the sampling frequency to one
measurement every 90min, which is typical for ESM studies.
When analyzing the ideal time series we found that the popular VAR model
(and the GGM fitted on its residuals) can in principle not recover the global dy-
namics of the true bistable system, and only recovers some of its microdynamics.
However, we showed that descriptive statistics, data visualization and statistical
models which are based on mixtures (the HMM and threshold VAR) were able
to capture the global dynamics of the bistable system. The only method that re-
covered the full bistable system was a differential equation (DE) model building
procedure. Reducing the sampling frequency from every six seconds to every 90
minutes affected the considered methods differently. The VAR model and its ex-
154
5.5. Discussion
155
5. Recovering Bistable Systems from Psychological Time Series
much simpler than our bistable system) that completes one cycle within, say, 2
minutes, has to be sampled at least every minute to be recovered (e.g., Marks,
2012; Papoulis & Pillai, 2002). This suggests that it is futile to try use a time series
sampled every 90 minutes to directly recover dynamics of emotions that operate
on a time scale of seconds or minutes (Houben, Van Den Noortgate, & Kuppens,
2015) or even from moment to moment (Wichers et al., 2015). However, this also
means that ESM time series can certainly be used to recover processes that unfold
at a time scale of several hours or days.
To summarize, we identified two fundamental barriers to studying mental
disorders from a complex systems perspective. First, even with extremely high
sampling frequency it is generally unclear how to make inferences from a statis-
tical model to an unspecified dynamical systems model. Second, the sampling
frequency of the data collection constrains the type of processes one can recover.
Specifically, a process can only be recovered if the sampling frequency is suf-
ficiently high. Clearly, these are profound problems every empirical discipline
struggles with and no simple answers can be expected. Indeed, they might imply
that studying some aspects of mental disorders will always remain out of reach.
That said, we believe that much progress can be made by studying mental disor-
ders as complex systems and that acknowledging and studying the above issues
allows one to do so more efficiently. As a way forward, in the following section we
suggest a new research strategy based on proposing substantively plausible dy-
namical systems, which opens up avenues to creatively tackle the two problems
identified in this section.
156
5.5. Discussion
and the empirical model are in agreement, we have tentative evidence that the dy-
namical system model is correct; if not, we can use the nature of the disagreement
to improve the dynamical system model. Clearly, this modeling approach, which
is typical to more quantitative disciplines such as physics, chemistry and biology,
is different to the statistical modeling framework most psychological researchers
are familiar with. On the one hand these formal dynamical systems models
are harder to build, since they cannot be estimated directly from the data. On
the other hand, they are powerful enough to be plausible for complex phenom-
ena such as mental disorders, and have additional benefits such as synthesizing
knowledge, revealing unknowns, laying open hidden assumptions and enabling
checking of the internal consistency of a model (Epstein, 2008; Lewandowsky &
Farrell, 2010; Smaldino, 2017).
This modeling approach also allows to tackle the problem of sampling fre-
quencies that are too low to recover the process of interest directly, as one can
generate a time series from the specified dynamical systems model and reduce
the sampling frequency to a level that is also available in empirical data. Then,
similarly to above, one can again compute the statistical model of choice that
is implied by the dynamical systems model with a given sampling frequency,
compare it to the corresponding model fit on empirical data, and in the case of
disagreement adapt the dynamical systems model accordingly. Of course, this
approach is not a panacea. Less information is available when the sampling fre-
quency is low, which makes model identification more difficult. However, spec-
ifying an initial dynamical systems model allows one to gauge how difficult it is
to recover a given type of process on a given time scale with a given sampling
frequency.
In addition, starting out with a dynamical systems model also allows to study
the measurement function that defines the mapping from the variables in the
dynamical systems to the obtained measurements, a topic we only touched on
briefly in this paper. In our emulated ESM time series we took the measurement
function to return the exact values of variables at the time point of measurement.
However, different questions imply different measurement functions. For exam-
ple, if the phrasing of a particular question refers to the entire period since the
last measurement, one could instead define the measurement as a function of the
variable values since the last measurement, such as the average. Next to formaliz-
ing which experiences an ESM question refers to exactly, defining a measurement
function also allows to formalize known response and memory biases, such as the
recency effect (Ebbinghaus, 1913/2013).
Finally, having a plausible dynamical systems model allows one to explicitly
address a behavior that has been largely ignored in the psychological time series
modeling literature, that is, the fact that humans sleep. Sleep interacts with es-
sentially everything physiological and psychological, is part of the definition of
several mental disorders (e.g., Major Depression) and related to many more (e.g.,
Walker, 2017). Thus, for many mental disorders, it seems necessary for a plausi-
ble model to include sleep. This may also allow using existing data in new ways,
because data around the “day-night shift” does not have to be excluded anymore,
but instead can be used to test hypotheses about the sleep-related assumptions
157
5. Recovering Bistable Systems from Psychological Time Series
5.5.3 Limitations
Several limitations of our work require discussion. First, our goal was to explore
to which extent one can recover (bistable) dynamical systems for mental disor-
ders from psychological time series. However, we only studied a single bistable
system. Therefore, it could be that the fundamental problems identified in the
paper and summarized in Section 5.5.1 are in fact a particularity of the chosen
bistable system. This, however, seems extremely unlikely: First, because we iden-
tify the problems in our paper as examples of well-known issues such as model
misspecification and sampling systems with a sampling frequency that is suffi-
cient for recovery. Second, the bistable system we chose is arguably the simplest
bistable system for four variables one can find. Choosing a different model there-
fore results most likely in choosing a more complex model, and our intuition is
that the methodological difficulties discussed in this paper become more and not
less relevant in such models.
Second, a more specific criticism of our bistable system could be that the time
scale of the process is unrealistically small, and we therefore exaggerated the
problem of recovering dynamics of psychological processes from ESM time se-
ries. We agree that it is possible that some psychological processes are easier
to recover from ESM data than the dynamical system used in this paper. Thus,
strictly speaking, we only showed that it is impossible to recover a system if the
sampling frequency does not appropriately match the time scale of the system. In
principle, it is therefore an open question whether there is a mismatch between
the time scale of the system of interest and the available sampling frequency.
However, intuition — and the sampling theorems such as the one mentioned in
Section 5.5.1 — strongly suggest that it is impossible (or at least very difficult) to
recover a process that operates at a time scale of seconds or minutes from an ESM
time series that is measured every 1.5 hours. Clearly, however, our investigation
is only a first treatment of the important topic of sampling frequency, and much
work on it is required to establish a tight connection between psychological time
series and dynamical systems models.
Third, one could reverse the argument in the previous two paragraphs and
argue that our model is so ideal that many analyses perform better than in most
158
5.5. Discussion
realistic applications. This is certainly the case for the Threshold VAR model,
which performs well only because of the simple dynamics of the bistable system
as we discussed in Section 5.3.5. Other examples are the descriptive statistics
and data visualization which may not be as insightful if fixed points are closer to
each other and if there is more noise in the system. Also, the two-step approach
to estimating the differential equations in Section 5.3.6 may work less well for a
more complicated model. Thus, we would agree with this assessment, however
chose to use a simple bistable system in order to make the paper more accessible
to applied researchers.
Fourth, we analyzed a bistable system whose structural parameters do not
change over time. However, much of the framework of considering mental dis-
orders as complex systems is based on the idea that pathology is defined with
respect to a structural change in the underlying system, and therefore structural
change is of central interest. We expect that structural change renders the re-
covery of a system more difficult, and we therefore did not include this feature
in order to keep the paper at a reasonable length. However, we believe that fu-
ture methodological research into how to recover such structural changes both in
principle and with realistic time series would be extremely helpful to better un-
derstand phenomena such as early warning signals (Scheffer et al., 2009; van de
Leemput et al., 2014) and more generally structural change in mental disorders.
Fifth, in order to estimate a differential equation from data, we took a rather
simple two-step approach based on local linear approximation of the derivative
(cf. Boker, Deboeck, et al., 2010). This approach involves first estimating the
derivative itself using scaled difference scores, and then using this derivative as
an outcome variable in a regression model. While this method benefits from be-
ing extremely simple to implement, we could expect that it would perform poorly
in the presence of low sampling frequency as the quality of the derivative approx-
imation degrades (as we noted in Section 5.3.6.3 and observed in Section 5.4.4).
There are multiple alternative approaches to estimating DE equations which we
did not consider here. For example, approaches based on numerical integration
of the DE equation during estimation, such as implemented in dynR (Ou et al.,
2019) and stan (Carpenter et al., 2017) (with additional functionality in the ct-
sem package; Driver et al., 2017) may in general perform better than the two-step
procedure when the sampling frequency is low. However, for the analysis shown
in the present paper, neither the ctsem nor dynr package performed better than
the two-step approach. In general, however, more research is needed to map out
which method deals best with the problem of low sampling frequencies.
Lastly, throughout our paper we studied how well certain analysis methods
can recover the true bistable system in principle. We did this by studying the pop-
ulation properties of these methods, that is, the situation in which one has essen-
tially infinite sample size, which we approximated with a huge number (201600)
of measurements. This was necessary in order to study the more fundamental
questions of (1) whether a given method can recover our bistable system in prin-
ciple and (2) whether a given method can recover our bistable system based on
a time series with realistic sampling frequency. We did this because it would be
meaningless to study the performance of a method as a function of sample size,
159
5. Recovering Bistable Systems from Psychological Time Series
if the method alredy fails with infinite sample size. Clearly, however, to apply
any of the methods we studied in practice, one has to know how reliable they are
with which sample size, and much more research is necessary to map our these
sample size requirements (e.g., Dablander, Ryan, & Haslbeck, 2019).
5.5.4 Summary
In the present paper we identified two fundamental problems involved in study-
ing mental disorders from a complex systems perspective: first, it is generally
unclear what to conclude from a statistical model about an unspecified underly-
ing complex systems model. Second, if the sampling frequency of a time series
is not high enough, it is futile to attempt to recover the microdynamics of the
underlying complex system. In response to these problems, we proposed a new
modeling strategy that takes an initial substantively plausible dynamical systems
model as a starting point, and develops the dynamical systems model by testing
its predictions. In this approach it is much clearer what we can learn from data
and statistical models about an underlying dynamical system, and in addition it
provides avenues to move the field forward by formalizing the sampling process,
measurement, response and memory biases, measurement reactivity and the in-
fluence of sleep.
160
5.A. Determining Fixed Points
4
X
0 = r1 x1 + C1j xj x1 + a1
j=1
4
X
0 = r2 x2 + C2j xj x2 + a2
j=1
4
X
0 = r3 x3 + C3j xj x3 + a3
j=1
4
X
0 = r4 x4 + C4j xj x4 + a4
j=1
Since we have r1 , r2 = 1 and a = [1.6, 1.6, 1.6, 1.6] in all studied situations, we fill
in those values and write out the summation:
We can exploit the symmetries in r and C to simplify finding the fixed points.
The derivatives of x1 and x2 are actually identical, and the derivatives of x3 and
x4 are identical. Thus also their integrals are identical. Thus, we can substitute
x1 into x2 , and x3 into x4 to arrive at a simpler 2-dimensional system. Making the
substitutions, and filling in the parameter values, the differential equations then
reduce to
where r3 is the stress level for which the fixed points should be computed.
We now solve these systems for a number of stress values (r3 ) using Mathe-
matica (Wolfram Research, Inc., 2019). This way, we computed the fixed points
shown in Table 5.5, which are displayed in panel (a) of Figure 5.1 in Section 5.2.1.
161
5. Recovering Bistable Systems from Psychological Time Series
Table 5.5: Fixed points of the emotion model for different values of stress (rows), rounded to two
decimals. The 2nd and 3rd columns refer to the fixed points of the healthy fixed points for positive
and negative emotions; the 4th and 5th columns reger to the unhealthy fixed points; and the last two
columns refer to the unstable fixed point.
162
5.B. Mean-Switching Hidden Markov Model
163
5. Recovering Bistable Systems from Psychological Time Series
2500000
●
2000000
1500000
BIC ●
1000000 ●
●
500000 ●
●
●
●
● ●
0
1 2 3 4 5 6 7 8 9 10
Number of States
Figure 5.15: The figure depicts BIC values as a function of the number of states K, for HMMs fitted
to the ideal data.
a Gaussian mixture. However, in the present case the data is generated from
a bistable dynamical system. This failed attempt at model selection based on
statistical models again highlights the problems of using misspecified statistical
models to make inferences about dynamical systems models.
10
8
Emotion Strength
Cheerful
6
Content
Anxious
4 Sad
−2
0 5 10 14
Days
Figure 5.16: A time series of two weeks generated from the HMM estimated in Section 5.3.2.
164
5.C. Data Generated from Estimated Models
The generated time series looks similar to the original data in that it switches
between the two fixed points at around (1,6) and (6,1). However, there are also
differences. In the original data there are less switches that lead to a long-lived
change in fixed point, but more switches that are very short-lived. Second, due to
the form of the Mean-Switching HMM, there are no “intermediate” observations
leading from one fixed point to the other. These observations exist in the original
time series (see panel (b) in Figure 5.2).
15
Emotion Strength
10
Cheerful
Content
Anxious
5 Sad
−5
0 5 10 14
Days
Figure 5.17: A time series of two weeks generated from the VAR(1) model estimated in Section 5.3.4.
The generated data does not show bistability, which is expected because the
VAR(1) model exhibits only a single fixed point. What looks approximately like
oscillating behavior is a result of the high auto-regressive effects present in the
estimated VAR(1) model: given a stochastic input, the high auto-regressive ef-
fects ensure that the system is slow to eventually return to equilibrium. This
oscillating behavior is also evident in the eigenvalues of Φ, which consist of one
complex conjugate pair (Strogatz, 2015).
165
5. Recovering Bistable Systems from Psychological Time Series
10
8
Emotion Strength
Cheerful
6
Content
Anxious
4 Sad
−2
0 5 10 14
Days
Figure 5.18: A time series of two weeks generated from the TVAR(1) model estimated in Section 5.3.5.
−0.05 −0.05
−0.05 −0.05
Anxious Sad Anxious 0.02 Sad
Figure 5.19: Residual partial correlation networks for both regimes in the TVAR model described in
Section 5.3.5 in the main text.
166
5.E. Differential Equation Model Building
validation. First, the given dataset is randomly partitioned into ten mutually
exclusive training and test sets. Second, for each partitioned dataset, regression
models A through G, (defined by the expression in the second column of Table
5.6) are fit to the training set four times, once each of the four outcome variables
dxiˆ/dt, ∀i ∈ {1, 2, 3, 4}. Third, the resulting parameters are then used to predict the
values of the outcome variable in the test set dxiˆ/dt. The variance of the resulting
residuals V AR(dxi /dt − dxiˆ/dt) is then divided by the variance of the outcome
variable in the test set, V AR(dxi /dt) yielding an out-of-bag variance explained
for variable i based on model m in partition k, R2i,k,m . Averaging the explained
variance across each of the partitions yields an average explained variance for
variable i in model m, R2i,m , and averaging this number across all four outcome
variables yields the average out-of-bag explained variance for model m.
dxi,t
Model dt ∼ a + ri xi + . . . q R2
P
A j,i rj xj 5 .04464
P Pp
B Rij j + j Cij xj xi
x 9 .06874
P j,i Pp
C j,i Rij xj + (j,k) βj xj xk 15 .06870
P Pp Pp 3
D j,i Rij xj + (j,k) βj xj xk + j γj xj 19 .06871
P Pp Pp 3 P p
E j,i Rij xj + (j,k) βj xj xk + j γj xj + j,k,l ζj (xj xk xl ) 23 .06870
P Pp Pp 3 Pp
F j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl ) 35 .06860
P Pp Pp 3 Pp
j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl )
G Pp 70 .06846
+ (j,k,l,m) ηj (xj xk xl xm )
Table 5.6: Model fit results for each of the seven models described in text in Section 5.3.6 for the ideal
dataset. The second column gives the model equation for each variable, q denotes the number of pa-
rameters estimated per univariate regression model, and the final column indicates R2 , the explained
variance, as calculated based on the prediction error on a hold-out set, using 10-fold cross-validation.
167
5. Recovering Bistable Systems from Psychological Time Series
Table 5.7: Full parameter estimates, standard errors and p-values for Model B in Section 5.3.6, for
the DE model fit to ideal data.
dxi,t
Model ∼ a + ri xi + . . .
dt q R2
P
A j,i rj xj 5 0.13991
P Pp
B j,i Rij xj + j Cij xj xi 9 0.16827
P Pp
C j,i Rij xj + (j,k) βj xj xk 15 0.16928
P Pp Pp 3
D j,i Rij xj + (j,k) βj xj xk + j γj xj 19 0.19455
P Pp Pp 3 P p
E j,i Rij xj + (j,k) βj xj xk + j γj xj + j,k,l ζj (xj xk xl ) 23 0.19801
P Pp Pp 3 Pp
F j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl ) 35 0.19940
P Pp Pp 3 Pp
j,i Rij xj + (j,k) βj xj xk + j γj xj + (j,k,l) ζj (xj xk xl )
G Pp 70 0.20420
+ (j,k,l,m) ηj (xj xk xl xm )
Table 5.8: Model fit results for each of the seven models described in text, for the ESM time series,
described in Section 5.4. The second column gives the model equation for each variable, q denotes the
number of parameters estimated per univariate regression model. The final two columns indicate R2 ,
the explained variance, as calculated based on the prediction error on a hold-out set, using 10-fold
cross-validation, for the snapshot ESM data and the mean-aggregated ESM data, respectively. R2 for
Model G in the mean-aggregated ESM data case was not available due to multicollinearity problems
encountered when fitting the model.
168
5.E. Differential Equation Model Building
Table 5.9: Full parameter estimates, standard errors and p-values for Model G in Section 5.4.4, for
the DE model fit to the emulated ESM data.
169
5. Recovering Bistable Systems from Psychological Time Series
Density
Density
Density
0.3 0.3 0.3 0.3
0.2 0.2 0.2 0.2
0.1 0.1 0.1 0.1
0.0 0.0 0.0 0.0
0 2 4 6 8 0 2 4 6 8 0 2 4 6 8 0 2 4 6 8
Figure 5.20: The histograms of the emotion intensity of the four modeled emotions Cheerful, Content,
Anxious and Sad, for the ESM data
8 (a) 8 (b)
6 6
ρ = −0.97
Anxious
Content
4 4
2 ρ = 0.98 2
0 0
0 2 4 6 8 0 2 4 6 8
Cheerful Cheerful
−0.97 −0.23
−0.97 −0.24
Anxious 0.98 Sad Anxious 0.52 Sad
Figure 5.21: Panel (a) shows the relationship between Content and Cheerful, two emotions with the
same valence, at the same time point The red line indicates the best fitting regression model, for ESM
time series. Similarly, panel (b) shows the relationship between Anxious and Content, two emotions
with different valence. Panel (c) displays the correlation matrix as a network, and panel (d) displays
the partial correlation matrix as a network.
170
5.F. Additional Results ESM Time Series
State 1 State 2
8
Emotion Strength
6
Cheerful
Content
Anxious
4 Sad
0 5 10 14
Days
Figure 5.22: Time series of the four emotion variables, also shown in panel (a) of Figure 5.2, with
background color indicating whether a given time point is assigned to the first or second component
of the mean-switching HMM estimated from the ESM dataset.
171
Chapter 6
Modeling Psychopathology:
From Data Models to Formal
Theories
Abstract
Over the past decade there has been a surge of empirical research inves-
tigating mental disorders as complex systems. In this paper, we investigate
how to best make use of this growing body of empirical research and move
the field toward its fundamental aims of explaining, predicting, and con-
trolling psychopathology. We first review the contemporary philosophy
of science literature on scientific theories and argue that fully achieving
the aims of explanation, prediction, and control requires that we construct
formal theories of mental disorders: theories expressed in the language of
mathematics or a computational programming language. We then investi-
gate three routes by which one can use empirical findings (i.e. data models)
to construct formal theories: (a) using data models themselves as formal
theories, (b) using data models to infer formal theories, and (c) comparing
empirical data models to theory-implied data models in order to evaluate
and refine an existing formal theory. We argue that the third approach is
the most promising path forward and conclude by expanding on this ap-
proach, proposing a framework for theory construction that details how to
best use empirical research to generate, develop, and test formal theories
of mental disorders.
This chapter has been adapted from: Haslbeck, J. M. B.*, Ryan, O.*, Robinaugh, D.J.*, Waldorp,
L.J. and Borsboom, D. (under review). Modeling Psychopathology: From Data Models to Formal
Theories. Pre-print: https://psyarxiv.com/jgm7f/. Author contributions: JMBH, OR and DJR are
considered joint first authors and contributed equally to this project. LJW and DB helped develop the
ideas in the project, discussed progress and provided textual feedback.
173
6. Modeling Psychopathology: From Data Models to Formal Theories
6.1 Introduction
Mental disorders are complex phenomena: highly heterogeneous and massively
multifactorial (e.g., Kendler, 2019). Confronted with this complex etiological and
ontological picture, researchers have increasingly called for approaches to psy-
chiatric research that embrace this complexity (Gardner & Kleinman, 2019). The
“network approach” to psychopathology addresses these calls, conceptualizing
mental disorders as complex systems of interacting symptoms (e.g., Borsboom &
Cramer, 2013; Schmittmann et al., 2013; Borsboom, 2017). From this perspec-
tive, symptoms are not caused by an underlying disorder, rather the symptoms
themselves and the causal relations among them constitute the disorder.
In recent years, empirical research within the network approach literature
has rapidly grown (for reviews see e.g., Robinaugh, Hoekstra, et al., 2019; Con-
treras et al., 2019). Most of this work employs statistical models that allow re-
searchers to study the multivariate dependencies among symptoms, thereby pro-
viding rich information about the relationships among those symptoms. How-
ever, this quickly expanding empirical literature has raised a critical question:
how can we best make use of this growing number of empirical findings to ad-
vance the fundamental aims of psychiatric science? This problem is not unique to
the network approach. Psychiatry has produced countless empirical findings, yet
genuine progress in our efforts to explain, predict, and control mental disorders
has remained stubbornly out of reach.
In this paper, we will argue that empirical research can best advance these
aims by supporting the development of scientific theories. We will begin in Sec-
tion 6.2 by discussing the nature of scientific theories and how they achieve the
explanation, prediction and control sought by psychiatric science. We will argue
that to fully achieve these aims, psychiatry requires theories formalized as math-
ematical or computational models. In Section 6.3, we will explore how models
estimated from data can best be used to develop formal theories. We examine
three possible routes from data model to formal theory: first, treating data mod-
els themselves as formal theories; second, drawing inferences from data models
to generate a formal theory; and third, using data models to develop formal the-
ories with an abductive approach. We will argue that the third approach is the
most promising path forward. In Section 6.4, we will expand on this approach
and propose a framework for theory construction, detailing how best to use em-
pirical research to advance the generation, development, and testing of scientific
theories of mental disorders.
174
6.2. The Nature and Importance of Formal Theories
C2 C4
V2 V4
Figure 6.1: The figure illustrates the concepts target system, theory, data and data model. The target
system is the system consisting of interacting components that gives rise to phenomena. Phenomena
are robust features of the world captured by data models. Theories represent the structure of the
target system, proposing a set of components C and the relations among them and positing that they
give rise to the phenomena. Data for variables V are obtained by probing the target system.
175
6. Modeling Psychopathology: From Data Models to Formal Theories
and syndromes. For example, researchers seek to explain the tendency for some
individuals to experience panic attacks and the tendency for recurrent panic at-
tacks to be accompanied by persistent worry about those attacks and avoidance
of situations in which they may occur (Spitzer, Kroenke, & Williams, 1980). The
target system in psychiatric research comprises the components of the real world
that give rise to these symptoms and syndromes, and may include genetic, neuro-
biological, physiological, emotional, cognitive, behavioral or social components.
Psychiatric theories aim to represent these target systems, positing a specific set
of components and relationships among them that give rise to the phenomena
of interest. For example, researchers have generated numerous theories of panic
disorder, specifying a set of components that they believe interact to give rise to
panic attacks and panic disorder. Among these, perhaps the most influential is
Clark’s cognitive model of panic attacks, which posits that “if [stimuli] are per-
ceived as a threat, a state of mild apprehension results. This state is accompanied
by a wide range of body sensations. If these anxiety-produced sensations are in-
terpreted in a catastrophic fashion, a further increase in apprehension occurs.
This produces a further increase in body sensations and so on round in a vicious
circle which culminates in a panic attack” (Clark, 1986). This cognitive theory of
panic attacks specifies components (e.g., bodily sensations and a state of appre-
hension) and the relations among them (e.g., the “vicious cycle” of positive causal
effects), positing that this is the target system that gives rise to panic attacks.
Because theories represent the target system, we can reason from theory in
order to draw conclusions about the target system. It is this capacity for surroga-
tive reasoning (Swoyer, 1991) that allows theories to explain, predict, and control.
For example, we can explain the rise and fall of predator and prey populations in
the real world by appealing to the relationships between components specified
in mathematical models representing these populations (H. I. Freedman, 1980;
Nguyen & Frigg, 2017). We can predict what will occur when two atoms collide
by deriving the expected outcome from models of particle physics (Higgs, 1964).
We can determine how to intervene to prevent panic attacks by appealing to the
relationships posited in the cognitive model of panic attacks, determining that
an intervention modifying a patient’s “catastrophic misinterpretations” should
prevent the “vicious cycle” between arousal and perceived threat, thereby cir-
cumventing panic attacks (Clark, 1986). It is this ability to support surrogative
reasoning that makes theories such powerful tools.
176
6.2. The Nature and Importance of Formal Theories
177
6. Modeling Psychopathology: From Data Models to Formal Theories
0.25
Arousal Perceived Escape
(A) Threat (E) 0.00
(T)
0 3 6 9
Figure 6.2: The left panel displays the key components of the theory proposed by Robinaugh,
Haslbeck, et al. (2019) at play during panic attacks: Arousal, Perceived Threat, Escape Behavior and
arousal schema. The arrows indicate the direct causal relationships which are posited to operate
between these components in the formal theory. The middle panel displays the formal theory that
specifies the precise nature of the relations among these components. The right panel depicts the
simulated behavior implied by the theory.
By specifying the structure of the theory in this way, we are able to solve
the system numerically, thereby deducing the theory’s predictions about how the
178
6.2. The Nature and Importance of Formal Theories
target system will behave. For example, the theory shows that when the effect of
Arousal on Perceived Threat is sufficiently strong, the positive feedback between
these components is sufficient to send the system into runaway positive feed-
back, producing the characteristic surge of arousal, perceived threat, and escape
behavior that we refer to as a panic attack (see right panel in Figure 6.2). As this
example illustrates, specifying the theory as a computational model substantially
strengthens our ability to deduce the behavior implied by the theory. A full re-
alization of a theory’s usefulness thus all but requires that theory be formalized.
For that reason, we believe the ultimate goal of psychiatric research should not
only be the production of theories, but the production of formal theories.
179
6. Modeling Psychopathology: From Data Models to Formal Theories
paper. This role is especially noteworthy in the context of the data models most
commonly used in the network approach literature: the Ising model, the Gaus-
sian Graphical Model, and the Vector Autoregressive model. In Section 6.3 we
will describe each of these models in more detail, but here it is sufficient to note
their most salient feature: these analyses estimate the structure of relationships
among a set of variables; specifically, the structure of conditional dependence re-
lationships (see Figure 6.1; Top Right). There is a strong intuitive appeal to these
analyses as they seem to hold the promise of directly informing the very thing
we are after: the structure of relations among components of the mental disorder
(see Figure 6.1; Top Left). In Section 6.3, our overarching aim will be to critically
evaluate that promise and determine how best to use (network) data models to
guide the development of theories about specific mental disorders.
180
6.3. Identifying Formal Theories from Data
Figure 6.3: The figure provides an overview of three routes to developing formal theories using data
models. In the left panel, data models are treated as formal theories. In the middle panel, data models
are used to draw inferences about the target system and, thereby, to generate formal theories of that
system. In the right panel, data models used to develop formal theories by deducing implied data
models and comparing them with empirical data models.
The third possible route puts formal theories at the heart of theory develop-
ment. From this perspective, research is carried out by first generating an initial
formal theory. From this formal theory we simulate data which we use to ob-
tain the theory-implied data model. We subsequently compare the implied data
model with the empirical data model, and adapt the formal theory based on the
discrepancy between the two. This route thus leverages the “immense deduc-
tive fertility” of formal theories to make precise predictions that clarify how the
model must be revised to be brought in line with empirical data (Meehl, 1978).
From this perspective, formal theory is not only the ultimate goal of the research
process, but also plays an active role in theory development. We evaluate this
route in Section 6.3.3 by deriving predicted data models from a formal theory of
Panic disorder, and showing how the model can be improved by comparing the
predicted data models to empirical data models.
181
6. Modeling Psychopathology: From Data Models to Formal Theories
of three data models: the VAR model, the GGM, and the Ising model (Section
6.3.1.2).
182
6.3. Identifying Formal Theories from Data
Bringmann et al., 2013; Pe et al., 2015; A. J. Fisher et al., 2017; Snippe et al., 2017;
Groen et al., 2019). The estimated lagged effects of the VAR models indicate con-
ditional dependence relationships among variables over time. The dynamic of
the VAR model is such that the variables are perturbed by random input (typ-
ically Gaussian noise) and the variables return to their means, which represent
the single stable state of the system.
As depicted in Figure 6.4, the VAR model is able to represent some key
characteristics likely to be present in mental disorder target systems. Most
notably, it allows for feedback loops. Variables can affect themselves both di-
rectly (e.g., Xt → Xt+1 ), or via their effects on other variables in the system (e.g.,
Xt → Yt+1 → Xt+2 ). The VAR model also allows for asymmetric relationships,
since the effect Xt → Yt+1 does not have to be the same effect as Yt → Xt+1 in
direction or magnitude. However, because the lag-size (i.e. the distance between
time points) is fixed and consistent across all relationships, the VAR model does
not allow for different time scales. Moreover, because the VAR model only in-
cludes relations between pairs of variables, it is unable to represent higher-order
interactions involving more than two variables. Finally, the VAR model has a
single stable state defined by its mean vector and thus cannot represent multiple
stable states of a system, such as a healthy state and unhealthy state.
*
Figure 6.4: The figure shows whether the five properties of mental disorders discussed above can
be represented by the three most popular network data models, the VAR model, the GGM, and the
Ising model with Glauber dynamics. Note that there is a check mark at feedback loops for GGMs
because one could in principle endow the GGM with a dynamic similar to the Ising model, which
would essentially lead to a restricted VAR model but with symmetric relations. The asterisk is present
because this endowment of dynamics is not done in practice.
183
6. Modeling Psychopathology: From Data Models to Formal Theories
The second model we will consider is the Gaussian Graphical Model (GGM).
The GGM linearly relates pairs of variables in either cross-sectional (Haslbeck &
Fried, 2017) or time series data (Epskamp, Waldorp, et al., 2018). In the case of
time series data the GGM models the relationships between variables at the same
time point. Because it does not model any dependency across time, it is typically
not considered a dynamic model and, thus, could not be used to represent the
behavior of a mental disorder target system as it evolves over time. In principle
the GGM could be augmented by a dynamic rule similar to one commonly used
with the Ising model (i.e. “Glauber dynamics”, see below). However, in that
case, the GGM would become a model similar to, but more limited than, the
VAR model described above (e.g., it would be limited to symmetric relationships).
Accordingly, the GGM is similarly unable to represent key features we expect to
observe in a mental disorder target system.
The final model we will consider is the Ising model. The Ising model again
represents pairwise conditional dependence relations between variables (Ising,
1925), however, it is a model for multivariate binary data. While the original Ising
model does not model dependencies over time, it can be turned into a dynamic
model by augmenting it with Glauber dynamics (Glauber, 1963).3 Like the VAR
model, the Ising model is able to represent feedback loops. Moreover, due to its
non-linear form it is able to exhibit multiple stable states (and the behavior that
accompanies such stable states, such as hysteresis and sudden shifts in levels of
symptom activation, see e.g., Cramer et al., 2016; Lunansky, van Borkulo, & Bors-
boom, 2019; Dalege et al., 2016). It is perhaps not surprising then, that the Ising
model is used as a theoretical model across many sciences (Stutz & Williams,
1999), and to our knowledge, is the only of the three data models examined here
that has been used as a formal theory of a mental disorder target system (Cramer
et al., 2016). Unfortunately, the Ising model falls short in its ability to repre-
sent the remaining characteristics likely to be present in mental disorders. The
relationships in the Ising model are exclusively symmetric; with the standard
Glauber dynamics, there is only a single time scale; and the Ising model includes
exclusively pairwise relationships, precluding any representation of higher-order
interactions.
184
6.3. Identifying Formal Theories from Data
185
6. Modeling Psychopathology: From Data Models to Formal Theories
nects parameters in particular data models to the structure of the target system.
For some simple systems, such a rule is available. For example, if the target sys-
tem can be represented as a Directed Acyclic Graph (DAG), then under certain
circumstances its structure can be inferred from conditional (in)dependence re-
lations between its components: Conditional independence implies causal inde-
pendence, and conditional dependence implies either direct causal dependence
or a common effect (Pearl, 2009; Ryan, Bringmann, & Schuurman, 2019). How-
ever, it is generally unclear how we can use the parameters of typical data models
to make inferences about the types of non-linear dynamic systems we expect in
a psychiatric context (although Mooij, Janzing, & Schölkopf, 2013 and Forré &
Mooij, 2018 have established some links in this regard). The consequences of
this are twofold. First, any inference from data model to target system must rely
instead on some simplified heuristic(s) in an attempt to approximate the link be-
tween the two. Second, it is unclear how well the combination of common data
models and simple heuristics perform in allowing us to make inferences about
the target system.
In this section, we evaluate whether the three data models introduced above
can be used to make inferences about mental disorder target systems. To do this,
we treat the Panic Model discussed in Section 6.2 as the data-generating target
system and compare the causal structure inferred from the data models to the
true causal structure. To yield these inferences we use a very simple and intu-
itive set of heuristics: a) if two variables are conditionally dependent in the data
model, we will infer that the corresponding components in the target system are
directly causally dependent; b) if there is a positive linear relationship, we will
infer that the causal relation between the corresponding components is positive
(i.e. reinforcing); c) if there is a negative linear relationship, we will infer that the
causal relationship among components is negative (i.e. suppressing).
186
6.3. Identifying Formal Theories from Data
erate data from the target system (See Appendix 6.A). Specifically, we use four
weeks of minute-to-minute time-series data for 1000 individuals. These individ-
uals differ in their initial value of Arousal Schema, with the distribution chosen
so that the proportion of individuals for whom a panic attack is possible was
equivalent to the lifetime history prevalence of panic attacks in the general popu-
lation (R. R. Freedman, Ianni, Ettedgui, & Puthezhath, 1985). For the VAR model
analysis, we create a single-subject experience-sampling-type dataset by choos-
ing the individual who experiences the most (16) panic attacks in the four-week
period. To emulate ESM measurements, we divide the four week period into 90-
minute intervals, taking the average of each component in that interval, yielding
448 measurements. For the GGM analysis, we create a continuous cross-sectional
dataset by taking the mean of each component for each individual over the four
weeks. For the Ising model analysis, we obtain cross-sectional binary data by tak-
ing a median split of those same variables. The resulting VAR, GGM and Ising
model networks are displayed in Figure 6.5 panels (b), (c) and (d), respectively.4
PT Arousal
AS Context
Avoidance Escape
(b) VAR Model (c) Gaussian Graphical Model (d) Ising Model
Figure 6.5: Panel (a) shows the true model in terms of local dependencies between components; panel
(b) shows the VAR model estimated from ESM data sampled from the true model; panel (c) shows
the GGM estimated from the cross-sectional data of 1000 individuals, generated from the true model;
panel (d) shows the Ising model estimated on the same data after being binarized with a median split.
Solid edges indicate positive relationships, dotted indicate negative relationships. For panels (b) to
(d), the widths of edges are proportional to the absolute value of the corresponding parameter. Note
that in panel (b) we do not depict the estimated auto-regressive parameters as the primary interest is
in inferring relationships between variables.
4 Note that in the Ising model the parameter estimates are somewhat unstable due to near-
deterministic relationships between some binarized variables.
187
6. Modeling Psychopathology: From Data Models to Formal Theories
188
6.3. Identifying Formal Theories from Data
For example, consider the relationships between Perceived Threat and Arousal.
The VAR model (panel (b) in Figure 6.5), identifies a negative lagged relationship
from Perceived Threat to Arousal in the data generated by the target system. Yet
in the target system, this effect is positive. This “discrepancy” occurs because of
a very specific dynamic between these components: After a panic attack (i.e. a
brief surge of Perceived Threat and Arousal) there is a “recovery” period in which
arousal dips below its mean level for a period of time. As a result, when we av-
erage observations over a 90 minute window, a high average level of Perceived
Threat is followed by a low average level of Arousal whenever a panic attack oc-
curs. That same property of the system produces the observed findings for the
GGM and Ising Model through yet another mapping (for details, see Appendix
6.B).
As this example illustrates, the mapping between target system and data
model is intricate, and it is unlikely that any simple heuristics can be used suc-
cessfully to work backwards from the data model to the exact relationships in
the target system. We can expect this problem to arise whenever we use rela-
tively simple statistical models to directly infer characteristics or properties of a
complex system (c.f. the problem of under-determination or indistinguishabil-
ity; Eberhardt, 2013; Spirtes, 2010). Indeed, the same problem arises even for
simpler dynamical systems when analyzed with more advanced statistical meth-
ods (e.g., Haslbeck & Ryan, 2019). Of course, in principle, it must be possible to
make valid inferences from data and data models to some properties of a target
system using a more principled notion of how one maps to the other. For exam-
ple, under a variety of assumptions, it has been shown that certain conditional
dependency relationships can potentially be used to infer patterns of local causal
dependencies in certain types of dynamic system (Mooij et al., 2013; Bongers &
Mooij, 2018; Forré & Mooij, 2018). However, the applicability of these methods
to the type of target system we expect to give rise to psychopathology (see Sec-
tion 6.3.1) is as yet unclear and even under the strict assumptions under which
they have been examined, these methods still do not recover the full structure of
the target system.5 This means that the intricacy of the mapping between target
system and data model currently precludes us from making reliable inferences
about the target system. Accordingly, we cannot use those inferences to build
formal theories.
5 Specifically, Mooij et al. (2013) and Bongers and Mooij (2018) have shown that cyclic causal mod-
els can be conceptualized as encoding causal dependencies between the equilibrium positions of de-
terministic differential equations and differential equations with random initial values. Forré and
Mooij (2018) formally link the conditional dependencies between equilibrium position values to the
causal dependencies in these cyclic causal models using a considerably more complex mapping rule
than that which holds for DAGs. Their applicability to the current context is limited in the sense
that 1) to our knowledge these rules have not been extended to dynamic systems with time-varying
stochastic terms (SDEs) as we would expect to see in complex psychological systems (and on which
the Panic Model is based), and 2) the use of these methods is reliant on data that reflects equilibrium
positions. Future developments in this area may prove to yield useful tools for psychological theory
development however, and we consider this area to be ripe for future research beyond the current
paper.
189
6. Modeling Psychopathology: From Data Models to Formal Theories
190
6.3. Identifying Formal Theories from Data
PA PA α = -6.14
α = −170.08
Av 625.85 PC Av 23.72 PC
Estimate Estimate
Av PA PC Av PA PC
Simulated 0 0 0 0 0 0
0 0 0 CPES Data 0 1 1
Data 1 1 1 1 1 1
0 0 0 0 0 0
1 1 1 1 1 0
Emulated Measurement
Measurement
Arousal
Schema
Panic Disorder
Target System
Represents
Avoid Arousal Perceived Escape
Threat
Context
Panic Attacks
Formal Theory
Figure 6.6: Illustration of the third route to formal theories. We take the Panic Model discussed in
Section 6.2 as our formal theory, representing the unknown target system that gives rise to panic
disorder. To obtain an implied data model from this theory, we first formalize how the components
of the theory produce the data of interest, emulating the measurement process. With this in place,
we can simulate data from the model in the form of cross-sectional binary symptom variables. We
obtain the theory-implied Ising Model by estimating it from these simulated data (top-left corner).
To estimate the empirical Ising Model (top-right corner) we make use of empirical measurements of
binary symptom variables from the CPES dataset.
intra-individual time series data for multiple individuals (as described in Ap-
pendix 6.A) and so we need to define how cross-sectional symptom variables can
be extracted from those time series. We specify that Recurrent Panic Attacks (PA
= 1) are present for an individual in our simulated data if there are more than
three panic attacks in the one month observation period. Persistent Concern is
191
6. Modeling Psychopathology: From Data Models to Formal Theories
determined using the average levels of jointly experienced arousal and perceived
threat (i.e. anxiety) following a panic attack. If an individual has a panic attack,
and their average anxiety following a panic attack exceeds a threshold determined
by “healthy” simulations (i.e. those without panic attacks), they are classified as
having Persistent Concern (PC = 1). Avoidance is defined similarly, with this
symptom present if an individual has a panic attack, and their average levels of
avoidant behavior following that attack are higher than we would expect to see
in the healthy sample. A more detailed account of how we generated these data
can be found in Appendix 6.C. This simulated cross-sectional data was used to
estimate the implied Ising Model (top left-hand corner, Figure 6.6).6
We obtained the corresponding empirical Ising model (right-hand column of
Figure 6.6) using the publicly available Collaborative Psychiatric Epidemiology
Surveys (CPES) 2001-2003 (Alegria, Jackson, Kessler, & Takeuchi, 2007). The
CPES is a nationally representative survey of mental disorders and correlates
in the United States, with a total sample size of over twenty thousand partici-
pants (of which n = 11367 are used in the current analysis; for details see Ap-
pendix 6.C). The CPES combines more than 140 items relating to panic attacks
and panic disorder, with a diagnostic manual describing how these items can be
re-coded into binary symptom variables reflecting Recurrent Panic Attacks, Per-
sistent Concern and Avoidance. PA is present if the participant reported more
than three lifetime panic attacks. PC is present if, following an attack, the partic-
ipant experienced a month or more of persistent concern or worry. Av is present
if the participant reports either a month of avoidance behavior following an at-
tack, or a general avoidance of activating situations in the past year.
192
6.3. Identifying Formal Theories from Data
PC Av Av
0 1 0 1 0 1
Empirical
PC
PA
PA
PC Av Av
0 1 0 1 0 1
Simulated
PC
PA
PA
Figure 6.7: Contingency tables showing percentages for each pair of symptom variables (one per col-
umn) for the empirical data (top row) and simulated data (bottom row). The CPES contingency tables
are based on nCP ES = 11367 observations. The simulated dataset contains nsim = 1000 observations.
193
6. Modeling Psychopathology: From Data Models to Formal Theories
194
6.4. An Abductive Approach to Formal Theory Construction
Abduction
Compare
Develop
Implied &
Theory
Empirical
Deduction
Figure 6.8: Flowchart depicting the process of developing a formal theory with the abductive ap-
proach put forward in this section. In the theory generation step we first establish the phenomenon
(Section 6.4.1.1) and then generate an initial verbal theory (Section 6.4.1.2) which is subsequently for-
malized (Section 6.4.1.3). In the second step (Section 6.4.2) the theory is validated by testing whether
it is consistent with existing empirical findings that are not part of the core phenomenon. If the for-
mal theory is not consistent with some findings, it is adapted accordingly. If these adaptations lead
to a “degenerative” theory (Meehl, 1990) we return to the first step; otherwise we continue to the
final step, in which we test the formal theory using risky predictions (Section 6.4.3). If many tests are
successful, we tentatively accept the theory. If not, the theory must either be adapted (step two) or a
new theory generated (step one).
195
6. Modeling Psychopathology: From Data Models to Formal Theories
196
6.4. An Abductive Approach to Formal Theory Construction
(Borsboom, 2017). For example, as we have seen in Section 6.2, the target sys-
tem could consist of Arousal, Perceived Threat, Avoidance and other symptoms
of panic attacks or panic disorder. Having identified the relevant components we
next specify the posited relations among them. For example, specifying that Per-
ceived Threat leads to Arousal. Within the domain of the network approach, this
second step will typically entail specifying causal relations among symptoms or
momentary experiences (e.g., thoughts, emotions, and behavior).
Notably, in psychiatry, we do not necessarily need to rely on creative insight
about the components and relations among them in order to generate an initial
theory. There are already a plethora of verbal theories about mental disorders. If
the initial verbal theory is well supported and specific, it will lend itself well to
formalization and subsequent theory development. However, even poor verbal
theories can be a useful starting point to developing a successful formal theory
(Wimsatt, 1987; Smaldino, 2017).
197
6. Modeling Psychopathology: From Data Models to Formal Theories
First, empirical research can inform specification of components and the re-
lations among them. For example, one could use the finding that sleep quality
predicts next-day affect, but daytime affect does not predict next-night sleep (de
Wild-Hartmann et al., 2013) to constrain the set of plausible relationships be-
tween those two variables in the formal theory. There could also be empirical data
on the rate of change of variables, for example, Siegle, Steinhauer, Thase, Stenger,
and Carter (2002) and Siegle, Steinhauer, Carter, Ramel, and Thase (2003) have
shown that depressed individuals exhibit longer sustained physiological reac-
tions to negative stimuli than healthy individuals, a finding which is echoed in
self-report measures of negative affect (Houben et al., 2015).
Second, we can possibly derive reasonable scales for variables and relation-
ships between variables from basic psychological science. For example, classical
results from psychophysics show that increasing the intensity of stimuli in almost
all cases leads to a nonlinear response in perception (e.g., Fechner, Howes, & Bor-
ing, 1966): When increasing the volume of music to a very high level, individuals
cannot hear an additional increase.
Third, in many cases we can use definitions, basic logic, or common sense to
choose formalizations. For example, by definition emotions should change at a
time scale of minutes (Houben et al., 2015), while mood should only change at
a time scale of hours or days (Larsen, 2000). And we can choose scales of some
variables using common sense, for example one cannot sleep less than 0 and more
than 24 hours a day, and heart rate should be somewhere between 50 and 180.
Fourth, we could use an existing formal model of another target system,
which we expect to have a similar structure as the target system giving rise to
the phenomenon of interest. This approach is called “analogical modeling”. For
example, Cramer et al. (2016) formulated a model for interactions between symp-
toms of Major Depression using the Ising model, which was originally formu-
lated to model magnetism on an atomic level (Ising, 1925). Similarly, Fukano
and Gunji formulated a model for interactions among core components of panic
attacks using a Lotka-Volterra model originally formulated to represent predator-
prey relationships (Fukano & Gunji, 2012).
Fifth, it is also important to note that there are methods by which we can
potentially estimate the parameters for a formal theory from empirical data.8
These approaches require considerable development of the formal theory (e.g.,
the form of a differential equation), suitable data (typically intensive longitudi-
nal data), and a clear measurement model relating observed variables to theory
components (as we did in Section 6.3.3). Accordingly, this approach already re-
quires considerable progress in generating a formal theory and may be limited
by practical considerations. Nonetheless, it remains a valuable resource that, if
successfully carried out, would likely strengthen subsequent efforts at theory de-
velopment.
8 For example, if the theory is formalized in a system of differential equations, the parameters of
such equations can in principle be estimated from time series data using, amongst others, Kalman
filter techniques and state-space approaches (e.g., Einicke, 2019; Kulikov & Kulikova, 2013; Durbin
& Koopman, 2012). For implementations of these estimation methods see Ou et al. (2019); Carpenter
et al. (2017); King, Nguyen, and Ionides (2015)
198
6.4. An Abductive Approach to Formal Theory Construction
The aim of this initial stage is to generate a formal theory able to explain a
set of core phenomena. As we have emphasized throughout this paper, formal
theories precisely determine the behavior implied by their theory. Accordingly,
explanation in this context means that the theory has demonstrated its ability to
produce the behavior of interest. For example, a theory of panic attacks must be
able to produce sudden surges of arousal and perceived threat; a theory of de-
pression must be able to produce sustained periods of low mood; and a theory of
borderline personality disorder must be able to produce affective instability. We
would note that there are very few theories in psychiatry that have reached this
stage of not merely positing, but demonstrating, that the theory can explain the
phenomena of interest. Accordingly, completing this stage of theory construction
would constitute a significant advance in psychiatric theories. Once a theory has
reached this stage, it is ready for the next stage of theory construction.
199
6. Modeling Psychopathology: From Data Models to Formal Theories
account for, different kinds of data and data models will be appropriate. In gen-
eral, however, more complex data models tend be more powerful tools to tease
apart competing theories. For example, a great many formal theories might be
consistent with a set of means, but it is likely that fewer are consistent with the
means and the conditional relationships between the variables, captured, for ex-
ample, by a GGM or Ising model. In other words, there are a more constrained
number of possible formal theories that may account for more complex data mod-
els, thereby doing more to guide theory development.
There are two important considerations when working through this stage.
First, it is important to consider how much trust to place in the empirical data
at hand. Discrepancies between the empirical data model and theory-implied
model may be due to shortcomings of the formal theory, but they may also be
due to poor measurement, insufficient samples, or poor estimation of the param-
eters of the data model. How do we know when to adapt the formal theory in the
face of some discrepancy? For some guidance on this question, we can draw on
the large literature on model evaluation and model comparison. A straightfor-
ward way to decide whether to adapt the theory would be to derive an implied
model of the adapted theory, and then compare the likelihood of the empirical
data given the initial theory and the adapted theory. In order to decide whether
to accept the adaptation we can use, for instance, a likelihood ratio test or a Bayes
factor. This procedure ensures that we only make adaptations to our theory if we
are certain enough that they actually lead to a better representation of the tar-
get system, and not only the idiosyncratic features of the empirical data at hand.
That is, whether we accept an adaptation of the formal theory depends both on
how large the improvement is, and how certain we are about it (i.e. how large the
sample size is).
Second, it is important to consider the danger of making too many ad hoc re-
visions to the model that account only for idiosyncratic features of a given data
model or, worse, yield new implications that are inconsistent with other empir-
ical findings. For some guidance on this question, we can draw on the litera-
ture on theory evaluation from the philosophy of science literature (Meehl, 1990;
Lakatos, 1976), which would suggest that the theory development phase has two
possible outcomes. If the theory is adapted almost every time it is tested against
empirical data, if those adaptations are making the theory increasingly unwieldy,
and if additional changes are increasingly difficult to make without causing the
theory to be inconsistent with earlier tested empirical findings, the theory can be
considered to be “degenerative” (Meehl, 1990; Lakatos, 1976). In such a situation,
the initial theory was inappropriate and we return to the first step to generate a
different initial formal theory. On the other hand if modifications to a theory ex-
pand, rather than contract, its ability to account for other empirical data beyond
those it was originally introduced to explain, then we can have greater confidence
in those modifications and, in turn, the formal theory. Ultimately, theorists must
strive for a balance between the simplicity of the model and its consistency with
empirical data models.
The aim of the theory development stage is a formal theory that not only
explains the core phenomena of interest, but is also consistent with a range of
200
6.4. An Abductive Approach to Formal Theory Construction
empirical data models. Our confidence in such a theory will grow the more data
models and the more complex the data models that are consistent with theory,
especially if the theory is able to achieve this consistency with minimal ad hoc
adjustments. In other words, if a theory has achieved these aims, we can be in-
creasingly confident that it is a good representation of the target system and can
prepare to subject the theory to more rigorous testing.
201
6. Modeling Psychopathology: From Data Models to Formal Theories
202
6.5. Conclusions
argue for a broader conceptualization that draws on a wider range of criteria for
theory appraisal, with particular emphasis on explanatory breadth. A formal
theory that can explain a range of phenomena should be retained more readily
than a theory that accounts for only a narrow set of phenomena. Nonetheless,
we believe that any failure of a strong test should be taken as a serious challenge
to the theory that, at a minimum, warrants careful consideration about how to
proceed.
If a theory passes a strong test, it is corroborated, with the strength of corrob-
oration proportional to the strength of the test. Notably, because strong tests all
but require the evaluation of predictions made by the theory, a theory that has
passed several such tests will have demonstrated a strong capacity for support-
ing prediction. Accordingly, a theory that has moved from generation, through
development, and testing will emerge well equipped to support not only the ex-
planation, but also the prediction and control of mental disorders.
6.5 Conclusions
In this paper, we have argued that psychiatry needs formal theories and we have
examined how data models can best inform the development of such theories.
We focused especially on the network approach to psychopathology and con-
sidered three possible routes by which conditional dependence networks may
inform formal theories about how mental disorders operate as complex systems.
We found that these data models were not themselves capable of representing the
structure we presume will be needed for a theory of mental disorders. Perhaps
more surprisingly, we also found that we were unable to draw clear and reli-
able inferences from data models about the underlying system. Together, these
findings suggest that merely gathering data models alone is unlikely to readily
inform a well-developed formal theory. Instead, we found that the most promis-
ing use of empirical data models for theory development was to compare them
to “implied data models” derived from an initial formal theory. In this approach,
formal theories play an active role in their own development, with initial formal-
ized theories being refined over time through ongoing comparison of implied and
empirical data models.
Importantly, our analysis is not a critique of the specific data models we ex-
amined here, nor is it a dismissal of their value. Quite the opposite. We believe
these data models provide rich and valuable information about the relationships
among components of a system. However, our analysis strongly suggests that
the network approach to psychopathology cannot survive on these data models
alone. Formal theory is needed if the network approach is to move toward the ex-
planation, prediction, and control of mental disorders. Indeed, there is growing
recognition that formal theories are needed if we are to avoid problems associated
with conflicting empirical results (i.e. the “Replication Crisis”; Collaboration et
al., 2015) and move toward an accumulation of knowledge in scientific research
(e.g., Muthukrishna & Henrich, 2019; Szollosi & Donkin, 2019; Yarkoni, 2019;
Ioannidis, 2014). Accordingly, as a field psychiatry must grapple not only with
203
6. Modeling Psychopathology: From Data Models to Formal Theories
methods for the collection and analysis of data, but also methods for the genera-
tion and development of formal theories
The research framework we proposed in Section 6.4 is intended to be a first
step toward such a method of theory construction. At the heart of this approach
is the use of formalized initial theories to start a cycle of theory development in
which empirical data informs ongoing theory development and these improved
theories inform subsequent empirical research. Critically, this research frame-
work is not intended to suggest that all researchers must develop expertise in
computational modeling. Data and the detection of robust empirical phenom-
ena are central to psychiatric research in our proposed framework. However, our
framework does suggest that, as a field, psychiatry must do more to develop ex-
pertise in computational modeling within its ranks. We suspect that it will only
be through ongoing collaboration among theorists and empirical researchers that
we will be able to leverage the empirical literature to produce genuine advances
in our ability to explain, predict, and control psychopathology.
204
6.A. Simulated Data from the Panic Model
205
6. Modeling Psychopathology: From Data Models to Formal Theories
Figure 6.9: Panel (a) depicts Arousal during a panic attack, showing the short sharp peak of arousal
levels, followed by a longer recovery period of low arousal, before the system returns to the usual
resting state. The dotted line indicates the mean level of arousal over the observation window (0 -
10 hrs). Panel (b) depicts the state-space plot of Perceived Threat and Arousal at the next measure-
ment occasion, as captured by the emulated ESM study and VAR model. Grey points indicate an
observation window of 90 min in which either part of a panic attack or the following recovery pe-
riod is captured. The solid grey line reflects the marginal lagged relationship. Panel (c) depicts the
cross-sectional marginal relationship between the mean of Arousal and mean of Perceived Threat, as
analyzed in the GGM model. Grey dots indicate individuals who suffer from panic attacks, and black
dots represent “healthy” individuals. The solid grey line shows the negative marginal relationship.
The dotted grey lines indicate the median of both variables, by which the binarized values used in
the Ising model analysis are defined.
In the VAR model in Figure 6.5 (b) in the main text, we observed a strong neg-
ative conditional relationship between Perceived Threat at time t and Arousal
at time t + 1, conditioning on all other variables at time t. The distribution of
these lagged variables is shown in Figure 6.9 (b), with the grey line representing
the also negative marginal relationship. This strong negative cross-lagged rela-
tionship is a direct consequence of the recovery period of Arousal: High values
of Perceived Threat are closely followed by a long period of low Arousal values.
This can be seen in Figure 6.9 (b), where observations over windows in which a
panic attack and recovery period occur are shaded in red. By averaging arousal
values over a window of 90 minutes, the strong positive causal effects operating
locally in time (i.e. over a very short time-interval) are not directly captured, but
instead the VAR(1) model describes correctly describes the negative relationship
between the means of each variable over this window.
In the GGM in Figure 6.5 (c) in the main text, we saw a positive linear re-
lationship between Arousal and Perceived Threat in the estimated GGM. This
dependency indicates that high mean levels of Arousal are associated with high
mean levels of Perceived Threat, conditional on all other variables. We stress the
conditional nature of this relationship, because the marginal relationship between
the two variables is in fact negative as can be seen in Figure 6.9 (c). This nega-
tive marginal relationship comes about by combining two groups of individuals
that have different mean values on both variables. Individuals who experience
panic attacks (grey points) have high average Perceived Threat, but low average
Arousal, due to the long recovery period of Arousal after a panic attack. On the
other hand, individuals who do not experience panic attacks have higher average
values of Arousal, and lower average values of Perceived Threat. When inspecting
206
6.C. Details Empirical vs Simulated Ising Model
the two groups separately, we see that there is a positive linear relationship be-
tween mean Arousal and Perceived Threat in the group without panic attacks; the
group with panic attacks is too small to determine a relationship. Since Escape
and Avoidance behavior only occur after Panic attacks, conditioning on those two
variables amounts to conditioning on whether an individual had panic attacks.
This conditional relationship is then driven mostly by the positive relationship
in the (much larger) group of individuals who have no panic attacks, indicated
by the black dots in Figure 6.9 (c).
Finally, we can explain the weak positive relationship between Arousal and
Perceived Threat in the Ising model (Figure 6.5 (d) in the main text): The levels
of these variables are defined by a median split of their mean values, depicted as
dotted lines in Figure 6.9 (c). Unlike in the GGM, there is a positive marginal rela-
tionship between these binarized variables, as the majority of individuals without
panic attacks (denoted by the black points) end up in the low Perceived Threat
and low Arousal groups (lower left quadrant Figure 6.9 (c)) or high Perceived
Threat and high Arousal groups (upper right quadrant). How then do we end up
with a weakly positive conditional relationship between these two binary vari-
ables? Similarly to the GGM above, it turns out that conditioning on variables
such as Escape behaviour and Avoidance almost entirely separates individuals
into either the low Arousal and low Perceived Threat category (for low Escape
values) or the high Arousal and high Perceived Threat category (for high Escape
value). This means that, once we have conditioned on other variables which have
direct and indirect causal connections to Arousal and Perceived Threat, there is
very little additional information which Arousal can add to predicting Perceived
Threat levels (and vice versa). This produces the weak positive conditional rela-
tionship between Arousal and Perceived Threat, as well as the stronger positive
connections between Avoidance and Perceived Threat.
207
6. Modeling Psychopathology: From Data Models to Formal Theories
2 Persistent Concern (PC): PC is present if, following at least one panic attack,
higher average levels of Anxiety are present than in the healthy population,
as defined by the 90th percentile of average Anxiety in the healthy popula-
tion.
3 Avoidance (Av): Av is present if, following at least one panic attack, higher
average levels of Avoid are present than in the healthy population, as de-
fined by the 90th percentile of average Avoid in the healthy population.
The Ising model of these three symptom variables is fit using the EstimateIsing
function from the IsingSampler package (Epskamp, 2015), that is, using a non-
regularized pseudolikelihood method.
208
6.C. Details Empirical vs Simulated Ising Model
Table 6.1: Description of the three criteria (A, B and C) necessary to code an individual as having one
lifetime panic attack based on items from the CPES survey, based on the CPES diagnostic manual
With this definition of a panic attack in place, we define the three binary
symptoms of panic disorder, following the definitions laid out in the diagonistic
manual for Panic Disorder.
1. Recurrent Panic Attacks (PA). PA is present if participant reports more than
three lifetime occurrences of an unexpected, short, sharp attack of fear or
panic (item PD4 and all three criteria in Table 6.1), more than one of which
is out of the blue (PD17a)
209
6. Modeling Psychopathology: From Data Models to Formal Theories
In coding this, if two out of three PA criteria were present, and the third was
missing, we assigned a positive value to the PA item. The empirical Ising model
was fit using the same procedure as the theory-implied Ising model.
210
References
Aalen, O. O. (1987). Dynamic modelling and causality. Scandinavian Actuarial
Journal, 1987(3-4), 177–190.
Aalen, O. O., Borgan, Ø., Keiding, N., & Thormann, J. (1980). Interaction be-
tween life history events. nonparametric analysis for prospective and retrospec-
tive data in the presence of censoring. Scandinavian Journal of Statistics, 161–
171.
Aalen, O. O., Gran, J., Røysland, K., Stensrud, M., & Strohmaier, S. (2018).
Feedback and mediation in causal inference illustrated by stochastic process
models. Scandinavian Journal of Statistics, 45, 62–86.
Aalen, O. O., Røysland, K., Gran, J., Kouyos, R., & Lange, T. (2016). Can we
believe the DAGs? A comment on the relationship between causal DAGs and
mechanisms. Statistical methods in medical research, 25(5), 2294–2314.
Aalen, O. O., Røysland, K., Gran, J., & Ledergerber, B. (2012). Causality, media-
tion and time: a dynamic viewpoint. Journal of the Royal Statistical Society: Series
A (Statistics in Society), 175(4), 831–861.
Abadir, K. M., & Magnus, J. R. (2005). Matrix Algebra (Vol. 1). Cambridge
University Press.
Angrist, J. D., Imbens, G. W., & Rubin, D. B. (1996). Identification of causal ef-
fects using instrumental variables. Journal of the American Statistical Association,
91(434), 444–455.
213
REFERENCES
Armour, C., Fried, E. I., Deserno, M. K., Tsai, J., & Pietrzak, R. H. (2017). A net-
work analysis of DSM-5 posttraumatic stress disorder symptoms and correlates
in us military veterans. Journal of Anxiety Disorders, 45, 49–59.
Asparouhov, T., Hamaker, E. L., & Muthén, B. (2018). Dynamic structural equa-
tion models. Structural Equation Modeling: A Multidisciplinary Journal, 25(3),
359–388.
Atkinson, K. E. (1989). An introduction to numerical analysis. New York, NY:
John Wiley & Sons.
Bisconti, T., Bergeman, C. S., & Boker, S. M. (2004). Emotional well-being in re-
cently bereaved widows: A dynamical system approach. Journal of Gerontology,
Series B: Psychological Sciences and Social Sciences, 59, 158-167.
Bogen, J., & Woodward, J. (1988). Saving the phenomena. The Philosophical
Review, 97(3), 303–352.
214
REFERENCES
Boker, S. M., Montpetit, M. A., Hunter, M. D., & Bergeman, C. S. (2010). Mod-
eling resilience with differential equations. In P. Molenaar & K. Newell (Eds.),
Learning and development: Individual pathways of change (p. 183-206). Washing-
ton, DC: American Psychological Association.
Boker, S. M., Neale, M., & Rausch, J. (2004). Latent differential equation model-
ing with multivariate multi-occasion indicators. In K. van Montfort, J. H. Oud,
& A. Satorra (Eds.), Recent developments on structural equation models (pp. 151–
174). Dordrecht, the Netherlands: Kluwer Academic.
Boker, S. M., & Nesselroade, J. R. (2002). A method for modeling the intrin-
sic dynamics of intraindividual variability: Recovering parameters of simulated
oscillators in multi-wave panel data. Multivariate Behavioral Research, 37, 127–
160.
Boker, S. M., Staples, A. D., & Hu, Y. (2016). Dynamics of change and change in
dynamics. Journal for person-oriented research, 2(1-2), 34–55.
Bolger, N., & Laurenceau, J.-P. (2013). Intensive longitudinal methods: An intro-
duction to diary and experience sampling research. NY: New York: The Guilford
Press.
Bollen, K. A. (1987). Total, direct, and indirect effects in structural equation
models. Sociological methodology, 37–69.
Bollen, K. A. (1989). Structural Equations with Latent Variables. Oxford, England,
John Wiley & Sons.
Bongers, S., & Mooij, J. M. (2018). From random differential equations to struc-
tural causal models: The stochastic case. arXiv preprint arXiv:1803.08784.
Borgatti, S. P. (2005). Centrality and network flow. Social networks, 27(1), 55–71.
Borsboom, D. (2017). A network theory of mental disorders. World Psychiatry,
16(1), 5–13.
Borsboom, D., & Cramer, A. O. (2013). Network analysis: an integrative ap-
proach to the structure of psychopathology. Annual Review of Clinical Psychol-
ogy, 9, 91–121.
Borsboom, D., Cramer, A. O., Schmittmann, V. D., Epskamp, S., & Waldorp, L. J.
(2011). The small world of psychopathology. PloS one, 6(11), e27407.
Borsboom, D., van der Sluis, S., Noordhof, A., Wichers, M., Geschwind, N.,
Aggen, S., . . . others (2012). What kind of causal modelling approach does
personality research need? European Journal of Personality, 26, 372–390.
Bos, F. M., Snippe, E., de Vos, S., Hartmann, J. A., Simons, C. J., van der Krieke,
L., . . . Wichers, M. (2017). Can we jump from cross-sectional to dynamic inter-
pretations of networks implications for the network perspective in psychiatry.
Psychotherapy and Psychosomatics, 86(3), 175–177.
215
REFERENCES
Boschloo, L., Schoevers, R. A., van Borkulo, C. D., Borsboom, D., & Oldehinkel,
A. J. (2016). The network structure of psychopathology in a community sample
of preadolescents. Journal of Abnormal Psychology, 125(4), 599–606.
Bramsen, R. H., Lasgaard, M., Koss, M. P., Shevlin, M., Elklit, A., & Banner,
J. (2013). Testing a multiple mediator model of the effect of childhood sexual
abuse on adolescent sexual victimization. American Journal of Orthopsychiatry,
83(1), 47–54.
Bringmann, L. F., Elmer, T., Epskamp, S., Krause, R. W., Schoch, D., Wichers, M.,
. . . Snippe, E. (2019). What do centrality measures measure in psychological
networks? The journal of abnormal psychology.
Bringmann, L. F., Lemmens, L., Huibers, M., Borsboom, D., & Tuerlinckx,
F. (2015). Revealing the dynamic network structure of the beck depression
inventory-ii. Psychological medicine, 45(4), 747–757.
Bringmann, L. F., Pe, M. L., Vissers, N., Ceulemans, E., Borsboom, D., Van-
paemel, W., . . . Kuppens, P. (2016). Assessing temporal emotion dynamics using
networks. Assessment, 23(4), 425–435.
Bringmann, L. F., Vissers, N., Wichers, M., Geschwind, N., Kuppens, P., Peeters,
F., . . . Tuerlinckx, F. (2013). A Network Approach to Psychopathology: New
Insights into Clinical Longitudinal Data. PloS One, 8(4), e60188. doi: 10.1371/
journal.pone.0060188
Browne, M. W., & Nesselroade, J. R. (2005). Representing psychological pro-
cesses with dynamic factor models: Some promising uses and extensions of
ARMA time series models. In A. Maydue-Olivares & J. J. McArdle (Eds.), Psy-
chometrics: A festschrift to Roderick P. McDonald (p. 415-452). Mahwah, NJ:
Lawrence Erlbaum Associates.
Bulteel, K., Tuerlinckx, F., Brose, A., & Ceulemans, E. (2016). Using raw var
regression coefficients to build networks can be misleading. Multivariate behav-
ioral research, 51(2-3), 330–344.
Carpenter, B., Gelman, A., Hoffman, M. D., Lee, D., Goodrich, B., Betancourt,
M., . . . Riddell, A. (2017). Stan: A probabilistic programming language. Journal
of Statistical Software, 76(1).
Cartwright, N. (1983). How the laws of physics lie. Oxford University Press.
Cartwright, N. (1999). Causal diversity and the markov condition. Synthese,
121(1-2), 3–27.
Cartwright, N. (2007). Hunting causes and using them: Approaches in philosophy
and economics. Cambridge University Press.
Chasalow, S. (2012). combinat: combinatorics utilities [Computer software
manual]. Retrieved from https://CRAN.R-project.org/package=combinat
(R package version 0.0-8)
216
REFERENCES
Chow, S.-M. (2019). Practical tools and guidelines for exploring and fitting lin-
ear and nonlinear dynamical systems models. Multivariate Behavioral Research,
1–29.
Chow, S.-M., Ferrer, E., & Hsieh, F. (2011). Statistical methods for modeling human
dynamics: An interdisciplinary dialogue. New York: Routledge.
Chow, S.-M., Ferrer, E., & Nesselroade, J. R. (2007). An unscented kalman filter
approach to the estimation of nonlinear dynamical systems models. Multivariate
Behavioral Research, 42(2), 283–321.
Chow, S.-M., Ou, L., Ciptadi, A., Prince, E. B., You, D., Hunter, M. D., . . .
Messinger, D. S. (2018). Representing sudden shifts in intensive dyadic in-
teraction data using differential equation models with regime switching. Psy-
chometrika, 83(2), 476–510.
Chow, S.-M., Ram, N., Boker, S., Fujita, F., Clore, G., & Nesselroade, J. (2005).
Capturing weekly fluctuation in emotion using a latent differential structural
approach. Emotion, 5(2), 208–225.
Christian, C., Perko, V., Vanzhula, I., Tregarthen, J., Forbush, K., & Levinson, C.
(2019). Eating disorder core symptoms and symptom pathways across develop-
mental stages: A network analysis. Journal of abnormal psychology.
Clark, D. M. (1986). A cognitive approach to panic. Behaviour Research and
Therapy, 24(4), 461–470.
Cole, D. A., & Maxwell, S. E. (2003). Testing mediational models with longitudi-
nal data: questions and tips in the use of structural equation modeling. Journal
of Abnormal Psychology, 112(4), 558–557.
Coleman, J. S. (1968). The mathematical study of change. Methodology in social
research, 428–478.
Collaboration, O. S., et al. (2015). Estimating the reproducibility of psychologi-
cal science. Science, 349(6251), aac4716.
Contreras, A., Nieto, I., Valiente, C., Espinosa, R., & Vazquez, C. (2019). The
study of psychopathology from the network analysis perspective: A systematic
review. Psychotherapy and Psychosomatics, 1–13.
Costantini, G., Epskamp, S., Borsboom, D., Perugini, M., Mõttus, R., Waldorp,
L. J., & Cramer, A. O. (2015). State of the aRt personality research: A tutorial
on network analysis of personality data in R. Journal of Research in Personality,
54, 13–29.
Cox, D. R., & Wermuth, N. (1996). Multivariate dependencies: Models, analysis
and interpretation. Chapman and Hall/CRC.
Cramer, A. O., van Borkulo, C. D., Giltay, E. J., van der Maas, H. L., Kendler,
K. S., Scheffer, M., & Borsboom, D. (2016). Major depression as a complex
dynamic system. PLoS One, 11(12), e0167490.
217
REFERENCES
Cramer, A. O., Waldorp, L. J., van der Maas, H. L. J., & Borsboom, D. (2010).
Comorbidity: A network perspective. Behavioral and Brain Sciences, 33(2-3),
137–150. doi: 10.1017/S0140525X09991567
Dablander, F., & Hinne, M. (2019). Node centrality measures are a poor substi-
tute for causal inference. Scientific reports, 9(1), 6846.
Dablander, F., Ryan, O., & Haslbeck, J. M. (2019, Jan). Choosing between AR(1)
and VAR(1) models in typical psychological applications. PsyArXiv. Retrieved from
psyarxiv.com/qgewy doi: 10.31234/osf.io/qgewy
Dahlhaus, R., & Eichler, M. (2003). Causality and graphical models in time
series analysis. Oxford Statistical Science Series, 115–137.
Dalege, J., Borsboom, D., van Harreveld, F., van den Berg, H., Conner, M., &
van der Maas, H. L. (2016). Toward a formalized account of attitudes: The
Causal Attitude Network (CAN) model. Psychological Review, 123(1), 2–22.
David, S. J., Marshall, A. J., Evanovich, E. K., & Mumma, G. H. (2018). Intraindi-
vidual dynamic network analysis–implications for clinical assessment. Journal
of psychopathology and behavioral assessment, 40(2), 235–248.
Dawid, A. P. (1979). Conditional independence in statistical theory. Journal of
the Royal Statistical Society. Series B (Methodological), 1–31.
218
REFERENCES
Deserno, M. K., Borsboom, D., Begeer, S., & Geurts, H. M. (2017). Multicausal
systems ask for multicausal approaches: A network perspective on subjective
well-being in individuals with autism spectrum disorder. Autism, 21(8), 960–
971.
de Wild-Hartmann, J. A., Wichers, M., van Bemmel, A. L., Derom, C., Thiery,
E., Jacobs, N., . . . Simons, C. J. (2013). Day-to-day associations between subjec-
tive sleep and affect in regard to future depressionin a female population-based
sample. The British Journal of Psychiatry, 202(6), 407–412.
Didelez, V. (2000). Graphical models for event history analysis based on local inde-
pendence. Logos Berlin.
Didelez, V. (2007). Graphical models for composable finite Markov processes.
Scandinavian Journal of Statistics, 34(1), 169–185.
Didelez, V. (2008). Graphical models for marked point processes based on
local independence. Journal of the Royal Statistical Society: Series B (Statistical
Methodology), 70(1), 245–264.
Didelez, V. (2019). Defining causal mediation with a longitudinal mediator and
a survival outcome. Lifetime data analysis, 25(4), 593–610.
Dormann, C., & Griffin, M. A. (2015). Optimal time lags in panel studies.
Psychological methods, 20(4), 489.
Driver, C. C., Oud, J. H., & Voelkle, M. C. (2017). Continuous time structural
equation modeling with R package ctsem. Journal of Statistical Software, 77(5),
1–35. doi: 10.18637/jss.v077.i05
Driver, C. C., & Voelkle, M. C. (2018). Understanding the time course of inter-
ventions with continuous time dynamic models. In K. L. Montfort, J. H. Oud,
& M. C. Voelkle (Eds.), Continuous time modeling in the behavioral and related
sciences (p. 179-203). New York: Springer.
Durbin, J., & Koopman, S. J. (2012). Time series analysis by state space methods.
Oxford university press.
Eichler, M., & Didelez, V. (2010). On Granger causality and the effect of inter-
ventions in time series. Lifetime Data Analysis, 16(1), 3–32.
Einicke, G. A. (2019). Smoothing, filtering and prediction: Estimating the past,
present and future. New York, NY: Prime Publishing.
219
REFERENCES
Epskamp, S., van Borkulo, C. D., van der Veen, D. C., Servaas, M. N., Isvoranu,
A.-M., Riese, H., & Cramer, A. O. (2018). Personalized network modeling in psy-
chopathology: The importance of contemporaneous and temporal connections.
Clinical Psychological Science, 6(3), 416–427.
Epskamp, S., Waldorp, L. J., Mõttus, R., & Borsboom, D. (2018). The Gaussian
graphical model in cross-sectional and time-series data. Multivariate Behavioral
Research, 53(4), 453–480.
Epstein, J. M. (2008). Why model? Journal of Artificial Societies and Social
Simulation, 11(4), 12.
220
REFERENCES
Fisher, A. J., Reeves, J. W., Lawyer, G., Medaglia, J. D., & Rubel, J. A. (2017). Ex-
ploring the Idiographic Dynamics of Mood and Anxiety via Network Analysis.
Journal of Abnormal Psychology, 126(8), 1044–1056. doi: 0.1037/abn0000311
Fisher, M. (2001). Modeling negative autoregression in continuous time.
(http://www.markfisher.net/ mefisher/papers/continuous ar.pdf)
Flake, J. K., & Fried, E. I. (2019, Jan). Measurement schmeasurement: Questionable
measurement practices and how to avoid them. PsyArXiv. Retrieved from psyarxiv
.com/hs7wm doi: 10.31234/osf.io/hs7wm
Fonseca-Pedrero, E., Ortuño, J., Debbané, M., Chan, R. C., Cicero, D., Zhang,
L. C., . . . Fried, E. I. (2018). The network structure of schizotypal personality
traits. Schizophrenia Bulletin, 44(2), S468–S479.
Forré, P., & Mooij, J. M. (2018). Constraint-based causal discovery for non-linear
structural causal models with cycles and latent confounders. arXiv preprint
arXiv:1807.03024.
Forré, P., & Mooij, J. M. (2019). Causal calculus in the presence of cycles, latent
confounders and selection bias. arXiv preprint arXiv:1901.00433.
Freedman, H. I. (1980). Deterministic mathematical models in population ecology
(Vol. 57). Marcel Dekker Incorporated.
Freedman, R. R., Ianni, P., Ettedgui, E., & Puthezhath, N. (1985). Ambulatory
monitoring of panic disorder. Archives of General Psychiatry, 42(3), 244–248.
Freeman, L. C. (1977). A set of measures of centrality based on betweenness.
Sociometry, 35–41.
Freeman, L. C. (1978). Centrality in social networks conceptual clarification.
Social networks, 1(3), 215–239.
Fried, E. I., Bockting, C., Arjadi, R., Borsboom, D., Amshoff, M., Cramer, A. O.,
. . . Stroebe, M. (2015). From loss to loneliness: The relationship between be-
reavement and depressive symptoms. Journal of Abnormal Psychology, 124(2),
256–265.
Fried, E. I., Boschloo, L., van Borkulo, C. D., Schoevers, R. A., Romeijn, J.-W.,
Wichers, M., . . . Borsboom, D. (2015). Commentary:“consistent superiority
of selective serotonin reuptake inhibitors over placebo in reducing depressed
mood in patients with major depression”. Frontiers in Psychiatry, 6, 117.
Fried, E. I., & Cramer, A. O. (2017). Moving forward: challenges and direc-
tions for psychopathological network theory and methodology. Perspectives on
Psychological Science, 12(6), 999–1020.
Fried, E. I., Epskamp, S., Nesse, R. M., Tuerlinckx, F., & Borsboom, D. (2016).
What are ’good’ depression symptoms? comparing the centrality of DSM and
non-DSM symptoms of depression in a network analysis. Journal of Affective
Disorders, 189, 314–320.
221
REFERENCES
Friedman, J., Hastie, T., & Tibshirani, R. (2001). The elements of statistical learn-
ing (Vol. 1) (No. 10). New York, NY: Springer Series in Statistics.
Friedman, J., Hastie, T., & Tibshirani, R. (2008). Sparse inverse covariance
estimation with the graphical lasso. Biostatistics, 9(3), 432–441.
Fukano, T., & Gunji, Y.-P. (2012). Mathematical models of panic disorder. Non-
linear Dynamics, Psychology, and Life Sciences, 16(4), 457–470.
Galles, D., & Pearl, J. (1995). Testing identifiability of causal effects. In Pro-
ceedings of the Eleventh Conference on Uncertainty in Artificial Intelligence (pp.
185–195).
Gardner, C., & Kleinman, A. (2019). Medicine and the mind — the consequences
of psychiatry’s identity crisis. New England Journal of Medicine, 381(18), 1697-
1699. Retrieved from https://doi.org/10.1056/NEJMp1910603 doi: 10.1056/
NEJMp1910603
Gault-Sherman, M. (2012). It’s a two-way street: The bidirectional relationship
between parenting and delinquency. Journal of Youth and Adolescence, 41, 121-
145.
Glauber, R. J. (1963). Time-dependent statistics of the Ising model. Journal of
Mathematical Physics, 4(2), 294–307.
Glymour, M. M. (2006). Using causal diagrams to understand common prob-
lems in social epidemiology. Methods in Social Epidemiology, 393–428.
Golino, H. F., & Epskamp, S. (2017). Exploratory graph analysis: A new ap-
proach for estimating the number of dimensions in psychological research. PloS
One, 12(6), e0174035.
Gollob, H. F., & Reichardt, C. S. (1987). Taking account of time lags in causal
models. Child Development, 58, 80–92.
Granger, C. W. (1969). Investigating causal relations by econometric models
and cross-spectral methods. Econometrica: Journal of the Econometric Society,
424–438.
Greenland, S., & Robins, J. M. (1986). Identifiability, exchangeability, and epi-
demiological confounding. International journal of epidemiology, 15(3), 413–419.
Grimm, V., & Railsback, S. F. (2005). Individual-based modeling and e ecology.
Princeton University press.
Groen, R. N., Snippe, E., Bringmann, L. F., Simons, C. J., Hartmann, J. A., Bos,
E. H., & Wichers, M. (2019). Capturing the risk of persisting depressive symp-
toms: A dynamic network investigation of patients’ daily symptom experiences.
Psychiatry Research, 271, 640–648. doi: 10.1016/j.psychres.2018.12.054
Grömping, U. (2006). Relative importance for linear regression in R: The pack-
age relaimpo. Journal of Statistical Software, 17(1), 1–27.
222
REFERENCES
223
REFERENCES
Hamerle, A., Nagl, W., & Singer, H. (1991). Problems with the estimation of
stochastic differential equations using structural equations models. Journal of
Mathematical Sociology, 16(3), 201–220.
Hamilton, J. D. (1989). A new approach to the economic analysis of nonstation-
ary time series and the business cycle. Econometrica: Journal of the Econometric
Society, 357–384.
Hamilton, J. D. (1994). Time series analysis (Vol. 2). Princeton, NJ: Princeton
University Press.
Haslbeck, J. M., & Fried, E. I. (2017). How predictable are symptoms in psy-
chopathological networks? A reanalysis of 18 published datasets. Psychological
Medicine, 47(16), 2767–2776.
Haslbeck, J. M., & Ryan, O. (2019, Sep). Recovering bistable systems from psy-
chological time series. PsyArXiv. Retrieved from psyarxiv.com/kcv3s doi:
10.31234/osf.io/kcv3s
Haslbeck, J. M., & Waldorp, L. J. (2018). How well do network models predict
observations? on the importance of predictability in network models. Behavior
Research Methods, 50(2), 853–861.
Hernan, M. A., & Robins, J. M. (2019). Causal inference. CRC Boca Raton, FL.
Higgs, P. W. (1964). Broken symmetries and the masses of gauge bosons. Phys-
ical Review Letters, 13(16), 508–509.
Hirsch, M. W., Smale, S., & Devaney, R. L. (2012). Differential equations, dynam-
ical systems, and an introduction to chaos. Academic press.
Hjelmeland, H., & Loa Knizek, B. (2018). The emperor’s new clothes? A critical
look at the interpersonal theory of suicide. Death Studies, 1–11. doi: 10.1080/
07481187.2018.1527796
224
REFERENCES
Ho, D. D., Neumann, A. U., Perelson, A. S., Chen, W., Leonard, J. M., &
Markowitz, M. (1995). Rapid turnover of plasma virions and cd4 lymphocytes
in hiv-1 infection. Nature, 373(6510), 123–126.
Hoge, E. A., Bui, E., Marques, L., Metcalf, C. A., Morris, L. K., Robinaugh, D. J.,
. . . Simon, N. M. (2013). Randomized controlled trial of mindfulness meditation
for generalized anxiety disorder: effects on anxiety and stress reactivity. The
Journal of clinical psychiatry, 74(8), 786–792.
Hoorelbeke, K., Marchetti, I., De Schryver, M., & Koster, E. H. (2016). The
interplay between cognitive risk and resilience factors in remitted depression: a
network analysis. Journal of Affective Disorders, 195, 96–104.
Horn, E. E., Strachan, E., & Turkheimer, E. (2015). Psychological distress and
recurrent herpetic disease: A dynamic study of lesion recurrence and viral shed-
ding episodes in adults. Multivariate behavioral research, 50(1), 134–135.
Hosenfeld, B., Bos, E. H., Wardenaar, K. J., Conradi, H. J., van der Maas, H. L.,
Visser, I., & de Jonge, P. (2015). Major depressive disorder as a nonlinear dy-
namic system: bimodality in the frequency distribution of depressive symptoms
over time. BMC Psychiatry, 15(222).
Houben, M., Van Den Noortgate, W., & Kuppens, P. (2015). The relation between
short-term emotion dynamics and psychological well-being: A meta-analysis.
Psychological Bulletin, 141(4), 901–930.
Ichii, K. (1991). Measuring mutual causation: Effects of suicide news on suicides
in Japan. Social Science Research, 20, 188-195.
Ioannidis, J. P. (2014). How to make more published research true. PLOS
Medicine, 11(10). doi: e1001747
Ising, E. (1925). Beitrag zur theorie des ferromagnetismus. Zeitschrift für Physik
A Hadrons and Nuclei, 31(1), 253–258.
Isvoranu, A.-M., van Borkulo, C. D., Boyette, L.-L., Wigman, J. T., Vinkers, C. H.,
Borsboom, D., & Investigators, G. (2016). A network approach to psychosis:
pathways between childhood trauma and psychotic symptoms. Schizophrenia
Bulletin, 43(1), 187–196.
Johnston, J., & DiNardo, J. (1972). Econometric methods. New York, 19(7), 22.
Jones, P. J. (2018). networktools: Tools for identifying important nodes in net-
works [Computer software manual]. Retrieved from https://CRAN.R-project
.org/package=networktools (R package version 1.2.0)
Jones, P. J., Mair, P., Riemann, B. C., Mugno, B. L., & McNally, R. J. (2018).
A network perspective on comorbid depression in adolescents with obsessive-
compulsive disorder. Journal of Anxiety Disorders, 53, 1–8.
225
REFERENCES
Kaiser, T., & Laireiter, A.-R. (2018). Daily dynamic assessment and modelling of
intersession processes in ambulatory psychotherapy: A proof of concept study.
Psychotherapy Research, 1–12.
Kalisch, M., & Bühlmann, P. (2007). Estimating high-dimensional directed
acyclic graphs with the PC-algorithm. Journal of Machine Learning Research,
8(Mar), 613–636.
Kalisch, M., Mächler, M., Colombo, D., Maathuis, M. H., & Bühlmann, P. (2012).
Causal inference using graphical models with the R package pcalg. Journal of
Statistical Software, 47(11), 1–26. Retrieved from http://www.jstatsoft.org/
v47/i11/
Kalisch, R., Cramer, A. O., Binder, H., Fritz, J., Leertouwer, I., Lunansky, G.,
. . . Van Harmelen, A.-L. (2019). Deconstructing and reconstructing resilience:
a dynamic network approach. Perspectives on Psychological Science, 14(5), 765–
777.
Kim, C.-J., & Nelson, C. R. (1999). State-space models with regime switching:
Classical and Gibbs-sampling approaches with applications. Cambridge, MA: The
MIT Press. doi: 10.2307/2669796
King, A. A., Nguyen, D., & Ionides, E. L. (2015). Statistical inference for
partially observed Markov processes via the R package pomp. arXiv preprint
arXiv:1509.00503.
Knefel, M., Tran, U. S., & Lueger-Schuster, B. (2016). The association of post-
traumatic stress disorder, complex posttraumatic stress disorder, and borderline
personality disorder from a network analytical perspective. Journal of Anxiety
Disorders, 43, 70–78.
Kossakowski, J., Groot, P., Haslbeck, J. M., Borsboom, D., & Wichers, M. (2017).
Data from ‘critical slowing down as a personalized early warning signal for de-
pression’. Journal of Open Psychology Data, 5(1).
Koval, P., Kuppens, P., Allen, N. B., & Sheeber, L. (2012). Getting stuck in de-
pression: The roles of rumination and emotional inertia. Cognition and Emotion,
26, 1412-1427.
Kraemer, N., Schaefer, J., & Boulesteix, A.-L. (2009). Regularized estimation
of large-scale gene regulatory networks using gaussian graphical models. BMC
Bioinformatics, 10(384).
226
REFERENCES
Kroeze, R., van der Veen, D. C., Servaas, M. N., Bastiaansen, J. A., Oude Voshaar,
R., Borsboom, D., & Riese, H. (2017). Personalized feedback on symptom
dynamics of psychopathology: A proof-of-principle study. Journal for Person-
Oriented Research, 3(1), 1–11.
Kuiper, R. M., & Ryan, O. (2018). Drawing conclusions from cross-lagged re-
lationships: Re-considering the role of the time-interval. Structural Equation
Modeling: A Multidisciplinary Journal, 25(5), 809–823.
Kulikov, G. Y., & Kulikova, M. V. (2013). Accurate numerical implementation of
the continuous-discrete extended Kalman filter. IEEE Transactions on Automatic
Control, 59(1), 273–279.
Kuppens, P., Allen, N. B., & Sheeber, L. B. (2010). Emotional inertia and psy-
chological maladjustment. Psychological Science, 21(7), 984–991.
Kuppens, P., Sheeber, L. B., Yap, M. B. H., Whittle, S., Simmons, J., & Allen,
N. B. (2012). Emotional inertia prospectively predicts the onset of depression
in adolescence. Emotion, 12, 283-289.
Lakatos, I. (1976). Falsification and the methodology of scientific research pro-
grammes. In S. Harding (Ed.), Can theories be refuted? (pp. 205–259). Dordrecht:
Springer.
227
REFERENCES
Lutz, W., Ehrlich, T., Rubel, J., Hallwachs, N., Röttger, M.-A., Jorasz, C., . . .
Tschitsaz-Stucki, A. (2013). The ups and downs of psychotherapy: Sudden
gains and sudden losses identified with session reports. Psychotherapy Research,
23(1), 14–24.
MacCallum, R. C., Wegener, D. T., Uchino, B. N., & Fabrigar, L. R. (1993). The
problem of equivalent models in applications of covariance structure analysis.
Psychological Bulletin, 114(1), 185–199.
Malinsky, D., & Danks, D. (2018). Causal discovery algorithms: A practical
guide. Philosophy Compass, 13(1), e12470.
Marks, R. J. I. (2012). Introduction to Shannon Sampling and Interpolation Theory.
Springer Science & Business Media.
Marsman, M., Borsboom, D., Kruis, J., Epskamp, S., van Bork, R., Waldorp, L., . . .
Maris, G. (2018). An introduction to network psychometrics: Relating ising net-
work models to item response theory models. Multivariate Behavioral Research,
53(1), 15–35.
Mauger, R., Tarbuck, E. J., & Lutgens, F. K. (1996). Earth: An introduction to
physical geology. Prentice-Hall.
Maxwell, S. E., & Cole, D. A. (2007). Bias in cross-sectional analyses of longitu-
dinal mediation. Psychological Methods, 12(1), 23–44.
McNally, R. J. (2016). Can network analysis transform psychopathology? Be-
haviour Research and Therapy, 86, 95–104.
McNally, R. J., Robinaugh, D. J., Wu, G. W., Wang, L., Deserno, M. K., & Bors-
boom, D. (2015). Mental disorders as causal systems: A network approach to
posttraumatic stress disorder. Clinical Psychological Science, 3(6), 836–849.
Meehl, P. E. (1978). Theoretical risks and tabular asterisks: Sir Karl, sir Ronald,
and the slow progress of soft psychology. Journal of Consulting and Clinical Psy-
chology, 46(4), 806–834.
Meehl, P. E. (1990). Appraising and amending theories: The strategy of
Lakatosian defense and two principles that warrant it. Psychological Inquiry,
1(2), 108–141.
Meier, B. P., & Robinson, M. D. (2004). Why the sunny side is up: Associations
between affect and vertical position. Psychological science, 15(4), 243–247.
Moberly, N. J., & Watkins, E. R. (2008). Ruminative self-focus and negative
affect: An experience sampling study. Journal of Abnormal Psychology, 117, 314-
323.
Molenaar, P. C. (2004). A manifesto on psychology as idiographic science:
Bringing the person back into scientific psychology, this time forever. Measure-
ment, 2(4), 201–218.
228
REFERENCES
Nguyen, J., & Frigg, R. (2017). Mathematics is not the only language in the book
of nature. Synthese, 1–22.
Opsahl, T., Agneessens, F., & Skvoretz, J. (2010). Node centrality in weighted
networks: Generalizing degree and shortest paths. Social networks, 32(3), 245–
251.
Oravecz, Z., & Tuerlinckx, F. (2011). The linear mixed model and the hierar-
chical ornstein–uhlenbeck model: Some equivalences and differences. British
Journal of Mathematical and Statistical Psychology, 64(1), 134–160.
Oravecz, Z., Tuerlinckx, F., & Vandekerckhove, J. (2009). A hierarchical
Ornstein-Uhlenbeck model for continuous repeated measurement data. Psy-
chometrika, 74, 395–418.
Oravecz, Z., Tuerlinckx, F., & Vandekerckhove, J. (2011). A hierarchical latent
stochastic difference equation model for affective dynamics. Psychological Meth-
ods, 16, 468–490.
229
REFERENCES
Oravecz, Z., Tuerlinckx, F., & Vandekerckhove, J. (2016). Bayesian data analysis
with the bivariate hierarchical ornstein-uhlenbeck process model. Multivariate
behavioral research, 51(1), 106–119.
Ou, L., Hunter, M. D., & Chow, S.-M. (2019). dynr: Dynamic modeling in R
[Computer software manual]. Retrieved from https://CRAN.R-project.org/
package=dynr (R package version 0.1.14-9)
Oud, J. H. (2007). Continuous time modeling of reciprocal relationships in the
cross-lagged panel design. In S. M. Boker & M. J. Wenger (Eds.), Data analytic
techniques for dynamic systems in the social and behavioral sciences (p. 87-129).
Mahwah, NJ: Lawrence Erlbaum Associates.
Oud, J. H., & Delsing, M. J. M. H. (2010). Continuous time modeling of panel
data by means of SEM. In K. van Montfort, J. H. Oud, & A. Satorra (Eds.), Lon-
gitudinal research with latent variables (pp. 201–244). New York, NY: Springer.
Oud, J. H., & Jansen, R. A. (2000). Continuous time state space modeling of
panel data by means of SEM. Psychometrika, 65(2), 199–215.
Oud, J. H., van Leeuwe, J., & Jansen, R. (1993). Kalman filtering in discrete and
continuous time based on longitudinal lisrel models. Advances in longitudinal
and multivariate analysis in the behavioral sciences, ITS, Nijmegen, Netherlands,
3–26.
Oud, J. H., Voelkle, M. C., & Driver, C. C. (2018). Sem based carma time series
modeling for arbitrary n. Multivariate behavioral research, 53(1), 36–56.
Papoulis, A., & Pillai, S. U. (2002). Probability, Random Variables, and Stochastic
Processes. Tata McGraw-Hill Education.
Pe, M. L., Kircanski, K., Thompson, R. J., Bringmann, L. F., Tuerlinckx, F.,
Mestdagh, M., . . . others (2015). Emotion-Network Density in Major De-
pressive Disorder. Clinical Psychological Science, 3(2), 292–300. doi: 10.1177/
2167702614540645
Pearl, J. (2009). Causality. Cambridge university press.
Pearl, J., & Verma, T. (1991). A theory of inferred causation. In J. Allen, R. Fikes,
& E. Sandewall (Eds.), Principles of knowledge representation and reasoning: Pro-
ceedings of the second international conference (pp. 441–452). San Mateo, CA.
Pero, F. (2015). Whither structuralism for scientific representation? (Unpublished
doctoral dissertation). Univeristy of Florence.
Rabiner, L. R. (1989). A tutorial on hidden markov models and selected appli-
cations in speech recognition. Proceedings of the IEEE, 77(2), 257–286.
Raykov, T., & Marcoulides, G. A. (2001). Can there be infinitely many models
equivalent to a given covariance structure model? Structural Equation Modeling,
8(1), 142–149.
230
REFERENCES
Richards, A., French, C. C., Johnson, W., Naparstek, J., & Williams, J. (1992). Ef-
fects of mood manipulation and anxiety on performance of an emotional stroop
task. British Journal of Psychology, 83(4), 479–491.
Richards, A., & Whittaker, T. M. (1990). Effects of anxiety and mood manipu-
lation in autobiographical memory. British Journal of Clinical Psychology, 29(2),
145–153.
Richardson, T. S., & Robins, J. M. (2013). Single world intervention graphs
(swigs): A unification of the counterfactual and graphical approaches to causal-
ity. Center for the Statistics and the Social Sciences, University of Washington Series.
Working Paper, 128(30), 2013.
Roberson-Nay, R., Gorlin, E. I., Beadel, J. R., Cash, T., Vrana, S., & Teachman,
B. A. (2017). Temporal stability of multiple response systems to 7.5% carbon
dioxide challenge. Biological psychology, 124, 111–118.
Robinaugh, D. J., Haslbeck, J. M., Waldorp, L. J., Kossakowski, J. J., Fried, E. I.,
Millner, A. J., . . . Borsboom, D. (2019). Advancing the network theory of men-
tal disorders:a computational model of panic disorder. PsyArXiv. Retrieved from
psyarxiv.com/km37w doi: 10.31234/osf.io/km37w
Robinaugh, D. J., Hoekstra, R. H., Toner, E. R., & Borsboom, D. (2019). The
network approach to psychopathology: a review of the literature 2008–2018
and an agenda for future research. Psychological Medicine, 1–14. doi: 10.1017/
S0033291719003404
Robinaugh, D. J., LeBlanc, N. J., Vuletich, H. A., & McNally, R. J. (2014).
Network analysis of persistent complex bereavement disorder in conjugally be-
reaved adults. Journal of abnormal psychology, 123(3), 510–522.
Robinaugh, D. J., Millner, A. J., & McNally, R. J. (2016). Identifying highly influ-
ential nodes in the complicated grief network. Journal of Abnormal Psychology,
125(6), 747–757.
Robins, J. M. (1999). Association, causation, and marginal structural models.
Synthese, 121(1-2), 151–179.
231
REFERENCES
Ryan, O., Bringmann, L. F., & Schuurman, N. (2019, Oct). The challenge of
generating causal hypotheses using network models. PsyArXiv. Retrieved from
psyarxiv.com/ryg69 doi: 10.31234/osf.io/ryg69
Ryan, O., & Hamaker, E. L. (2019). Time to intervene: A continuous-time approach
to network analysis and centrality. (Manuscript in Preparation)
Ryan, O., Kuiper, R. M., & Hamaker, E. L. (2018). A continuous time approach to
intensive longitudinal data: What, why and how? In K. L. Montfort, J. H. Oud,
& M. C. Voelkle (Eds.), Continuous time modeling in the behavioral and related
sciences (pp. 29–57). New York: Springer.
Scheffer, M., Bascompte, J., Brock, W. A., Brovkin, V., Carpenter, S. R., Dakos,
V., . . . Sugihara, G. (2009). Early-warning signals for critical transitions. Nature,
461(7260), 53.
Scheffer, M., Bolhuis, J. E., Borsboom, D., Buchman, T. G., Gijzel, S. M., Goulson,
D., . . . others (2018). Quantifying resilience of humans and other animals.
Proceedings of the National Academy of Sciences, 115(47), 11883–11890.
Schmittmann, V. D., Cramer, A. O., Waldorp, L. J., Epskamp, S., Kievit, R. A., &
Borsboom, D. (2013). Deconstructing the construct: A network perspective on
psychological phenomena. New ideas in psychology, 31(1), 43–53.
Schuurman, N. K., Ferrer, E., de Boer-Sonnenschein, M., & Hamaker, E. L.
(2016). How to compare cross-lagged associations in a multilevel autoregres-
sive model. Psychological Methods, 21(2), 206-221.
Schwarz, G., et al. (1978). Estimating the dimension of a model. The Annals of
Statistics, 6(2), 461–464.
Schweder, T. (1970). Composable Markov processes. Journal of Applied Proba-
bility, 7(2), 400–410.
232
REFERENCES
233
REFERENCES
Spirtes, P., & Zhang, K. (2016). Causal discovery and inference: concepts and
recent methodological advances. Applied Informatics, 3.
Spitzer, R. L., Kroenke, K., & Williams, J. B. (1980). Diagnostic and statistical
manual of mental disorders. Washington, DC: American Psychiatric Association.
Steele, J. S., & Ferrer, E. (2011). Latent differential equation modeling of
self-regulatory and coregulatory affective processes. Multivariate Behavioral Re-
search, 46(6), 956–984.
Steele, R. J., & Raftery, A. E. (2010). Performance of bayesian model selection
criteria for gaussian mixture models. Frontiers of Statistical Decision Making and
Bayesian Analysis, 2, 113–130.
Stiles, W. B., Leach, C., Barkham, M., Lucock, M., Iveson, S., Shapiro, D. A.,
. . . Hardy, G. E. (2003). Early sudden gains in psychotherapy under routine
clinic conditions: Practice-based evidence. Journal of Consulting and Clinical
Psychology, 71(1), 14–21.
Strauss, A., & Corbin, J. (1994). Grounded theory methodology. Handbook of
qualitative research, 17, 273–85.
Strogatz, S. H. (2015). Nonlinear dynamics and chaos: with applications to physics,
biology, chemistry, and engineering. Colorado, USA: Westview press.
Stutz, C., & Williams, B. (1999). Obituary: Ernst Ising. Physics Today, 52, 106–
108.
Suárez, M., & Pero, F. (2019). The representational semantic conception. Philos-
ophy of Science, 86(2), 344–365.
Suls, J., Green, P., & Hillis, S. (1998). Emotional reactivity to everyday prob-
lems, affective inertia, and neuroticism. Personality and Social Psychology Bul-
letin, 24(2), 127–136.
Suppes, P. (1962). Models of data. In E. Nagel, P. Suppes, & A. Tarski (Eds.),
Logic, methodology and the philosophy of science: Proceedings of the 1960 interna-
tional congress. CA, Stanford University Press.
Swoyer, C. (1991). Structural representation and surrogative reasoning. Syn-
these, 87(3), 449–508.
Szollosi, A., & Donkin, C. (2019). Neglected sources of flexibility in psycholog-
ical theories: From replicability to good explanations. Computational Brain &
Behavior, 1–3.
Tomarken, A. J., & Waller, N. G. (2003). Potential problems with “well fitting”
models. Journal of Abnormal Psychology, 112(4), 578–598.
Tong, H., & Lim, K. (1980). Threshold autoregression, limit cycles and cyclical
data. Journal of the Royal Statistical Society: Series B (Methodological), 42(3), 245–
268.
234
REFERENCES
Van Borkulo, C. D., Borsboom, D., Epskamp, S., Blanken, T. F., Boschloo, L.,
Schoevers, R. A., & Waldorp, L. J. (2014). A new method for constructing net-
works from binary data. Scientific Reports, 4, 5918.
van de Leemput, I. A., Wichers, M., Cramer, A. O., Borsboom, D., Tuerlinckx,
F., Kuppens, P., . . . others (2014). Critical slowing down as early warning for
the onset and termination of depression. Proceedings of the National Academy of
Sciences, 111(1), 87–92.
Van Der Maas, H. L., Dolan, C. V., Grasman, R. P., Wicherts, J. M., Huizenga,
H. M., & Raijmakers, M. E. (2006). A dynamical model of general intelligence:
the positive manifold of intelligence by mutualism. Psychological Review, 113(4),
842–861.
235
REFERENCES
236
REFERENCES
237
Nederlandse Samenvatting
239
Nederlandse Samenvatting
betrekking tot conclusies over causale relaties. Hiernaast worden nieuwe cen-
trality statistieken specifiek voor CT netwerken ontwikkeld. Geı̈nspireerd door
de interventionele causale inferentie literatuur, stellen deze statistieken onder-
zoekers in staat om variabelen voor verschillende interventies (acuut of continu)
optimaal te identificeren.
De focus van hoofdstuk 5 is bistabiele systemen, een specifiek type dynamisch
systeem dat veel aandacht heeft gekregen in de psychologie literatuur. De ca-
paciteit van verschillende gangbare statistische modellen om de eigenschappen
van zulke systemen te achterhalen wordt onderzocht, waarbij wordt uitgegaan
van twee scenario’s: ideale data en meer realistische (met langere tijdsintervallen
tussen metingen) data. Met de ideale data wordt aangetoond dat sommige statis-
tische modellen gebruikt kunnen worden om de globale dynamiek (de meer sta-
biele relaties tussen variabelen) te achterhalen maar dat het moeilijk is om dit
correct te doen voor de microdynamiek (moment-tot-moment relaties). Voor de
realistische data geldt dat de globale dynamiek nog steeds gevonden kan worden
maar dat dit helemaal niet mogelijk is voor microdynamiek. Deze resultaten be-
nadrukken a) hoe moeilijk het is om inferenties maken op basis van statistische
modellen zonder een sterke theorie, en b) de fundamentele rol van de frequentie
van de dataverzameling bij het statistisch modelleren van tijdreeksen.
In hoofdstuk 6 wordt betoogd dat formele theorieën van kritiek belang zijn
om onderzoek te doen naar psychiatrische stoornissen. Formele theorieën –
geoperationaliseerd als een set differentiaalvergelijkingen – zijn gebruikelijk in
disciplines die dynamische systeemtheorie toepassen, maar ontbreken bijna al-
tijd in de klinische psychologie. Eerst wordt een kort overzicht gegeven van de
wetenschapsfilosofische literatuur om het belang van formele theorieën te be-
nadrukken. Daarna worden drie manieren onderzocht om formele theorieën
te construeren op basis van statische modellen: a) het gebruik van statistische
modellen als formele theorieën, b) het gebruik van statistische modellen om
formele theorieën van af te leiden, en c) het gebruik van statistische modellen
om bestaande formele theorieën te verbeteren. De derde benadering blijkt het
meeste veelbelovende te zijn maar staat ook het verst af van de huidige praktijk.
Tenslotte wordt een kader voorgesteld dat beschrijft hoe empirisch onderzoek het
beste kan worden ingezet om formele theorieën over psychiatrische stoornissen
te genereren, testen, en verbeteren.
Dit proefschrift eindigt met een duidelijke boodschap voor de klinische psy-
chologie, de psychiatrie en de methodologen die binnen deze disciplines werken:
als we hopen een beter begrip te krijgen over de complexe dynamische processen
die ten grondslag liggen aan psychopathologie, moeten we de huidige onder-
zoekspraktijken met betrekking tot de ontwikkeling van formele theorie radicaal
heroriënteren.
240
About the Author
Oisı́n Ryan was born on November 23rd 1991 in Kilkenny, Ireland. In 2013
he obtained his BSc. in Psychology from the University of Limerick with first-
class honors. During his bachelor program he worked as a research assistant for
Dr. Timothy D. Ritchie and visited Utrecht University in the Netherlands for
six months as part of an Erasmus program. In September 2013 he returned to
Utrecht University and enrolled in the research masters Methodology and Statis-
tics for the Behavioral, Biomedical and Social Sciences, graduating cum laude in
2015.
Awarded a talent grant from the Netherlands Organization for Scientific Re-
search (NWO), he began his PhD project in September 2015 under the supervi-
sion of Prof. dr. Ellen Hamaker at the Department of Methodology and Statistics
in Utrecht. He has given a variety of invited presentations and workshops to
different groups, and has presented his research at several international confer-
ences, including the International Meeting of the Psychometrics Society (IMPS),
the Association for Psychological Science (APS) annual convention, and the Con-
ference on Complex Systems (CCS). In 2019 he spent one month as a visiting
scholar with Dr. Donald J. Robinaugh at the Department of Psychiatry, Mas-
sachusetts General Hospital, Harvard Medical School, Boston.
As of January 2020, Oisı́n holds a post-doctoral position at Utrecht University,
allowing him to continue his research on dynamical systems modeling in clinical
psychology and psychiatry.
241
Publications & Working Papers
Ryan, O., Bringmann, L. F., & Schuurman, N. K. (under review). The chal-
lenge of generating causal hypotheses using network models. Pre-print DOI:
10.31234/osf.io/ryg69
Haslbeck, J. M. B.*, Ryan, O.*, Robinaugh, D.*, Waldorp, L. J., & Borsboom,
D. (under review). Modeling psychopathology: From data models to formal
theories. Pre-print DOI: 10.31234/osf.io/jgm7f
Haslbeck, J. M. B.* & Ryan, O.* (under review). Recovering within-person dy-
namics from psychological time series. Pre-print DOI: 10.31234/osf.io/dymhw
Dablander, F.*, Ryan, O.*, & Haslbeck, J. M. B.* (under review). Choosing be-
tween AR(1) and VAR(1) models in typical psychological applications. Pre-print
DOI: 10.31234/osf.io/qgewy
Haslbeck, J. M. B.*, Ryan, O.*, & Dablander, F.* (under review). The sum of
all fears: Comparing networks based on symptom sum-scores. Pre-print DOI:
10.31234/osf.io/3nxu9
Robinaugh, D. J., Haslbeck, J. M. B., Ryan, O., Fried, E. I., & Waldorp, L. J. (under
review) Invisible hands and fine calipers: A call to use formal theory as a toolkit
for theory construction. Perspectives on Psychological Science. Pre-print available
from https://psyarxiv.com/ugz7y.
Groen, R. N., Ryan, O., Wigman, J. T. W., Riese, H., Penninx, B. W. J. H., Giltay,
E. J., Wichers, M., & Hartman, C. A. (in press) Comorbidity between depression
and anxiety: Assessing the role of bridge mental states in dynamic psychological
networks. BMC Medicine.
Bastiaansen, J. A., Kunkels, Y. K., Blaauw, F., Boker, S. M., Ceulemans, E.,
Chen, M., Chow, S. M, de Jonge, P., Emerencia, A. C., Epskamp, S., Fisher, A.
J., Hamaker, E.L., Kuppens, P., Lutz, W., Meyer, M. J., Moulder, R., Oravecz, Z.,
Riese, H., Rubel, J., Ryan, O., Servaas, M. N., Sjobeck, G., Snipper, E., Trull,
243
Publications & Working Papers
T. J., Tschacher, W., van der Veen, D. C., Wichers, M., Wood, P. K., Woods,
W. C., Wright, A. G. C., Albers, C. J. & Bringmann, L. F. (2020). Time to get
personal? The impact of researchers’ choices on the selection of treatment targets
using the experience sampling methodology. Journal of Psychosomatic Research.
https://doi.org/10.1016/j.jpsychores.2020.110211
van de Schoot, R., Winter, S. D., Ryan, O., Zondervan-Zwijnenburg, M., &
Depaoli, S. (2017). A systematic review of Bayesian articles in psychology: The
last 25 years. Psychological Methods, 22(2), 217–239.
Hepper, E. G., Wildschut, T., Sedikides, C., Ritchie, T. D., Yung, Y. F., Hansen, N.,
. . . Ryan, O., & Stephan, E., & Vingerhoets, A . J. J. (2014). Pancultural nostalgia:
prototypical conceptions across cultures. Emotion, 14(4), 733–747.
244
Acknowledgements
Ellen. Since most people only ever do one PhD, I guess everyone feels that their
experience was unique, whether it was or not. But I’m pretty sure mine was, and
a large part of that was down to you. Thank you for sharing your knowledge, for
your support, your openness, your honesty, your humor and your endless supply
of (in)appropriate idioms. Thanks for giving me plenty of rope, and for stepping
in every now and then to ensure I didn’t hang myself with it. I’m not sure how
this all would have turned out if I didn’t have you to act as my mentor, confidant,
“work mom”, and critic. I’m glad we don’t have to deal with that counterfactual.
Jonas. If I were to make a highlight reel of the last few years of my PhD, you
would feature in a striking amount of it. In Utrecht and Amsterdam, Boston,
Greece and Bordeaux; In our apartments, offices and Airbnbs, spas, bars, and too
many restaurants. Our dinners, debates, discussions and arguments all helped
keep me motivated, focused, and having fun. I’m proud of the work we did to-
gether, and I feel tremendously lucky to have you as a friend. My hope is to enjoy
many more years of friendship and collaboration: May we never tire of ranting at
each other, never agree on how to use semicolons, and never stop surreptitiously
editing each other’s punctuation.
Noémi and Laura, if I were to extend the uncomfortable “work family”
metaphor, you two would be my work big-sisters. Laura, thanks for your will-
ingness to discuss any topic at any time, and do it all with good humor. Noémi,
thanks for your willingness to argue with me about any topic at any time. I
learned a lot from both your empathy and your general intolerance of nonsense.
Don, the month I spent working with you and Jonas in Boston was probably
the best experience I’ve had in academia. Thanks for putting up with us demand-
ing your time, shouting so loudly that your boss installed a white noise machine
outside your office, and for the marathon skype sessions that have followed every
few weeks since we left the US. They have been a true inspiration.
Fabian, I’m very happy that we ended up working with each other so much
in the last couple of years. I hope to write many more lengthy, informative, and
ultimately unnecessary appendices with you. I consider myself a fan of your
work, and I’m delighted that I can count you as a good friend too.
Thanks to everyone at the M&S department, especially Fayette, Thomas, Kees,
Jolien, Anne, Erik-Jan and Ayoub, for all the lunches, coffees and walks around
the uithof. All of these were essential distractions and stress-release valves which
kept me sane in the office. Thanks especially to Erik-Jan and Ayoub in the last
couple of years for sharing the office with me, feeding my caffeine addiction,
dealing with my grumpy behaviour, and engaging me in general tomfoolery when
it was needed most. Thanks to Rebecca for always being willing to help with the
245
Acknowledgements
most obscure mathematical questions that came to mind, and thanks to Gerko
for all the random pieces of advice over the years. Thanks to Flip and Chantal
for the cookies and impromptu Dutch lessons, and to Kevin for putting up with
my constant requests for office equipment. Thanks to Irene Klugkist, who has
somehow conspired to play an important role in many big moments in my life
in recent years: You gave me my first job in the Netherlands, helping me be able
to move here in the first place, you introduced me to Roline for the first time
(although it was not the smoothest first meeting), and now you’ve been a member
of my reading committee. I’m grateful for all of your help.
Thanks to everyone outside of work for distracting me, giving me many great
memories and experiences, and making me feel at home in the past years. Thanks
to Thomas, Millitza and Fayette for all the trips, the marathon board game ses-
sions, and the willingness to engage in all sorts of random children’s pastimes.
Thanks to Joris for all the coffees, visits to the uithof sheep, and your boundless
enthusiasm for new adventures. Thanks to EJ and Lara, Marloes and Vincent, for
all the dinners, games, bottles of wine, late nights and shared recipes. Especially
thanks to all the Kamphuises for making me feel part of the family.
Thanks to all my friends in Ireland, especially Ali, Eoin, Declan, Luke, the
Steves, Kate, Niamh, Cillian, Úna, Zoe and Sally. Thanks for all the meet-ups,
whatsapp calls, facebook messages and occasional reminders of the most em-
barrassing moments of my past. Thanks in particular to Luke for designing the
cover of this book, and Ali for your never-ending enthusiasm for planning shared
holidays and visits to London or Utrecht.
Thanks to all my family for putting up with my too-infrequent visits home
and for ferrying me all over the place when I do arrive back. Thanks to my par-
ents for all the support in moving and setting up a new life in the Netherlands;
Colin, Pia and the kids for having myself and Roline stay for as long as we want
every Christmas; thanks to Niamh for “checking in to see if I’m still alive” on a
frequent basis, and thanks to Paudie for always reminding me how similar we
are, and giving me the peace of mind that everything is taken care of back home.
Finally, thanks to Roline. I have threatened many ways of messing with you
in my acknowledgements, all of which you have successfully vetoed: This para-
graph is longer than one word, it won’t feature a proposal of any kind, and there’s
no need to search for cryptic messages spelled out across the first letter of each
preceding line (turns out that’s much more difficult to pull off than you’d think).
You were there for everything the last few years: The joy, the despair, frustrations
and elations, the bouts of self-doubt, stress, mania and lethargy. You’re in the
top two people who suffered most in my bad moments during this PhD, and I’m
genuinely not sure if I’m number one on that list or you. Through all that you
built us an amazing life together in Utrecht. I promise not to do another PhD, I
promise not to use the doctor title during petty disagreements, and I promise not
to argue with you about what constitutes a “real job” anymore. I couldn’t have
done this without you. You’re my rock, my best friend, and the love of my life.
Thanks.
246