ROBINS-I Detailed Guidance
ROBINS-I Detailed Guidance
detailed guidance
Edited by Jonathan AC Sterne, Julian PT Higgins, Roy G Elbers and Barney C Reeves
on behalf of the development group for ROBINS-I
Updated 20 October 2016
To cite the ROBINS-I tool: Sterne JAC, Hernán MA, Reeves BC, Savović J, Berkman ND, Viswanathan M, Henry
D, Altman DG, Ansari MT, Boutron I, Carpenter JR, Chan AW, Churchill R, Deeks JJ, Hróbjartsson A, Kirkham J,
Jüni P, Loke YK, Pigott TD, Ramsay CR, Regidor D, Rothstein HR, Sandhu L, Santaguida PL, Schünemann HJ, Shea
B, Shrier I, Tugwell P, Turner L, Valentine JC, Waddington H, Waters E, Wells GA, Whiting PF, Higgins JPT.
ROBINS-I: a tool for assessing risk of bias in non-randomized studies of interventions. BMJ 2016; 355; i4919.
To cite this document: Sterne JAC, Higgins JPT, Elbers RG, Reeves BC and the development group for ROBINS-
I. Risk Of Bias In Non-randomized Studies of Interventions (ROBINS-I): detailed guidance, updated 12 October
2016. Available from http://www.riskofbias.info [accessed {date}]
Contents
1 Contributors .......................................................................................................................................................... 2
2 Background .............................................................................................................................................................3
2.1 Context of the tool .................................................................................................................................. 3
2.2 Assessing risk of bias in relation to a target trial................................................................................. 3
2.3 Domains of bias ...................................................................................................................................... 4
2.4 Study designs .......................................................................................................................................... 8
2.5 Risk of bias assessments should relate to a specified intervention effect ......................................... 8
2.6 Structure of this document .................................................................................................................... 8
3 Guidance for using the tool: general considerations......................................................................................... 9
3.1 At protocol stage ..................................................................................................................................... 9
3.2 Preliminary considerations for each study ......................................................................................... 11
3.3 Signalling questions ............................................................................................................................. 16
3.4 Domain-level judgements about risk of bias ..................................................................................... 16
3.5 Reaching an overall judgement about risk of bias ............................................................................. 17
3.6 Assessing risk of bias for multiple outcomes in a review ................................................................. 18
4 Guidance for using the tool: detailed guidance for each bias domain .......................................................... 20
4.1 Detailed guidance: Bias due to confounding ..................................................................................... 20
4.2 Detailed guidance: Bias in selection of participants into the study ................................................ 28
4.3 Detailed guidance: Bias in classification of interventions ................................................................ 32
4.4 Detailed guidance: Bias due to deviations from intended interventions........................................ 34
4.5 Detailed guidance: Bias due to missing data ..................................................................................... 43
4.6 Detailed guidance: Bias in measurement of outcomes .....................................................................46
4.7 Detailed guidance: Bias in selection of the reported result .............................................................49
5 References ............................................................................................................................................................. 53
1
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
1 Contributors
(Listed alphabetically within category)
Core group: Julian Higgins, Barney Reeves, Jelena Savović, Jonathan Sterne, Lucy Turner.
Additional core research staff: Roy Elbers, Alexandra McAleenan, Matthew Page.
Bias due to confounding: Nancy Berkman, Miguel Hernán, Pasqualina Santaguida, Jelena Savović, Beverley Shea,
Jonathan Sterne, Meera Viswanathan.
Bias in selection of participants into the study: Nancy Berkman, Miguel Hernán, Pasqualina Santaguida, Jelena
Savović, Beverley Shea, Jonathan Sterne, Meera Viswanathan.
Bias due to departures from intended interventions: David Henry, Julian Higgins, Peter Jüni, Lakhbir Sandhu,
Pasqualina Santaguida, Jonathan Sterne, Peter Tugwell.
Bias due to missing data: James Carpenter, Julian Higgins, Terri Piggott, Hannah Rothstein, Ian Shrier, George
Wells.
Bias in measurement of outcomes or interventions: Isabelle Boutron, Asbjørn Hróbjartsson, David Moher, Lucy
Turner.
Bias in selection of the reported result: Doug Altman, Mohammed Ansari, Barney Reeves, An-Wen Chan, Jamie
Kirkham, Jeffrey Valentine.
Cognitive testing leads: Nancy Berkman, Meera Viswanathan.
Piloting and cognitive testing participants: Katherine Chaplin, Hannah Christensen, Maryam Darvishian, Anat
Fisher, Laura Gartshore, Sharea Ijaz, J Christiaan Keurentjes, José López-López, Natasha Martin, Ana Marušić,
Anette Minarzyk, Barbara Mintzes, Maria Pufulete, Stefan Sauerland, Jelena Savović, Nandi Seigfried, Jos Verbeek,
Marie Wetwood, Penny Whiting.
Other contributors: Belinda Burford, Rachel Churchill, Jon Deeks, Toby Lasserson, Yoon Loke, Craig Ramsay,
Deborah Regidor, Jan Vandenbroucke, Penny Whiting.
2
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
2 Background
The goal of a systematic review of the effects of an intervention is to determine its causal effects on one or more
outcomes. When the included studies are randomized trials, causality can be inferred if the trials are
methodologically sound, because successful randomization of a sufficiently large number of individuals should
result in intervention and comparator groups that have similar distributions of both observed and unobserved
prognostic factors. However, evidence from randomized trials may not be sufficient to answer questions of
interest to patients and health care providers, and so systematic review authors may wish to include non-
randomized studies of the effects of interventions (NRSIs) in their reviews.
Our ROBINS-I tool (“Risk Of Bias In Non-randomized Studies - of Interventions”) is concerned with evaluating
the risk of bias (RoB) in the results of NRSIs that compare the health effects of two or more interventions. The
types of NRSIs that can be evaluated using this tool are quantitative studies estimating the effectiveness (harm or
benefit) of an intervention, which did not use randomization to allocate units (individuals or clusters of
individuals) to comparison groups. This includes studies where allocation occurs during the course of usual
treatment decisions or peoples’ choices: such studies are often called “observational”. There are many types of
such NRSIs, including cohort studies, case-control studies, controlled before-and-after studies, interrupted-time-
series studies and controlled trials in which intervention groups are allocated using a method that falls short of
full randomization (sometimes called “quasi-randomized” studies). This document provides guidance for using
the ROBINS-I tool specifically for studies with a cohort-type of design, in which individuals who have received
(or are receiving) different interventions are followed up over time.
The ROBINS-I tool is based on the Cochrane RoB tool for randomized trials, which was launched in 2008 and
modified in 2011 (Higgins et al, 2011). As in the tool for randomized trials, risk of bias is assessed within specified
bias domains, and review authors are asked to document the information on which judgements are based.
ROBINS-I also builds on related tools such as the QUADAS 2 tool for assessment of diagnostic accuracy studies
(Whiting et al, 2011) by providing signalling questions whose answers flag the potential for bias and should help
review authors reach risk of bias judgements. Therefore, the ROBINS-I tool provides a systematic way to organize
and present the available evidence relating to risk of bias in NRSI.
3
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
patients, health care professionals and outcome assessors to intervention received throughout follow up;
ascertained outcomes in all randomized participants; and reported intervention effects for all measured outcomes.
Defined in this way, bias is distinct from issues of generalizability (applicability or transportability) to types
of individual who were not included in the study. For example, restricting the study sample to individuals free of
comorbidities may limit the utility of its findings because they cannot be generalized to clinical practice, where
comorbidities are common.
Evaluations of risk of bias in the results of NRSIs are therefore facilitated by considering each NRSI as an attempt
to emulate (mimic) a hypothetical trial. This is the hypothetical pragmatic randomized trial that compares the
health effects of the same interventions, conducted on the same participant group and without features putting
it at risk of bias (Hernán 2011; Institute of Medicine 2012). We refer to such a hypothetical randomized trial
as a “target” randomized trial (see section 3.1.1 for more details). Importantly, a target randomized trial need
not be feasible or ethical.
ROBINS-I requires that review authors explicitly identify the interventions that would be compared in the target
trial that the NRSI is trying to emulate. Often the description of these interventions will require subject-matter
knowledge, because information provided by the investigators of the observational study is insufficient to define
the target trial. For example, authors may refer to “use of therapy [A],” which does not directly correspond to the
intervention “initiation of therapy [A]” that would be tested in an intention-to-treat analysis of the target trial.
Meaningful assessment of risk of bias is problematic in the absence of well-defined interventions. For example, it
would be harder to assess confounding for the effect of obesity on mortality than for the effect of a particular
weight loss intervention (e.g., caloric restriction) in obese people on mortality.
To keep the analogy with the target trial, this document uses the term “intervention” groups to refer to
“treatment” or “exposure” groups in observational studies even though in such studies no actual intervention
was implemented by the investigators.
Variation in terminology between contributors and between research areas proved a challenge to development of
ROBINS-I and to writing guidance. The same terms are sometimes used to refer to different types of bias, and
different types of bias are often described by a host of different terms. Table 1 explains the terms that we have
chosen to describe each bias domain, and related terms that are sometimes used. The term selection bias is a
particular source of confusion. It is often used as a synonym for confounding (including in the current Cochrane
tool for assessing RoB in randomized trials), which occurs when one or more prognostic factors also predict
whether an individual receives one or the other intervention of interest. We restrict our use of the term selection
bias to refer to a separate type of bias that occurs when some eligible participants, or the initial follow up time of
some participants, or some outcome events, are excluded in a way that leads to the association between
intervention and outcome differing from the association that would have been observed in complete follow up of
the target trial. We discourage the use of the term selection bias to refer to confounding, although we have done
this in the past, for example in the context of the RoB tool for randomized trials. Work is in progress to resolve
this difference in terminology between the ROBINS-I tool and the current Cochrane tool for assessing RoB in
randomized trials.
By contrast with randomized trials, in NRSIs the characteristics of study participants will typically differ between
intervention groups. The assessment of the risk of bias arising from uncontrolled confounding is therefore a major
component of the ROBINS-I assessment. Confounding of intervention effects occurs when one or more
prognostic factors (factors that predict the outcome of interest) also predict whether an individual receives one
or the other intervention of interest. As an example, consider a cohort study of HIV-infected patients that
4
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
compares the risk of death from initiation of antiretroviral therapy A versus antiretroviral therapy B. If
confounding is successfully controlled, the effect estimates from this observational study will be identical, except
for sampling variation, to those from a trial that randomly assigns individuals in the same study population to
either intervention A or B. However, failure to control for key confounders may violate the expectation of
comparability between those receiving therapies A and B, and thus result in bias. A detailed discussion of
assessment of confounding appears in section 4.1
Selection bias may arise when the analysis does not include all of the participants, or all of their follow-up after
initiation of intervention, that would have been included in the target randomized trial. The ROBINS-I tool
addresses two types of selection bias: (1) bias that arises when either all of the follow-up or a period of follow-up
following initiation of intervention is missing for some individuals (for example, bias due to the inclusion of
prevalent users rather than new users of an intervention), and (2) bias that arises when later follow-up is missing
for individuals who were initially included and followed (for example, bias due to differential loss to follow-up
that is affected by prognostic factors).We consider the first type of selection bias under “Bias in selection of
participants into the study” (section 4.2), and aspects relating to loss to follow up are covered under “Bias due to
missing data” (section 4.5). Examples of these types of bias are given within the relevant sections.
5
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Table 1. Bias domains included in the ROBINS-I tool
randomized trials
bias assessment is mainly distinct from assessments of
Pre-intervention or at-intervention domains for which risk of
Bias due to Selection bias as it is sometimes used in Baseline confounding occurs when one or more prognostic variables (factors
confounding relation to clinical trials (and currently in that predict the outcome of interest) also predicts the intervention received at
widespread use within Cochrane); Allocation baseline. ROBINS-I can also address time-varying confounding, which occurs
bias; Case-mix bias; Channelling bias. when individuals switch between the interventions being compared and when
post-baseline prognostic factors affect the intervention received after baseline.
Bias in selection Selection bias as it is usually used in relation to When exclusion of some eligible participants, or the initial follow up time of
of participants observational studies and sometimes used in some participants, or some outcome events, is related to both intervention and
into the study relation to clinical trials; Inception bias; Lead- outcome, there will be an association between interventions and outcome even
time bias; Immortal time bias. Note that this if the effects of the interventions are identical. This form of selection bias is
bias specifically excludes lack of external distinct from confounding. A specific example is bias due to the inclusion of
validity, which is viewed as a failure to prevalent users, rather than new users, of an intervention.
generalize or transport an unbiased (internally
valid) effect estimate to populations other
than the one from which the study population
arose.
At intervention
Bias in Misclassification bias; Information bias; Recall Bias introduced by either differential or non-differential misclassification of
classification of bias; Measurement bias; Observer bias. intervention status. Non-differential misclassification is unrelated to the
interventions outcome and will usually bias the estimated effect of intervention towards the
null. Differential misclassification occurs when misclassification of
intervention status is related to the outcome or the risk of the outcome, and is
likely to lead to bias.
6
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Post-intervention
randomized trials
substantial overlap with assessments of
Post-intervention domains for which there is
Bias due to Performance bias; Time-varying confounding Bias that arises when there are systematic differences between experimental
deviations from intervention and comparator groups in the care provided, which represent a
intended deviation from the intended intervention(s). Assessment of bias in this domain
interventions will depend on the type of effect of interest (either the effect of assignment to
intervention or the effect of starting and adhering to intervention).
Bias due to Attrition bias; Selection bias as it is sometimes Bias that arises when later follow-up is missing for individuals initially included
missing data used in relation to observational studies and followed (e.g. differential loss to follow-up that is affected by prognostic
factors); bias due to exclusion of individuals with missing information about
intervention status or other variables such as confounders.
Bias in Detection bias; Recall bias; Information bias; Bias introduced by either differential or non-differential errors in measurement
measurement of Misclassification bias; Observer bias; of outcome data. Such bias can arise when outcome assessors are aware of
outcomes Measurement bias intervention status, if different methods are used to assess outcomes in
different intervention groups, or if measurement errors are related to
intervention status or effects.
Bias in selection Outcome reporting bias; Analysis reporting Selective reporting of results in a way that depends on the findings.
of the reported bias
result
7
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
2.4 Study designs
This document relates most closely to NRSIs with cohort-like designs, such as cohort studies, quasi-randomized
trials and other concurrently controlled studies. Much of the material is also relevant to designs such as case-
control studies, cross-sectional studies, interrupted time series and controlled before-after studies, although we
are currently considering whether modifications to the signalling questions are required for these other types of
studies.
8
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
3 Guidance for using the tool: general considerations
3.1.2 Listing the confounding domains relevant to all or most studies eligible for the review
Relevant confounding domains are the prognostic factors that predict whether an individual receives one or the
other intervention of interest. They are likely to be identified both through the knowledge of subject matter
experts who are members of the review group, and through initial (scoping) reviews of the literature. Discussions
with health professionals who make intervention decisions for the target patient or population groups may also
be helpful. These issues are discussed further in section 4.1.
3.1.3 Listing the possible co-interventions that could differ between intervention groups and have an
impact on study outcomes
Relevant co-interventions are the interventions or exposures that individuals might receive after or with initiation
of the intervention of interest, which are related to the intervention received and which are prognostic for the
outcome of interest. These are also likely to be identified through the expert knowledge of members of the review
group, via initial (scoping) reviews of the literature, and after discussions with health professionals. These issues
are discussed further in section 4.4.
9
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 1: The ROBINS-I tool (Stage 1): At protocol stage
List co-interventions that could be different between intervention groups and that could impact on outcomes
10
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
3.2 Preliminary considerations for each study
11
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
the intervention as described in the trial protocol is starting and adhering to experimental intervention versus
starting and adhering to comparator intervention unless medical reasons (e.g. toxicity) indicate discontinuation.
For example, in a study of cancer screening the effect of interest might relate either to receipt (or not) of an
invitation to screening (the effect estimated in an ITT analysis of a randomized trial of screening), or to uptake
(or not) of an invitation to screening.
For both randomized trials and NRSI, unbiased estimation of the effect of starting and adhering to intervention
requires appropriate adjustment for prognostic factors that predict deviations from the intended interventions
(“time-varying confounders”, see detailed discussion in sections 4.1.9 and 4.4). Review authors should seek
specialist advice when assessing intervention effects estimated using methods that adjust for time-varying
confounding.
In both randomized trials and NRSI, risk of bias assessments should be in relation to a specified effect of interest.
When the effect of interest is that of assignment to the intervention at baseline (randomized trials) or
starting intervention at baseline (NRSI), risk of bias assessments for both types of study need not be
concerned with post-baseline deviations from intended interventions that reflect the natural course of
events (for example, a departure from randomized intervention that was clinically necessary because of a sudden
worsening of the patient’s condition) rather than potentially biased actions of researchers. When the effect of
interest is starting and adhering to the intended intervention, risk of bias assessments of both randomized trials
and NRSI may have to consider adherence and differences in additional interventions (“co-interventions”)
between intervention groups. More detailed discussions of these issues are provided in sections 4.1.8, 4.1.9 and
4.4.
12
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 2: The ROBINS-I tool (Stage 2, part 1): For each study: setting up the assessment
13
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Box 3: The ROBINS-I tool (Stage 2, part 2): For each study: evaluation of confounding domains and co-interventions
(ii) Additional confounding domains relevant to the setting of this particular study, or which the study authors identified as important
Confounding domain Measured variable(s) Is there evidence that controlling Is the confounding domain OPTIONAL: Is failure to adjust for this variable
for this variable was measured validly and reliably by this (alone) expected to favour the experimental
unnecessary?* variable (or these variables)? intervention or the comparator?
Favour experimental / Favour comparator / No
Yes / No / No information information
* In the context of a particular study, variables can be demonstrated not to be confounders and so not included in the analysis: (a) if they are not predictive of the outcome; (b) if they are not predictive
of intervention; or (c) because adjustment makes no or minimal difference to the estimated effect of the primary parameter. Note that “no statistically significant association” is not the same as “not
predictive”
14
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Preliminary consideration of co-interventions
Complete a row for each important co-intervention (i) listed in the review protocol; and (ii) relevant to the setting of this particular study, or which the study authors identified as
important.
“Important” co-interventions are those for which, in the context of this study, adjustment is expected to lead to a clinically important change in the estimated effect of the
intervention.
(ii) Additional co-interventions relevant to the setting of this particular study, or which the study authors identified as important
Co-intervention Is there evidence that controlling for this co-intervention Is presence of this co-intervention likely to favour
was unnecessary (e.g. because it was not administered)? outcomes in the experimental intervention or the
comparator
Favour experimental / Favour comparator / No
information
Favour experimental / Favour comparator / No
information
Favour experimental / Favour comparator / No
information
15
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
3.3 Signalling questions
A key feature of the tool is the inclusion of signalling questions within each domain of bias. These are reasonably
factual in nature and aim to facilitate judgements about the risk of bias.
The response options for the signalling questions are:
(1) Yes;
(2) Probably yes;
(3) Probably no;
(4) No; and
(5) No information.
One exception to this system is the opening signalling question (1.1, in the assessment of bias due to confounding)
does not have a “No information” option.
Some signalling questions are only answered in certain circumstances, for example if the response to a previous
question is “Yes” or “Probably yes” (or “No” or “Probably no”). When questions are not to be answered, a response
option of “Not applicable” may be selected. Responses underlined in green in the tool are potential markers for
low risk of bias, and responses in red are potential markers for a risk of bias. Where questions relate only to sign
posts to other questions, no formatting is used.
Responses of “Yes” and “Probably yes” (also of “No” and “Probably no”) have similar implications, but allow for a
distinction between something that is known and something that is likely to be the case. The former would imply
that firm evidence is available in relation to the signalling question; the latter would imply that a judgement has
been made. If measures of agreement are applied to answers to the signalling questions, we recommend grouping
these pairs of responses.
3.3.1 Free-text boxes alongside signalling questions
There is space for free text alongside each signalling question. This should be used to provide support for each
answer. Brief direct quotations from the text of the study report should be used when possible to support
responses.
16
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
(3) Serious risk of bias (the study has some important problems in this domain);
(4) Critical risk of bias (the study is too problematic in this domain to provide any useful evidence on the
effects of intervention); and
(5) No information on which to base a judgement about risk of bias for this domain.
The “low risk of bias” category exists to emphasize the distinction between randomized trials and non-
randomized studies. These distinctions apply in particular to the “pre-intervention” and “at-
intervention” domains (see Table 1). In particular, we anticipate that only rarely design features of a non-
randomized study will lead to a classification of low risk of bias due to confounding. Randomization does
not protect against post-intervention biases, and we therefore expect more overlap between assessments of
randomized trials and assessments of NRSI for the post-intervention domains. However other features of
randomized trials, such as blinding of participants, health professionals or outcome assessors, may protect against
post-intervention biases.
3.4.1 Free-text boxes alongside risk of bias judgements
There is space for free text alongside each RoB judgement to explain the reasoning that underpins the judgement.
It is essential that the reasons are provided for any judgements of “Serious” or “Critical” risk of bias.
17
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Table 2. Reaching an overall RoB judgement for a specific outcome.
Declaring a study to be at a particular level of risk of bias for an individual domain will mean that the
study as a whole has a risk of bias at least this severe (for the outcome being assessed). Therefore, a judgement
of “Serious risk of bias” within any domain should have similar implications for the study as a whole, irrespective
of which domain is being assessed.
Because it will be rare that an NRSI is judged as at low risk of bias due to confounding, we anticipate that
most NRSI will be judged as at least at moderate overall risk of bias.
The mapping of domain-level judgements to overall judgements described in Table 2 is a programmable
algorithm. However, in practice some “Serious” risks of bias (or “Moderate” risks of bias) might be considered to
be additive, so that “Serious” risks of bias in multiple domains can lead to an overall judgement of “Critical” risk
of bias (and, similarly, “Moderate” risks of bias in multiple domains can lead to an overall judgement of “Serious”
risk of bias).
18
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Table 3. Reaching an overall RoB judgement for a specific outcome.
This would give us the RoB profiles (which might accompany meta-analyses and/or GRADE assessments) shown
in Table 4.
Domain O1 O2 O3
Bias due to confounding Serious risk Moderate risk Serious risk
Bias in selection of participants into the study Low risk Low risk Low risk
Bias in classification of interventions Low risk Low risk Low risk
Bias due to deviations from intended
Moderate risk Moderate risk Moderate risk
interventions
Bias due to missing data Low risk No info No info
Bias in measurement of outcomes Low risk Low risk Serious risk
Bias in selection of the reported result Moderate risk Moderate risk Serious risk
Overall* Serious risk Moderate risk Serious risk
19
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4 Guidance for using the tool: detailed guidance for each bias domain
4.1.1 Background
A confounding domain is a pre-intervention prognostic factor that predicts whether an individual
receives one or the other intervention of interest. Some common examples are severity of pre-existing
disease, presence of comorbidities, health care utilization, adiposity, and socioeconomic status. Confounding
domains can be characterised by measuring one or more of a range of specific variables. The relevant confounding
domains vary across study settings. For example, socioeconomic status might not introduce confounding in
studies conducted in countries in which access to the interventions of interest is universal and therefore
socioeconomic status does not influence intervention received.
The tool addresses two types of confounding: baseline confounding and time-varying confounding.
We can identify two broad reasons that confounding is not fully controlled. Residual confounding occurs when
a confounding domain is measured with error, or when the relation between the confounding domain and the
outcome or exposure (depending on the analytic approach being used) is imperfectly modelled. For example, in
a NRSI comparing two antihypertensive drugs, we would expect residual confounding if pre-intervention blood
pressure was measured 3 months before the start of intervention, but the blood pressures used by clinicians to
decide between the drugs at the point of intervention were not available in our dataset. Unmeasured
confounding occurs when a confounding domain has not been measured, or when it is not controlled in the
analysis. This would be the case if no pre-intervention blood pressure measurements were available, or if the
analysis failed to control for pre-intervention blood pressure despite it being measured.
Note that when intervention decisions are made by health professionals, measurement error in the information
available to them does not necessarily introduce residual confounding. For example, pre-intervention blood
pressure will not perfectly reflect underlying blood pressure. However, if intervention decisions were made based
on two pre-intervention measurements, and these measurements were available in our dataset, it would be
possible to adjust fully for the confounding.
For some review questions the confounding may be intractable, because it is not possible to measure all the
confounding domains that influence treatment decisions. For example, consider a study of the effect of treating
type 2 diabetes with insulin when oral antidiabetic drugs fail. The patients are usually older, and doctors may,
without recording their decisions, prescribe insulin treatment mostly to those without cognitive impairment and
with sufficient manual dexterity. This creates potentially strong confounding that may not be measurable.
21
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.1.6 Control of confounding
When all confounders are measured without error, confounding may be controlled either by design (for example
by restricting eligibility to individuals who all have the same value of the baseline confounders) or through
statistical analyses that adjust (“control”) for the confounding factor(s). If, in the context of a particular study, a
confounding factor is unrelated to intervention or unrelated to outcome, then there is no need to control for it in
the analysis. It is however important to note that in this context “unrelated” means “not associated” (for example,
risk ratio close to 1) and does not mean “no statistically significant association”.
Appropriate control of confounding requires that the variables used are valid and reliable measures of the
confounding domains. In this context, “validity” refers to whether the variable or variables fully measures the
domain, while “reliability” refers to the precision of the measurement (more measurement error means less
reliability) (Streiner and Norman, 2003). For some topics, a list of valid and reliable measures of confounding
domains will be available in advance and should be specified in the review protocol. For other topics, such a list
may not be available. Study authors may cite references to support the use of a particular measure: reviewers can
then base their judgment of the validity and reliability of the measure based on these citations (Cook and
Beckman, 2006). Some authors may control for confounding variables with no indication of their validity or
reliability. In such instances, review authors should pay attention to the subjectivity of the measure. Subjective
measures based on self-report may tend to have lower validity and reliability relative to objective measures such
as clinical reports and lab findings (Cook et al, 1990).
It is important to consider whether inappropriate adjustments were made. In particular, adjusting for post-
intervention variables is usually not appropriate. Adjusting for mediating variables (those on the causal
pathway from intervention to outcome) restricts attention to the effect of intervention that does not go via the
mediator (the “direct effect”) and may introduce confounding, even for randomized trials. Adjusting for common
effects of intervention and outcome causes bias. For example, in a study comparing different antiretroviral drug
combinations it will usually be essential to adjust for pre-intervention CD4 cell count, but it would be
inappropriate to adjust for CD4 cell count 6 months after initiation of therapy.
23
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 4: The ROBINS-I tool (Stage 2, part 3): Risk of bias due to confounding
24
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Signalling questions Elaboration Response options
Questions relating to baseline confounding only
1.4. Did the authors use an appropriate Appropriate methods to control for measured confounders include stratification, regression, NA / Y / PY / PN / N /
analysis method that controlled for all the matching, standardization, and inverse probability weighting. They may control for individual NI
important confounding domains? variables or for the estimated propensity score. Inverse probability weighting is based on a
function of the propensity score. Each method depends on the assumption that there is no
unmeasured or residual confounding.
1.5. If Y/PY to 1.4: Were confounding Appropriate control of confounding requires that the variables adjusted for are valid and reliable NA / Y / PY / PN / N /
domains that were controlled for measured measures of the confounding domains. For some topics, a list of valid and reliable measures of NI
validly and reliably by the variables available confounding domains will be specified in the review protocol but for others such a list may not be
in this study? available. Study authors may cite references to support the use of a particular measure. If authors
control for confounding variables with no indication of their validity or reliability pay attention to
the subjectivity of the measure. Subjective measures (e.g. based on self-report) may have lower
validity and reliability than objective measures such as lab findings.
1.6. Did the authors control for any post- Controlling for post-intervention variables that are affected by intervention is not appropriate. NA / Y / PY / PN / N /
intervention variables that could have been Controlling for mediating variables estimates the direct effect of intervention and may introduce NI
affected by the intervention? bias. Controlling for common effects of intervention and outcome introduces bias.
Questions relating to baseline and time-varying confounding
1.7. Did the authors use an appropriate Adjustment for time-varying confounding is necessary to estimate the effect of starting and NA / Y / PY / PN / N /
analysis method that controlled for all the adhering to intervention, in both randomized trials and NRSI. Appropriate methods include those NI
important confounding domains and for time- based on inverse probability weighting. Standard regression models that include time-updated
varying confounding? confounders may be problematic if time-varying confounding is present.
1.8. If Y/PY to 1.7: Were confounding See 1.5 above. NA / Y / PY / PN / N /
domains that were controlled for measured NI
validly and reliably by the variables available
in this study?
25
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Risk of bias judgement See Table 5. Low / Moderate /
Serious / Critical / NI
Optional: What is the predicted direction of bias Can the true effect estimate be predicted to be greater or less than the estimated effect in the Favours experimental
due to confounding? study because one or more of the important confounding domains was not controlled for? / Favours comparator
Answering this question will be based on expert knowledge and results in other studies and / Unpredictable
therefore can only be completed after all of the studies in the body of evidence have been
reviewed. Consider the potential effect of each of the unmeasured domains and whether all
important confounding domains not controlled for in the analysis would be likely to change the
estimate in the same direction, or if one important confounding domain that was not controlled
for in the analysis is likely to have a dominant impact.
26
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 5: Reaching risk of bias judgements for bias due to confounding
Moderate risk of bias (the study is (i) Confounding expected, all known important confounding
sound for a non-randomized study domains appropriately measured and controlled for;
with regard to this domain but cannot and
be considered comparable to a well- (ii) Reliability and validity of measurement of important domains
performed randomized trial) were sufficient, such that we do not expect serious residual
confounding.
Serious risk of bias (the study has some (i) At least one known important domain was not appropriately
important problems) measured, or not controlled for;
or
(ii) Reliability or validity of measurement of an important domain
was low enough that we expect serious residual confounding.
Critical risk of bias (the study is too (i) Confounding inherently not controllable
problematic to provide any useful or
evidence on the effects of intervention) (ii) The use of negative controls strongly suggests unmeasured
confounding.
27
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.2 Detailed guidance: Bias in selection of participants into the study
4.2.1 Introduction
Selection bias occurs when some eligible participants, or the initial follow up time of some participants, or some
outcome events, are excluded in a way that leads to the association between intervention and outcome differing
from the association that would have been observed in the target trial. As explained in section 2.3, this
phenomenon is distinct from that of confounding, although the term selection bias is sometimes used to
mean confounding.
Our use of the term “selection bias” is intended to refer only to biases that are internal to the study, and not
to issues of indirectness (generalizability, applicability or transferability to people who were excluded from
the study) (Schunemann et al, 2013). For example, restricting the study sample to individuals free of comorbidities
may limit the generalizability of its findings to clinical practice, where comorbidities are common. However it
does not bias the estimated effect of intervention in individuals free of comorbidities.
4.2.2 When selection of participants into the study may introduce bias
Selective recruitment of participants into a study does not necessarily cause bias. For example, consider a study
that selected (at random) only half of eligible men, but included all eligible women. The effect of intervention in
men will be less precisely estimated than if all men had been included, but if the true effect of the intervention in
women is the same as in men, the effect estimate will not be biased by the selection.
Selection bias occurs when selection of participants is related to both intervention and outcome. For example,
studies of folate supplementation to prevent neural tube defects were biased because they were restricted to live
births (Hernán et al, 2002). The bias arises because stillbirths and therapeutic abortions (which were excluded
from the sample), are related to both the intervention and the outcome (Velie and Shaw, 1996, Hernán et al, 2002).
For the same reason, selection bias can occur when some follow up time is excluded from the analysis. For
example, consider the potential for bias when prevalent, rather than new (incident), users of the intervention are
included in analyses. This is analogous to starting the follow-up of the target trial some time after the start of
intervention, so that some individuals who experienced the outcome after starting the intervention will have been
excluded. This is a type of selection bias that has also been termed inception bias or lead time bias. If
participants are not followed from the start of the intervention (inception), as they would be in a randomized
trial, then a period of follow up has been excluded, and individuals who experienced the outcome soon after
intervention will be missing from analyses. The key problem is that there is no reason to expect the effect of the
intervention to be constant over time. Therefore, excluding follow up immediately after intervention may bias the
estimated effect either upwards or downwards. Studies that report estimates of the effect of intervention stratified
into follow up periods may provide information on the extent to which the effect of intervention varies with time
since the start of intervention.
For example, analyses based on prevalent users of a drug may tend to select those who tolerate the drug well:
“depletion of the susceptible” will already have taken place. As a result we may underestimate the rate of adverse
effects in the intervention group: pharmacoepidemiological studies therefore often specify that there should have
been no record of use of the drug in the previous 12 months. For example, there was an apparently increased risk
of venous thromboembolism with the newer oral contraceptive progestogens when investigated in NRSI (Ray et
al, 2003; Suissa et al, 2000).
Users of the newer agents had started treatment more recently than users of older agents and the risk of venous
thromboembolism is greatest early in the course of treatment. Contemporary methodological standards
emphasize the importance both of identifying cohorts of new users of health technologies and of commencing
follow-up from the date of the treatment decision, not commencement of treatment, in order to avoid biases like
this (Ray et al, 2003; Suissa, 2008).
A related bias – immortal time bias – occurs when the interventions are defined in such a way that there is a
period of follow up during which the outcome cannot occur. For example, a study followed cohorts of subjects
with chronic obstructive pulmonary disease or chronic heart failure and considered them to be in two groups
according to whether they received telehomecare or standard care. However, to get telehomecare, patients had
to survive for several weeks after the index hospitalization: therefore the time between hospitalization and start
of telehomecare was “immortal time”. Exclusion of this follow up period, and of the deaths that occur during the
28
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
period, will bias the study towards finding that telehomecare reduces mortality. Comparison with a target trial
should facilitate identification of such bias, because in a trial participants would be followed from the time of
randomization even if implementation of intervention occurred some time later.
4.2.4 Risk of bias assessment for bias in selection of participants into the study
The signalling questions and risk of bias assessments are given in Box 5 and Table 6.
29
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 5: The ROBINS-I tool (Stage 2, part 4): Risk of bias in selection of participants into the study
30
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 6: Reaching risk of bias judgements in selection of participants into the study
Low risk of bias (the study is (i) All participants who would have been eligible for the target trial
comparable to a well-performed were included in the study;
randomized trial with regard to this and
domain) (ii) For each participant, start of follow up and start of intervention
coincided.
Moderate risk of bias (the study is (i) Selection into the study may have been related to intervention
sound for a non-randomized study and outcome;
with regard to this domain but cannot and
be considered comparable to a well- The authors used appropriate methods to adjust for the selection
performed randomized trial) bias;
or
(ii) Start of follow up and start of intervention do not coincide for all
participants;
and
(a) the proportion of participants for which this was the case
was too low to induce important bias;
or
(b) the authors used appropriate methods to adjust for the
selection bias;
or
(c) the review authors are confident that the rate (hazard)
ratio for the effect of intervention remains constant over time.
Serious risk of bias (the study has some (i) Selection into the study was related (but not very strongly) to
important problems) intervention and outcome;
and
This could not be adjusted for in analyses;
or
(ii) Start of follow up and start of intervention do not coincide;
and
A potentially important amount of follow-up time is missing from
analyses;
and
The rate ratio is not constant over time.
Critical risk of bias (the study is too (i) Selection into the study was very strongly related to intervention
problematic to provide any useful and outcome;
evidence on the effects of intervention) and
This could not be adjusted for in analyses;
or
(ii) A substantial amount of follow-up time is likely to be missing
from analyses;
and
The rate ratio is not constant over time.
No information on which to base a No information is reported about selection of participants into the
judgement about risk of bias for this study or whether start of follow up and start of intervention coincide.
domain
31
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.3 Detailed guidance: Bias in classification of interventions
4.3.1 Introduction
Bias may be introduced if intervention status is misclassified. This is seldom a problem in randomized trials and
other experimental studies, because interventions are actively assigned by the researcher and their accurate
recording is a key feature of the study. However, in observational studies information about interventions
allocated or received needs to be collected.
Possible methods for data collection include:
systematic assessment of patients (clinical examinations, interviews, diagnostic tests);
administrative or in-house databases (prospective recording of data with no pre-specified purpose);
extraction from medical records; and
organizational records or policy documents (e.g. for organizational or public health interventions).
32
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 6: The ROBINS-I tool (Stage 2, part 5): Risk of bias in classification of interventions
3.2 Was the information used to define intervention In general, if information about interventions received is available from sources that could Y / PY / PN / N / NI
groups recorded at the start of the intervention? not have been affected by subsequent outcomes, then differential misclassification of
intervention status is unlikely. Collection of the information at the time of the intervention
makes it easier to avoid such misclassification. For population-level interventions (e.g.
measures to control air pollution), the answer to this question is likely to be ‘Yes’.
3.3 Could classification of intervention status have Collection of the information at the time of the intervention may not be sufficient to avoid Y / PY / PN / N / NI
been affected by knowledge of the outcome or risk bias. The way in which the data are collected for the purposes of the NRSI should also avoid
of the outcome? misclassification.
33
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 7: Reaching risk of bias judgements for bias in classification of interventions
Low risk of bias (the study is (i) Intervention status is well defined;
comparable to a well-performed and
randomized trial with regard to this (ii) Intervention definition is based solely on information collected at
domain) the time of intervention.
Moderate risk of bias (the study is (i) Intervention status is well defined;
sound for a non-randomized study and
with regard to this domain but cannot (ii) Some aspects of the assignments of intervention status were
be considered comparable to a well- determined retrospectively.
performed randomized trial)
Serious risk of bias (the study has some (i) Intervention status is not well defined;
important problems) or
(ii) Major aspects of the assignments of intervention status were
determined in a way that could have been affected by knowledge of
the outcome.
Critical risk of bias (the study is too (Unusual) An extremely high amount of misclassification of
problematic to provide any useful intervention status, e.g. because of unusually strong recall biases.
evidence on the effects of intervention)
4.4.1 Introduction
We consider in this domain biases that arise when there are systematic differences between the care
provided to experimental intervention and comparator groups, beyond the assigned interventions.
These differences reflect additional aspects of care, or intended aspects of care that were not delivered.
It is important to distinguish between:
(a) deviations from intended intervention that arise because of knowledge of the intervention applied and
because of expectation of finding a difference between experimental intervention and comparator
consistent with the hypothesis being tested in the study. Such deviations are not part of usual practice.
(b) deviations from intended intervention that happen during usual clinical care following the intervention (for
example, cessation of a drug intervention because of acute toxicity); and
34
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
from the intended interventions for deviations of type (b). Specifically, if all deviations from intended intervention
are part of usual practice, then we can still evaluate the effect of assignment to intervention, regardless of the
actual implementation of the interventions.
On the other hand, if the goal is the unbiased estimation of starting and adhering to intervention, then all
deviations from the target “protocol” will lead to bias. For example, an open-label study compared respiratory
tract infection (RTI) rates after minimally invasive or open surgery for oesophageal cancer. There were two
important differences between intervention groups in the delivery of co-interventions. First, one-lung mechanical
ventilation (which is thought to increase respiratory complications, including RTIs) was used in the open surgery
group, whereas the minimally invasive group underwent two lung ventilation. Second, epidural analgesia was
used more frequently in the open surgery group: patients with epidurals are generally less mobile and thus at
increased risk of developing an RTI. These deviations from the intended interventions put the result of the study
at risk of bias in relation to the effect of starting and adhering to the intended interventions.
4.4.4 Deviations from intended intervention when assessing the effect of starting and adhering to
intervention
Consideration of co-interventions, implementation of the intervention and adherence by participants should be
assessed only when interest is in the effect of starting and adhering to the intervention.
35
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
from the intended intervention. Similarly, addition of other treatments aimed at diabetes control may be pre-
specified as part of usual clinical practice in the context of the intended intervention.
In some instances the protocol for the intended intervention specifies that the addition of other treatments is at
the discretion of physicians, but such additions differ between the intervention groups. For example, consider a
cohort study comparing rates of gastrointestinal ulcers in aspirin users and non-users. The use of proton pump
inhibitors (PPIs) to prevent bleeding in those taking aspirin is part of usual practice. If their use is pre-specified
in the study protocol then the comparison is of aspirin plus PPIs (as necessary) with non-use of aspirin, and the
PPIs should not be considered a co-intervention. However if the study aims to compare aspirin use with no use,
then PPIs may be considered a co-intervention because their greater use in the aspirin group leads to an
underestimate of the effect of aspirin on gastrointestinal ulcers. Similarly, PPIs cause diarrhoea, and a higher
frequency of diarrhoea in aspirin users may be due to proton pump inhibitor co-intervention, rather than the
aspirin.
Review authors should make in advance a list of important co-interventions that could differ between intervention
groups and could have an impact on study outcomes (see section 3.1.3). They should then consider whether they
are likely to be administered in the context of each particular study.
We suggest that review authors consider whether the critical co-interventions are balanced between intervention
groups. If effective co-interventions are not balanced, performance bias is probably present. However, if the co-
interventions are balanced across intervention groups, there is still a risk that intervention groups will differ in
their management or behaviour beyond the intervention comparison of interest. This will be the case if a co-
intervention interacts with the experimental and comparator intervention in different ways (for example, it
enhances the effect of one intervention, but has no effect on the other).
37
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.4.5 Risk of bias assessment for bias due to deviations from intended interventions
The signalling questions and risk of bias assessments are given in Box 7 and Table 8.
We are aware that review authors would find it extremely useful if we could provide guidance on criteria that
should be used to judge co-interventions to be “balanced”, or the amount of adherence to intervention to be
“high”. Unfortunately, we do not believe that simple guidance will be generally applicable: a small absolute
difference in the numbers of patients receiving an important co-intervention might lead to substantial bias if the
co-intervention strongly influenced the outcome and patients in whom the outcome occurred were usually those
who received the co-intervention.
We recommend that review teams ensure that judgements of “balanced” co-intervention, “successful”
implementation and lack of adherence are applied consistently across the studies included in their
review.
38
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 7: The ROBINS-I tool (Stage 2, part 6): Risk of bias due to deviations from intended interventions
4.2. If Y/PY to 4.1: Were these deviations Deviations from intended interventions that do not reflect usual practice will be important if NA / Y / PY / PN / N / NI
from intended intervention unbalanced they affect the outcome, but not otherwise. Furthermore, bias will arise only if there is
between groups and likely to have imbalance in the deviations across the two groups.
affected the outcome?
If your aim for this study is to assess the effect of starting and adhering to intervention, answer questions 4.3 to 4.6
4.3. Were important co-interventions Risk of bias will be higher if unplanned co-interventions were implemented in a way that Y / PY / PN / N / NI
balanced across intervention groups? would bias the estimated effect of intervention. Co-interventions will be important if they
affect the outcome, but not otherwise. Bias will arise only if there is imbalance in such co-
interventions between the intervention groups. Consider the co-interventions, including any
pre-specified co-interventions, that are likely to affect the outcome and to have been
administered in this study. Consider whether these co-interventions are balanced between
intervention groups.
4.4. Was the intervention implemented Risk of bias will be higher if the intervention was not implemented as intended by, for Y / PY / PN / N / NI
successfully for most participants? example, the health care professionals delivering care during the trial. Consider whether
implementation of the intervention was successful for most participants.
39
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Signalling questions Elaboration Response options
4.5. Did study participants adhere to the Risk of bias will be higher if participants did not adhere to the intervention as intended. Lack Y / PY / PN / N / NI
assigned intervention regimen? of adherence includes imperfect compliance, cessation of intervention, crossovers to the
comparator intervention and switches to another active intervention. Consider available
information on the proportion of study participants who continued with their assigned
intervention throughout follow up, and answer ‘No’ or ‘Probably No’ if this proportion is high
enough to raise concerns. Answer ‘Yes’ for studies of interventions that are administered
once, so that imperfect adherence is not possible.
We distinguish between analyses where follow-up time after interventions switches
(including cessation of intervention) is assigned to (1) the new intervention or (2) the original
intervention. (1) is addressed under time-varying confounding, and should not be considered
further here.
4.6. If N/PN to 4.3, 4.4 or 4.5: Was an It is possible to conduct an analysis that corrects for some types of deviation from the NA / Y / PY / PN / N / NI
appropriate analysis used to estimate the intended intervention. Examples of appropriate analysis strategies include inverse
effect of starting and adhering to the probability weighting or instrumental variable estimation. It is possible that a paper reports
intervention? such an analysis without reporting information on the deviations from intended
intervention, but it would be hard to judge such an analysis to be appropriate in the absence
of such information. Specialist advice may be needed to assess studies that used these
approaches.
If everyone in one group received a co-intervention, adjustments cannot be made to
overcome this.
Risk of bias judgement See Table 8. Low / Moderate /
Serious / Critical / NI
Optional: What is the predicted direction If the likely direction of bias can be predicted, it is helpful to state this. The direction might Favours experimental /
of bias due to deviations from the be characterized either as being towards (or away from) the null, or as being in favour of one Favours comparator /
intended interventions? of the interventions. Towards null /Away from
null / Unpredictable
40
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 8: Reaching risk of bias judgements for bias due to deviations from intended interventions
41
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Critical risk of bias (the study is too Effect of assignment to intervention:
problematic to provide any useful There were substantial deviations from usual practice that were
evidence on the effects of intervention) unbalanced between the intervention groups and likely to have
affected the outcome.
No information on which to base a No information is reported on whether there is deviation from the
judgement about risk of bias for this intended intervention.
domain
42
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.5 Detailed guidance: Bias due to missing data
4.5.1 Introduction
Missing data may arise, among other reasons, through attrition (loss to follow up), missed appointments,
incomplete data collection and by participants being excluded from analysis by primary investigators. In NRSI,
data may be missing for baseline characteristics (including interventions received or baseline confounders), for
outcome measurements, for other variables involved in the analysis or a combination of these. A general rule for
consideration of bias due to missing data is that we should assume that an analysis using the data we intended to
collect (were they available) would produce an unbiased effect estimate, so that we concentrate only on biases
that might be introduced by the missing data.
The starting point for considering risk of bias due to missing outcome data is to clarify the nature of the
comparison of interest, particularly with regard to the distinction between assignment to (or start of) intervention
and starting and adhering to intervention (see section 3.2.2). For example, the “complete” data set would be
different for a comparison between those who were and were not offered screening and a comparison between
those who did and did not attend screening. Therefore the definition of missing data would also be different. In
order to consider missing outcome data, it is therefore important that a study sample is clearly defined at the
outset. This can be achieved through consideration of the target randomized trial.
43
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 8: The ROBINS-I tool (Stage 2, part 7): Risk of bias due to missing data
44
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 9: Reaching risk of bias judgements for bias due to missing data
Low risk of bias (the study is (i) Data were reasonably complete;
comparable to a well-performed or
randomized trial with regard to this (ii) Proportions of and reasons for missing participants were similar
domain) across intervention groups;
or
(iii) The analysis addressed missing data and is likely to have
removed any risk of bias.
Moderate risk of bias (the study is (i) Proportions of and reasons for missing participants differ slightly
sound for a non-randomized study across intervention groups;
with regard to this domain but cannot and
be considered comparable to a well- (ii) The analysis is unlikely to have removed the risk of bias arising
performed randomized trial) from the missing data.
Serious risk of bias (the study has some (i) Proportions of missing participants differ substantially across
important problems) interventions;
or
Reasons for missingness differ substantially across interventions;
and
(ii) The analysis is unlikely to have removed the risk of bias arising
from the missing data;
or
Missing data were addressed inappropriately in the analysis;
or
The nature of the missing data means that the risk of bias cannot
be removed through appropriate analysis.
Critical risk of bias (the study is too (i) (Unusual) There were critical differences between interventions in
problematic to provide any useful participants with missing data;
evidence on the effects of intervention) and
(ii) Missing data were not, or could not, be addressed through
appropriate analysis.
No information on which to base a No information is reported about missing data or the potential for
judgement about risk of bias for this data to be missing.
domain
45
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.6 Detailed guidance: Bias in measurement of outcomes
4.6.1 Introduction
Bias may be introduced if outcomes are misclassified or measured with error. Possible methods for data
collection include:
systematic assessment of patients (clinical examinations, interviews, diagnostic tests);
administrative or in-house databases (prospective recording of data with no pre-specified purpose); and
extraction from medical records; and
organizational records or policy documents (e.g. for organizational or public health outcomes).
46
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 9: The ROBINS-I tool (Stage 2, part 8): Risk of bias in measurement of outcomes
47
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 10: Reaching risk of bias judgements for bias in measurement of outcomes
Low risk of bias (the study is (i) The methods of outcome assessment were comparable across
comparable to a well-performed intervention groups;
randomized trial with regard to this and
domain) (ii) The outcome measure was unlikely to be influenced by
knowledge of the intervention received by study participants (i.e. is
objective) or the outcome assessors were unaware of the intervention
received by study participants;
and
(iii) Any error in measuring the outcome is unrelated to intervention
status.
Moderate risk of bias (the study is (i) The methods of outcome assessment were comparable across
sound for a non-randomized study intervention groups;
with regard to this domain but cannot and
be considered comparable to a well- (ii) The outcome measure is only minimally influenced by knowledge
performed randomized trial) of the intervention received by study participants;
and
(iii) Any error in measuring the outcome is only minimally related to
intervention status.
Serious risk of bias (the study has some (i) The methods of outcome assessment were not comparable across
important problems) intervention groups;
or
(ii) The outcome measure was subjective (i.e. vulnerable to influence
by knowledge of the intervention received by study participants);
and
The outcome was assessed by assessors aware of the intervention
received by study participants;
or
(iii) Error in measuring the outcome was related to intervention
status.
Critical risk of bias (the study is too The methods of outcome assessment were so different that they
problematic to provide any useful cannot reasonably be compared across intervention groups.
evidence on the effects of intervention)
48
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
4.7 Detailed guidance: Bias in selection of the reported result
4.7.1 Introduction
In this document we define: an outcome domain as a true state or endpoint of interest, irrespective of how it is
measured (e.g. presence or severity of depression), an outcome measurement as a specific measurement made
on the study participants (e.g. measurement of depression using the Hamilton rating scale 6 weeks after initiation
of treatment) and an outcome analysis as a specific result obtained by analysing one or more outcome
measurements (e.g. the difference in mean change in Hamilton rating scale scores from baseline to 6 weeks
between intervention and control groups).
4.7.2 Exclusion of outcome non-reporting bias from the risk of bias tool
Selective reporting within clinical trials has to date mainly been described with respect to the failure to report, or
partial reporting of, outcome domains that were measured and analysed (Kirkham et al, 2010). Outcome reporting
bias (ORB) arises when the outcome domain is not reported or partially reported based on the direction,
magnitude or statistical significance of its association with intervention group. The presence of such bias in one
or more of the studies included in a systematic review puts the treatment effect estimate reported by the
systematic review at risk of bias (usually in the direction of exaggeration of the magnitude of effect).
The proposed new structure of the RoB tool considers this kind of selective outcome reporting as analogous to
publication bias. Therefore, it is proposed to appraise this kind of selective outcome reporting using a different
mechanism (e.g. as part a GRADE assessment in the Summary of Findings), not as part of the RoB tool. This is a
notable departure from the current Cochrane RoB tool for randomized trials. We therefore do not include
signalling questions for selective non-reporting (or insufficient reporting) of outcome domains in this document.
We recommend the Kirkham et al (2010) framework for considering this kind of selective outcome reporting.
49
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Selective reporting can arise for both harms and benefits, although the motivations (and direction of bias)
underlying selective reporting of effect estimates for harms and benefits may differ. Selective reporting typically
arises from a desire for findings to be newsworthy, or sufficiently noteworthy to merit publication, and this could
be the case if previous evidence (or a prior hypothesis) is either supported or contradicted.
These types of selective reporting apply to all cohort study designs, irrespective of whether they involve clustering.
Selective reporting is more likely to arise in studies which have exploratory objectives because, by their nature,
such studies often involve inspecting many associations between multiple interventions or multiple outcomes.
However, an exploratory study that fully reported all associations investigated would not be at risk of selective
reporting: it is selective reporting that it is the problem, not the exploratory nature of the objective per se.
4.7.5 Risk of bias assessment for bias in selection of the reported result
The signalling questions and risk of bias assessments are given in Box 10 and Table 11.
50
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
Box 10: The ROBINS-I tool (Stage 2, part 9): Risk of bias in selection of the reported result
51
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-NoDerivatives 4.0 International License.
Table 11: Reaching risk of bias judgements for bias in selection of the reported result
Low risk of bias (the study is There is clear evidence (usually through examination of a pre-
comparable to a well-performed registered protocol or statistical analysis plan) that all reported
randomized trial with regard to this results correspond to all intended outcomes, analyses and sub-
domain) cohorts.
Moderate risk of bias (the study is (i) The outcome measurements and analyses are consistent with an a
sound for a non-randomized study priori plan; or are clearly defined and both internally and externally
with regard to this domain but cannot consistent;
be considered comparable to a well- and
performed randomized trial) (ii) There is no indication of selection of the reported analysis from
among multiple analyses;
and
(iii) There is no indication of selection of the cohort or subgroups for
analysis and reporting on the basis of the results.
Serious risk of bias (the study has some (i) Outcomes are defined in different ways in the methods and
important problems) results sections, or in different publications of the study;
or
(ii) There is a high risk of selective reporting from among multiple
analyses;
or
(iii) The cohort or subgroup is selected from a larger study for
analysis and appears to be reported on the basis of the results.
Critical risk of bias (the study is too (i) There is evidence or strong suspicion of selective reporting of
problematic to provide any useful results;
evidence on the effects of intervention) and
(ii) The unreported results are likely to be substantially different
from the reported results.
No information on which to base a There is too little information to make a judgement (for example if
judgement about risk of bias for this only an abstract is available for the study).
domain.
52
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.
5 References
Cook DA, Beckman TJ. Current concepts in validity and reliability for psychometric instruments: theory and
application. The American Journal of Medicine 2006; 119: 166.e7-166.e16.
Cook TD, Campbell DT, Peracchio L. Quasi experimentation. In: Dunnette MD, Hough LM (Eds.), Handbook of
Industrial and Organizational Psychology (pp. 491-567). Palo Alto, CA, US: Consulting Psychologists Press, 1990.
Davey Smith G, Phillips AN, Neaton JD. Smoking as “independent” risk factor for suicide: illustration of an
artifact from observational epidemiology? Lancet 1992; 340: 709-12.
Feinstein AR. An additional basic science for clinical medicine: II. The limitations of randomized trials. Annals
of Internal Medicine 1983;99:544-50.
Hernán MA. With great data comes great responsibility. Publishing comparative effectiveness research in
Epidemiology [editorial]. Epidemiology 2011; 22: 290-291.
Hernán MA, Hernandez-Diaz S, Robins JM. A structural approach to selection bias. Epidemiology 2004;15(5):615-
25.
Hernán MA, Hernandez-Diaz S, Werler MM, et al, Causal knowledge as a prerequisite for confounding
evaluation: an application to birth defects epidemiology. Am J Epidemiol 2002;155(2):176-84.
Higgins JPT, Altman DG, Gøtzsche PC, Jüni P, Moher D, Oxman AD, Savovic J, Schulz KF, Weeks L, Sterne JAC,
Cochrane Bias Methods Group, Cochrane Statistical Methods Group. The Cochrane Collaboration's tool for
assessing risk of bias in randomised trials. BMJ 2011; 343: d5928.
Institute of Medicine. Ethical and Scientific Issues in Studying the Safety of Approved Drugs. Washington, DC:
The National Academies Press, 2012.
Jackson LA, Jackson ML, Nelson JC, Neuzil KM, Weiss NS. Evidence of bias in estimates of influenza vaccine
effectiveness in seniors. International Journal of Epidemiology 2006; 35: 337–344.
Kirkham JJ, Dwan KM, Altman DG, Gamble C, Dodd S, Smyth R, Williamson PR. The impact of outcome
reporting bias in randomised controlled trials on a cohort of systematic reviews. BMJ 2010; 340: c365.
Lipsitch M, Tchetgen Tchetgen E, Cohen T. Negative controls: a tool for detecting confounding and bias in
observational studies. Epidemiology 2010; 21: 383-8.
Magid DJ, Shetterly SM, Margolis KL, Tavel HM, O’Connor PJ, Selby JV, Ho PM. Comparative effectiveness of
angiotensin-converting enzyme inhibitors versus beta-blocker as second-line therapy for hypertension.
Circulation: Cardiovascular Quality and Outcomes 2010; 3: 453-458.
Majeed AW, Troy G, Nicholl JP, et al, Randomized, prospective, single-blind comparison of laparascopic versus
small-incision cholesystectomy. Lancet 1996; 347: 989-94.
McMahon AJ, Russell IT, Baxter JN, et al, Laparascopic versus minilaparotomy cholesystectomy: a randomised
trial. Lancet 1994; 343: 135-8.
Ray WA. Evaluating medication effects outside of clinical trials: new-user designs. American Journal of
Epidemiology 2003;158(9):915-20.
Schünemann HJ, Tugwell P, Reeves BC, et al, Non-randomized studies as a source of complementary, sequential
or replacement evidence for randomized controlled trials in systematic reviews on the effects of interventions.
Research Synthesis Methods 2013; 4: 49-62.
Streiner DL, Norman GR. Health measurement scales: a practical guide to their development and use. 3rd ed. New
York: Oxford University Press, 2003.
Suissa S, Spitzer WO, Rainville B, et al, Recurrent use of newer oral contraceptives and the risk of venous
thromboembolism. Human Reproduction 2000;15(4):817-21.
Suissa S. Immortal time bias in pharmaco-epidemiology. Am J Epidemiol 2008;167(4):492-9.
Rothman KJ and Greenland S. Modern Epidemiology. 2nd ed. Philadelphia: Lippincott Williams & Wilkins 1998.
Velie EM, Shaw GM. Impact of prenatal diagnosis and elective termination on prevalence and risk estimates of
neural tube defects in California, 1989-1991. American Journal of Epidemiology 1996;144(5):473-9.
Whiting PF, Rutjes AW, Westwood ME, Mallett S, Deeks JJ, Reitsma JB, Leeflang MM, Sterne JA, Bossuyt PM;
QUADAS-2 Group. QUADAS-2: a revised tool for the quality assessment of diagnostic accuracy studies. Annals
of Internal Medicine 2011; 155: 529-36.
53
© 2016 by the authors. This work is licensed under a Creative Commons Attribution-NonCommercial-
NoDerivatives 4.0 International License.