Ebp CH-5
Ebp CH-5
Many physiotherapists experience a common frustration. When they consult the research literature for answers
to clinical questions they are confronted by a range of studies with very different conclusions.
Consider, for example, the findings that confront a physiotherapist who would like to know whether
acupuncture protects against exercise- induced asthma. One study, by Fung et al (1986) concluded ‘acupuncture
provided better protection against exercise-induced asthma than did sham acupuncture’. On the other hand,
Gruber et al (2002) concluded ‘acupuncture treatment offers no protection against exercise-induced
bronchoconstriction’.
These conclusions appear inconsistent. It seems implausible that both could be true. Situations like this, where
similar studies draw contradictory conclusions, often arise.
Why is the literature apparently so
inconsistent?
One set of criteria that is of particular interest is the Delphi list of criteria for
assessing the quality of clinical trials, developed by Verhagen and colleagues
(Verhagen et al 1998a)
CRITICAL APPRAISAL OF EVIDENCE ABOUT THE
EFFECTS OF INTERVENTION
control
groups comparability can only be assured by randomly assigning
comparable subjects to groups.
Concealment
of allocation (Concealment of allocation is commonly misunderstood
to mean blinding. Blinding and concealment are quite
different features of clinical trials. It would probably be
clearer if concealment of allocation was called
concealment of recruitment) .
This means that the researcher is not aware, at the time
a decision is made about eligibility of a person to
participate in the trial, if that per- son would
subsequently be randomized to the treatment or control
group.
Concealment is important because, even though most
trials specify inclusion and exclusion criteria that
determine who is and who is not eligible to participate
in the trial, there is sometimes uncertainty about
whether a particular patient satisfies those criteria, and
Concealment often the researcher responsible for entering new
patients into the trial has some latitude in such
of allocation decisions.
• Doing clinical trials is hard and often mundane work. One of the difficulties is
ensuring that the trial protocol is adhered to. And one of the hardest parts of the
trial protocol to adhere to is the planned measurement of outcomes (‘follow-up’).
Was there complete or near complete
follow-up?
• Most clinical trials involve interventions that are implemented over days
or weeks or months. Usually outcomes are assessed at the end of the
intervention, and they are often also assessed at one or several times after
the intervention has ceased. Trials of chronic conditions may assess out-
comes several years after the intervention period has ceased.
Problems with Follow-up
• A problem that arises in most trials is that it is not always possible to obtain
outcome measures as planned. Occasionally subjects die. Others become too sick
to measure, or they move out of town, or go on long holidays. Some may lose
interest in participating in the study or simply be too busy to attend for follow-up
appointments
Loss to follow-up/ Dropouts
• For these and a myriad of other reasons it may be impossible for the researchers
to obtain outcome measures from all subjects as planned, no matter how hard
the researchers try to obtain follow-up measures from all patients. This
phenomenon of real-life clinical trials is termed ‘loss to follow-up’. Subjects lost to
follow-up are sometimes called ‘dropouts
Random Vs Non Random dropouts
• When this occurs, differences between groups are no longer attributable just to
the intervention and chance
Hypothetical Example
• Imagine a hypothetical trial of treatment for cervical headache. The trial compares the
effect of six sessions of manual therapy to a no-intervention control condition, and
outcomes in both groups are assessed 2 weeks after randomization.
• Some subjects in the control group may experience little resolution of their symptoms.
Understandably, these subjects may become dissatisfied with participation in the trial and
may be reluctant to return for outcome assessment after not having received any
intervention.
• The consequence is that there may be a tendency for those subjects in the control group
with the worst outcomes to be lost to follow-up, more so than in the treated group. In that
case, estimates of the effects of intervention are likely to be biased and the treatment will
appear less effective than it really is.
Acceptable range of Dropouts
• Protocol violations occur when the trial is not carried out as planned. In trials of physiotherapy
interventions, the most common protocol violation is the failure of subjects to receive the intended
intervention.
• For example, subjects in a trial of exercise may be allocated to an exercise group but may fail to do
their exercises, or fail to exercise according to the protocol (this is sometimes called ‘non-
compliance’ or ‘non-adherence’), or subjects allocated to the control condition may take up exercise.
• Protocol violations are undesirable, but usually some degree of protocol violations cannot be
avoided. Usually they present less of a problem than loss to follow-up.
What to do in case of Protocol violation ?
• The most satisfactory solution is the least obvious one. It involves ignoring the protocol violations and
analysing the data of all subjects in the groups to which they were allocated. This is called ‘analysis by
intention to treat’
• it maintains the comparability of groups. Also, from a pragmatic point of view, analysis by intention to
treat provides the most meaningful estimates of effects of intervention. This is because, pragmatically
speaking, interventions can only be effective if patients comply
3. Was there blinding to allocation of patients and assessors?
• There is reason to prefer that, in clinical trials, subjects are unaware of whether
they received the intervention or control condition. This is called blinding of
subjects.
• Blinding of subjects is considered important because it provides a means of
controlling for placebo effects.
• The general approach involves giving a ‘sham’ intervention to the control group.
• Sham interventions are those that look, feel, sound, smell and taste like the
intervention but could not effect the presumed mechanism of the intervention.
Assessor Masking
• If a systematic review is to produce valid conclusions it must identify most of the relevant studies
that exist and produce a balanced synthesis of their findings. To determine if this goal has been
achieved, readers can ask three questions.
1. Was it clear which trials were to be reviewed?
2. Were most relevant studies reviewed?
3. Was the quality of the reviewed studies taken into account?
Was it clear which trials were to be reviewed?
• When we read systematic reviews we need to be satisfied that the reviewer has not
selectively reviewed those trials which support his or her own point of view
• To reduce the possibility of selective reviewing, reviewers should clearly define the scope
of the review prior to undertaking a search for relevant trials. The best way to do this is
to clearly describe criteria that are used to decide what sorts of trials will be included in
the review, and perhaps also which trials will not. The inclusion and exclusion criteria
usually refer to the population, interventions and outcomes of interest.
Was it clear which trials were to be
reviewed?
• Systematic reviews which specify clear inclusion and exclusion
criteria provide stronger evidence of effects of therapy than
those that do not.
Were most relevant studies reviewed?
• Randomized trials in physiotherapy are indexed across a range of partially overlapping major
medical literature databases such as Medline, Embase, CINAHL, AMED, and PsycINFO. The
Cochrane Collaboration’s Register of Clinical Trials and the Centre for Evidence-Based Physio-
therapy’s PEDro database attempt to provide more complete indexes of the clinical trial literature,
but they rely on other databases to locate trials.
Were most relevant studies reviewed?
• Some trials are not indexed on any databases, or are so poorly indexed that they
are unlikely ever to be found. they may be published in obscure journals, or they
may not have been published at all.
Publication/ Language Bias
• Systematic reviews which search only for published trials are said to be exposed
to ‘publication bias’, and systematic reviews which search only for trials reported
in English are said to be exposed to ‘language bias’.
Was the quality of the reviewed studies taken into account?
• Many randomized trials are poorly designed and provide potentially seriously
biased estimates of the effects of intervention. Consequently, if a systematic
review is to obtain an unbiased estimate of the effects of intervention, it must
ignore low quality studies.
Consequently, the criteria used to appraise qualitative research must differ from
those used to appraise quantitative research.
Critical Appraisal of Evidence about
Experience
• When critically appraising the methodological quality of qualitative research, you
need to ask questions that focus on other elements and issues than those that
are relevant to research which includes numbers and graphs.
• individual people,
• situations,
• social settings,
• social interactions or documents.
The sample is usually strategically selected to contain subjects with relevant roles,
perspectives or experiences.
purposive sampling
researcher to another person, and that person to one more, and so on. This is
• why the participants in the study were the most appropriate to provide access to
the type of knowledge sought by the study.
• If there have been any problems with recruitment (for example, if there were
many people that were invited to participate but chose not to take part), this
should be reported.
2. Was the data collection sufficient to cover the phenomena?
• Were the data detailed enough to interpret what was being researched?
2. Was the data collection sufficient to
cover the phenomena?
• A range of very different methods is used to collect data in qualitative research. These vary from,
for example,
• participant observations,
• in-depth interviews,
• focus groups,
• document analysis.
The data collection method should be relevant and address the questions raised, and should be
justified in the research report.
Was the data collection sufficient to cover
the phenomena?
• Data collection should be comprehensive enough in both breadth (type of
observations) and depth (extent of each type of observation) to generate and
support the interpretations.
• Was it clear how the researchers derived categories or themes from the
data, and how they arrived at the conclusion?
• The aim of this process is often to make sense of an enormous amount of text,
tape recordings or video materials by reducing, summarizing and interpreting the
data.
Triangulation
• There are several features that can strengthen a reader’s trust in the findings of a
study.
• One is the use by the researchers of more than one source for information when
studying the phenomena, for example the use of both observation and
interviews. This is often called triangulation.
Triangulation
• Triangulation might involve the use of more than one method, more than
one researcher or analyst, or more than one theory.
• The use of more than one investigator to collect and analyse the raw data
(multiple coders) also strengthens the study.
CRITICAL APPRAISAL OF EVIDENCE ABOUT PROGNOSIS
• When we read studies of prognosis we first need to know which population the
study is seeking to provide a prognosis for (the ‘target population’).
• The target population is defined by the criteria used to determine who was
eligible to participate in the study. Most studies of prognosis describe a list of
inclusion and exclusion criteria that clearly identify the target population.
1. Was there representative sampling from
a well-defined population?
• When you read studies looking for information about
prognosis, start by looking to see whether the study
recruited ‘all’ patients or ‘consecutive cases’. If it did not, the
study may provide biased estimates of the true prognosis.
2. Was there an inception cohort?
• survivor cohorts of life-threatening diseases might generate unrealistically good prognoses. Either
way, survivor cohorts potentially provide biased estimates of prognosis.
• The solution is to recruit subjects at a uniform (usually early) point in the course of the disease.39
Studies which recruit subjects in this way are said to recruit ‘inception cohorts’ because subjects
were identified as closely as possible to the inception of the condition. The advantage of inception
cohorts is that they are not exposed to the biases inherent in studies of survivor cohorts.
• Studies that recruit inception cohorts may provide less biased estimates of prognosis than
studies that recruit survivor cohorts.
3. Was there complete or near-complete follow-up?
• Like clinical trials of effects of therapy, prognostic studies can be biased by loss to
follow-up. Bias occurs if those lost to follow-up have, on average, different
outcomes to those who were followed up.