Quota Sampling Guidance 4
Quota Sampling Guidance 4
In this context, this guidance on quota sampling is both necessary and timely. The aim of the
guidance is not to produce an academic treatise – if you want detailed theory, please refer to Patrick
Sturgis’s recent work1 – but to provide practical advice.
Decisions about sampling methods are complex, hence this guidance cannot provide definitive
instructions for each situation, or proffer a simple step-by-step approach. Instead it provides
information on what to consider so that decisions on sampling methodology are fully informed.
Such decisions are rarely straightforward and are likely now more complex than ever because of
declining response rates, emerging evidence of a weak relationship between response rates and
non-response bias and developments of new online alternatives to traditional face-to-face
household random probability surveys.
These issues are all explained in the guidance along with some information on where quota samples
may be more appropriately used and how they can be improved. All this information should help to
ensure that decisions around sampling approaches are the right ones for the evidence needed.
1
A dizzying number of publications, including the inquiry he chaired to investigate inaccuracies in the opinion
polls during the 2015 General Election campaign
2
What is sampling and what are the key methods2?
Sampling is how to select ‘sample units’ (usually people, households or businesses – hereto referred
to as respondents) to participate in research. Sampling is cheaper and less burdensome than
surveying every respondent in the target population of interest (a census).
All sampling methods can be categorized as either probability or non-probability. The distinguishing
factor being, respectively, whether the probability of a respondent being sampled is known or
unknown. This guidance focuses mainly on a key non-probability method – quota sampling – but also
provides information on probability sampling methods as a contrast.
The aim of quota sampling is to control the composition of the final achieved sample ‘by design’. The
design may replicate the true composition of the population of interest, have equal numbers of
different types of respondent, or over-sample a particular type of respondent.
For example, the quota may require equal numbers from different ethnic minority groups
even though their population proportions differ.
Through appropriate design, a quota sample will be more representative (in terms of the estimates it
produces) of the population of interest than:
a self-selecting sample – where any respondent can choose to take part, usually those who
feel very strongly about a particular topic
a convenience sample – where interviewers can sample any respondent they choose and
don’t have particular quotas to meet, usually those who are easiest to contact
Quota samples vary in terms of quality. Section 5 sets out practical tips for improving the quality of
quota samples in key areas such as: the source of the sample; the quotas themselves; survey
procedures; and technical adjustments such as weighting.
Probability sampling
Probability sampling requires a list of every respondent in the population – ‘sampling frame’ – from
which respondents are selected with a known probability. See appendix 1 for the main methods.
For example, a sample of households selected from a list of all the residential addresses.
The aim of probability sampling is to generalise, or make inferences, about the whole population
sampled from, and be able to quantify the precision of this inference. The statistical theory which
underpins this only applies if the sample is unbiased – i.e. there is no systematic difference from the
‘true’ population – and each respondent in the population has a known non-zero probability of
selection.
2
See Appendix 1 for a quick guide to sampling methods
3
What are the issues with different sampling methods?
Bias and non-response
The main statistical issue is bias – whether survey results are systematically different from the
population ‘truth’. Non-response errors arise (in both probability and quota samples) when
different types of potential respondent are more or less likely to take part.
For example, a quota sample of men and women to assess prevalence of mental illness may suffer
higher non-response from people suffering mental illness, and therefore falsely under report its
prevalence.
The problem with quota samples is the sampling method itself. Until a quota is filled, the
interviewer (or in the case of some self-complete online surveys, the respondents’ themselves)
determines which respondent to sample. Interviewers will naturally target the easiest respondents
to contact, e.g. those walking by in the street where they are interviewing or those who have a
greater desire to participate in the research, i.e. those with a high ‘propensity’ to respond. This
means that they exclude respondents not easy to contact or with a low propensity to respond.
Therefore, some sections of the population will never be sampled, and if these have different survey
characteristics, the results will be biased. Even if an interviewer makes an equal effort to interview
every contactable respondent, the respondents with a high propensity to respond – by definition –
are more likely to participate, so the achieved sample would still be unbalanced and potentially
biased3.
There are particular concerns about bias in online panel surveys. In recent years, many
organisations have set up large online panels of potential respondents. They then use quota
sampling to conduct large-scale surveys with a sample of these web panel members, which
can result in achieved sample sizes of thousands of respondents within short timescales for
relatively little cost. Panel samples may be particularly prone to bias due to self-selection in
joining the panel, exclusion of the off-line population and panellists becoming “experienced”
respondents to repeated surveys.
The problem with probability samples is participation – unlike quota sampling, interviewers can't
determine who to interview and must make repeated efforts to contact their sample and can’t
substitute refusals with any person who fits the quota. Non-response does occur and evidence
suggests that it is getting harder achieve high response rates (though random probability samples
tend to have higher response rates than quota samples). Non- response may be due to the lack of
an accurate and up-to-date sampling frame, or simply propensity to respond. Importantly, a low
response rate does not automatically mean a sample is biased. Many studies have shown a weak
relationship between response rates and non-response bias (for more information on this see
Appendix 3). However, if the non-response is not random – i.e. it is associated with the intended
outcome of the survey research – and it cannot be adjusted for, it may lead to bias. In practice, it
remains difficult to measure bias for random samples.
3
Many studies have shown certain parts of the population, such as those from more disadvantaged socio-
economic backgrounds or young men are less likely to respond
4
Note that survey estimates are prone to other types of non-sampling error which might be more
substantial, in particular measurement error, which can apply to both random and quota sampling
approaches.
Although quota samples are traditionally considered quicker and cheaper than probability samples,
if due care and attention is given to training interviewers, randomly selecting data collection points –
i.e. where an interviewer samples their quota, then the costs and time spent implementing a good
quality quota sample will also be high.
4
There is no complete list or sampling frame of telephone numbers for the general population, hence
probability surveys of households are only appropriate if it’s a survey of a sub-group for whom sampling
frames with telephone contact details are available. Therefore telephone surveys for probability sample
household surveys are rare, though they are more common for business surveys where sample frames with
telephone contact details are more commonly available
5
When should quota sampling be used?
Quota sampling can be useful for a broad brush picture of attitudes, behaviours or circumstances:
understanding the range of concerns facing respondents about potential policy decisions
testing a communications intervention before wider roll-out
testing questions5 in the pilot stage of a survey
getting preliminary information on an issue
helping to scope a policy intervention
However, quota sampling should only be used in government if there are compelling reasons for not
using a probability sampling approach – which should be the default choice for survey research.
In areas like quantitative commercial research and political opinion polling quota sampling is
used as matter of course – though for the latter this is now a matter of some controversy
after recent polling data led to incorrect predictions of a number of election results (e.g. the
2015 General Election, Brexit)!
You should particularly try to avoid quota sampling in the following situations:
When you want to draw robust conclusions about the population as a whole and quantify
your precision. This includes analysing differences between sub-groups or changes over
time.
o Quota sampling only provides robust information about the responding sample
which cannot be generalised to the wider population – unlike probability sampling,
which is designed to provide information about the wider population with a certain
level of confidence.
o Quota sampling results will not be truly representative of the population even after
a great deal of effort in the design of quotas and weighting to account for all
expected differences between domains, and only pseudo-measures of precision are
possible.
If your research is about informing important policies, government forecasts,
important/controversial/high profile debate or is relied on as evidence for a select
committee. Quota samples are not suited for measuring key issues due to inherent
problems with respondent bias – participants are more likely to be willing, easily accessible
and interested in the survey topic.
5
There would need to be compelling reasons even for this use. At the Office for National Statistics, the
Questionnaire Design team in Methodology use probability sampling, not quota, for selecting samples for
piloting questions via cognitive interviewing.
6
Quick guide to: What sampling method to use?
In practice, decisions about sampling methods come down to many different factors and often need
to be taken on a case-by-case basis, ideally in consultation with a relevant statistician or researcher.
To help guide often complex decisions about which sampling methods to use, we have come up with
four golden questions plus an additional consideration. These simplify many nuances that exist
between probability and quota sampling, but are a useful quick guide.
4. How much scope is there for collecting a good quality quota sample?
Generally, the more information known about a population and subject area, and the more
focused the research question, the more potential there is for a good quality quota sample
NB Do not automatically rule out probability approaches just because budgets are small
and/or timescales are limited, and do remember that it is possible to conducting a random
probability sample household survey via online methods which is both quicker and cheaper
than a traditional face to face household survey (more information is earlier in the
document)
7
What is best practice for quota sampling?
Once you have decided to use a quota sample, it is important to ensure both high quality inputs
(design, sample sources and training) and carefully considered outputs (weighting, reporting).
Quota samples can vary considerably in their quality, and this section highlights how you can ensure
the most robust findings possible.
Design
The population characteristics used as the basis of setting quotas are most often demographic –
gender, age, ethnicity, socio-economic group, working status etc. However, you can also use
characteristics relating to opinions, attitudes and behaviours – political ideology, whether cycles,
whether goes to the cinema etc. If you don’t have high quality data to inform the sample design,
you may need to consider a probability sample.
Once you have decided characteristics and quotas, you can introduce some randomisation by, for
example, selecting locations at random and/or selecting a systematic quota sample (for example,
every tenth individual). For quotas selected from sampling frames which include contact details of
respondents, you can specify in advance a minimum number of re-contacts to be made before
replacement.
Sample sources
In some situations you may be able to use a good quality random survey sample as a starting point
for your quota sample. This will help to reduce selection biases. Use multiple sample sources can
also help to reduce selection bias.
Training
Quota samples are biased towards people who are willing, easily accessible and interested in the
survey topic. Bias is reduced for interviewer administered surveys, if interviewers convert refusals,
8
rather than replace. This means that interviews should practice and use refusal conversion
techniques and collect information on refusal rates to assess bias.
In general, interviewer recruitment and training procedures should be rigorous – and make efforts to
avoid ‘contamination’ between sampling techniques. Interviewers are generally skilled in either
quota sample interviews or random sample interviews, but often not in both, so check whether
interviewers also work on random sample surveys.
Weighting
For good quality research based on quota samples, a fully specified weighting model should adjust
the results for non-response. In order to do this, information on refusal rates should be augmented
by demographics of non-responders (i.e. paradata).
Reporting
The important messages to convey are:
why quota sampling has been used
how to interpret the results
There may be financial and other pressures to use quota sampling and other non-probability
methods – this is fine and perfectly understandable. However, the potential issues with quota
sampling need to be made clear in reporting – and not hidden.
NB There are countless examples that go against these best practice reporting guidelines.
What can be said about the population as a whole? Unless the research participants
represent non-participants – for example, share the same attitudes, beliefs, have the same
characteristics, behaviours, experiences – then quota sample results cannot be generalised
to the population. Unfortunately, quota sampling will always be biased towards respondents
willing to take part in the research – which may be very different and think very differently
about the research topic than non-participants. Therefore, quota sampling research should
only refer to survey respondents. Any generalisations to the whole population should be
strongly caveated with statements such as “If the quota sample was representative of the
whole population, this would mean…” .
Are the results ‘statistically significant? Any differences observed in the data need to be
interpreted with caution. Try to triangulate findings with other evidence (such as other
quantitative sources or themes from earlier qualitative research or comparable evidence
from other organisations or countries) or review against the theory to see if the results make
sense. There are a number of issues with using formal significance tests on quota sample
data – bias, lack of known sampling probability, unknown population. Therefore, you must
report any statistical significance testing of quota sampling results with great care. Any
reference to statistical significance should be strongly caveated with statements such as “If
the data were generated from a probability sample, the following results would be… ”. NB
Many market research agencies use significance tests as a matter of course, even though
they are, strictly, invalid.
9
Can confidence intervals be used? Again, confidence intervals do not apply for quota
samples – as they also rely on statistical theory and a known probability of selection.
Therefore, you must report on confidence intervals for quota samples with great care. Any
reference to confidence intervals should be strongly caveated with statements such as “If the
data were generated from a random probability sample, the following confidence intervals
would apply. As the data were generated from a quota sample, confidence intervals are
unknown. ”
See appendix 6 for some examples of good and bad reporting of quota sampling research.
It is also important to be as transparent as possible about survey procedures in any technical reports
or appendices. Survey contractors should be able to provide information on6:
people's willingness to take part in the survey - If lots of people are refusing to take part in
the research, this may mean that participants are more likely to be 'unusual' in some
unknown way. Contractors should be able to provide information on feasible expected, and
actual, refusal rates and outline their approaches to reduce non-response and bias.
Whether survey participants are different from non-participants with respect to the
research topic – Contractors should be able to provide insight into how key characteristics
are distributed, and whether certain groups are more/less willing to respond.
Whether the sample profile matches that of the population- it is worth checking that the
sample profile matches the population for characteristics, which have not for some reason
been used to set quotas. Mismatches in demographic characteristics may be an indicator of
other hidden imbalances or bias, which may not be corrected by weighting or other
adjustment techniques.
6
Wretman, J. (2010) Reflections on Probability vs. Non-Probability Sampling. In M. Carlson, H. Nyquist & M.
Villani (eds.), Official Statistics - Methodology and Applications in Honour of Daniel Thorburn, pp. 29-35
10
APPENDIX 1: Quick guide to key sampling methods
7
The International Passenger Survey is an exception – a systematic sample is selected of passengers as they
arrive at/depart from ports, airports, train terminals, where the only ordering is their arrival time
12
APPENDIX 3: Response rates and non-response bias
Response rates have traditionally been used as a proxy for how representative a survey is, and have
generally been viewed as being inversely related to non-response bias. That is, a high response rate
is seen as an indication that the survey has minimal non-response bias and vice versa. Response
rates across most surveys have declined over the last 15 years, with more effort needed to achieve
and maintain them, leading to rising fieldwork costs.
Recent studies8 have explored the relationship between fieldwork efforts, response rates and non-
response bias to assess whether increasing fieldwork efforts (and costs) is worth it. Such studies
indicate that there is generally a weak relationship between response rates and non-response bias.
For example, Sturgis et al (2016) compared weighted survey estimates after different numbers of
field calls (i.e. with different response rates) with the final estimates and found that higher response
rates had only a marginal impact on estimates overall. However, there were occasional exceptions
with some variables being particularly susceptible to non-response bias at lower response rates,
such as volunteering rates.
There is therefore no easy answer to what response rate is ‘good enough’ for your survey. Bolling
and Smith (2017)9 outline 3 key questions to consider:
How vulnerable are your variables of interest to non-response bias? Existing evidence
should be used to try to identify these.
Will marginal increases in non-response bias compromise your conclusions? Often trends
are more important than point estimates, and non-response bias is likely to be constant over
time.
Are there better ways of reducing non-response bias in your study? For example, instead
of trying to increase your overall response rate, target increased field efforts on selected
non-respondents.
SOURCES:
Groves and Peytcheva (2008), The impact of nonresponse rates on nonresponse bias, a meta-
analysis, Public Opinion Quarterly Vol.72 No.2, pp 167-189. Accessed at:
http://askchisne.ucla.edu/chis/tac2015/Documents/SDSM%20TAC/Groves_Peytcheva%202008.pdf
Sturgis et al (2016), Fieldwork effort, response rate, and the distribution of survey outcomes: a multi-
level meta-analysis. Accessed at:
https://eprints.soton.ac.uk/398569/1/Paper_Draft_final_RR_April16_submitted.pdf
8
For example, Groves and Peytcheva (2008) and Sturgis et al (2016).
9 th
Declining response rates and their impact, presentation at the SRA on 29 June 2017, accessible at:
http://the-sra.org.uk/wp-content/uploads/keith-bolling-and-patten-smith-declining-response-rates-and-their-
impact.pdf
13
APPENDIX 4: Golden questions
1. How are the results going to be used?
Use a quota sample if… Use a probability approach if…
… aim is a broad brush picture of people's … aim is precise estimates to inform important
attitudes, behaviours or circumstances policies, government forecasts, to monitor and
report trends or feed into wider public debate; or if
Sometimes, it's not vital to get an exact the research is controversial or high profile
result… and it is not always necessary
to be able to quantify accuracy Sometimes, it’s important that results are
representative of the population … and that
For example: accuracy can be quantified
testing questions in the piloting stage
of a survey For example:
getting preliminary information on an The Family Resources Survey collects data on
issue Social Security benefits, assets, savings, housing
helping to scope a policy intervention costs and income – because it is based on a
understanding the range of concerns large probability sample and has good response,
people face about potential policy the results are representative of the UK and
decisions their accuracy can be quantified. Hence it is
testing a communications intervention used to inform government policy and forecasts,
before wider roll-out as well as wider public debate
Research used in Select Committees needs to
use probability sampling to ensure results stand
up to scrutiny: quota sampling is easier to
discredit
2. Are conclusions needed about the wider population?
Use a quota sample if… Use a probability sample if…
... generalisation to the population is not … firm precise conclusions are needed about
needed, or indicative estimates for the wider the population as a whole
population are acceptable with some
Most of the time, survey results need to be
Survey results do not always need to be generalised to the population as a whole
generalised to the wider population
For example:
For example: the National Travel Survey uses probability
research may be in early stages – ie scoping sampling and therefore results can be
a policy or communications intervention – generalised to the whole UK population, ie
with light needed on issues re the intended “60% of women in the UK own a car”
approach before wider testing and roll-out Understanding Society uses probability
for rare subgroups for which a sampling sampling, which enables small differences
frame does not exist, probability sampling identified between subgroups to be tested
would not work but quota sampling could for statistical significance
achieve sufficient numbers to analyse – and the Labour Force Survey uses probability
point to differences, without being able to sampling, and so is able to accurately
test whether they are statistically measure changes in UK unemployment rate
significant at a point in time or in terms of changes over time
change over time
14
3. Are there technical reasons preventing probability sampling?
Use a quota sample if… Use a probability approach if…
… there is no sample frame (or contact details … you have a sample frame with suitable
on the frame are unsuitable) or a probability contact details
sample would have non-response bias
If there is a good sample frame, and can
If a sample frame is not available and achieve a good response within budget, use
cannot be built (due to time or resources) a probability sampling approach
or the frame lacks contact information, a
quota sample is a sensible alternative For example:
If non-response bias would be an issue for a the Student Income and Expenditure Survey
probability sample, a quota sample is a uses probability sampling because
sensible alternative10 universities maintain databases of students.
Hence, researchers are able to access
For example: emails, telephone numbers and postal
DfT recently conducted a quota survey of addresses, facilitating an effective survey
commuters as there was no available with sufficient budget, a longer fieldwork
population list. Instead, commuters were period and more effective data collection
intercepted at key locations, such as service modes – i.e. face-to-face surveys including
stations (car drivers) and train stations an increased number of visits to each
(train passengers) household – will ensure good response
response rates are very low for random information on non-responders from the
digit dialling, and bias can only be assessed sampling frame means potential bias can
by comparing the achieved sample to other often be adjusted (i.e. weighted) for
sources of population information
4. How much scope is there for collecting a good quality quota sample?
Use a quota sample if… Use a probability approach if…
… good knowledge of the population and … only limited knowledge of the population and
subject being researched is available, or the subject being researched is available, or the
research is focussed on a narrow policy area survey covers a broad range of unrelated areas
A good quality quota sample requires good Without knowledge of population variables
prior knowledge of population variables and sizes, appropriate quotas cannot be set
and sizes in order to set quotas so probability sampling should be used
10
Cumming (2010): "Quota sampling is not an acceptable alternative to probability sampling with a
reasonable response rate. It can however be an acceptable alternative if probability sampling is
unlikely to produce a reasonable response rate"
15
a targeted quota sample to get a snapshot and local action, volunteering, charitable
of attitudes towards a certain policy topic giving, and well-being
Final sanity check. How much time and money exists for the survey?
First investigate all options around using a If after initial investigation a probability sample
probability sampling approach within the time clearly cannot be used, a quota sample should
and budget… not be the automatic first step…
16
APPENDIX 5: More detail on post-stratification of quota samples
Simple post-stratification – to be avoided So how should post-stratification be used?
Assume an achieved quota sample from a targeted To ensure data generated through a quota
rare population happens to be 100 male and 200 sample are as representative of the general
female, with a population of 400 of each gender. population as possible, all relevant
Simple post-stratification would assign a weight of stratification domains should be used. For
0.5 to the female responses. example, if collecting information known to be
Problem – a weight of 0.5 assumes that influenced by age, sex, ethnicity, education,
the samples are drawn randomly from the and occupation – then the results should be
populations, and this is precisely what we post-stratified to known population totals for
know is not the case for quota samples each of these variables: which implicitly relies
Solution – the first 100 sampled of each on the quota sampling questionnaire collecting
sex would be likely to have the same enough demographic information to properly
‘propensity to respond’ and so would post-stratify.
probably be similar in other Knowing all key variables impacting on
characteristics, but the second 100 women response could be through a pilot
sampled would be likely to have a lower survey, previous studies or expert
propensity, so would not be similar, but analysis.
weighting would equate them to the first Post-stratification requires a highly
100 men. The unpalatable solution would regarded independent data source.
be to only use the interviews from the first While this may be feasible in some
100 women for population results. studies, in others it may be less
Rather than assume a quota sample is random, it is straightforward.
more realistic to describe it as the "lowest hanging It is also important thing that the survey data
fruit". What the quota sample actually delivers is measures the characteristic in the same way as
the sub-population with the lowest probability of the benchmark data, and that there is
refusal. sufficient sample size for each characteristic.
Researchers used data from the National Sexual Attitudes and Lifestyles Survey (NatSal) to post-
stratify a (web panel) quota sample survey on sexual behaviour. The aim was to reduce the potential
bias of the quota sample survey.
Despite post-stratification, the results of the quota sample survey and NatSal were significantly
different on most key measures. The conclusion was that the quota sample survey must be biased to
some degree – i.e. the quota sample survey participants must be different in some way to non-
participants, despite researchers' best efforts to control findings.
While it is possible that the mode (web), nature of the sample (panel) and topic area (sexual
attitudes) exacerbated these biases, it is useful to bear in mind that biases may be present when
quota sample research is used (more information in this SRA pamphlet)
17
APPENDIX 6: Reporting do’s and don’ts
What can be said about the population as a whole?
Do say… Don’t say ×
11
A significance test estimates whether differences are statistically significant but this is an estimation only
18
Can confidence intervals be used?
Do say… Don’t say…×
"While it is difficult to quantify the “Survey findings are subject to a margin
accuracy of these results in the way of error as they are based on a sample.
that you can for random variability in Findings were statistically tested at the
probability sampling, this does not 5% significance level, and only
mean that errors/bias do not exist." differences which were statistically
If the data were generated from a significant at that level are referred to
random probability sample, the in the text unless otherwise stated.”
following confidence intervals would "60 per cent of women own a car"
apply. As the data were generated from (±3%)".
a quota sample, confidence intervals “We estimate that if all adults in the
are unknown. ” population had been asked, the
proportion thinking that climate change
is one of the biggest problems we face
would probably be between 23% and
27%.”
19