0% found this document useful (0 votes)
82 views

Survey Sample Sizes and Margin of Error

This document discusses key concepts about margins of error in election polls. It makes 5 main points: 1) The margin of error represents the range that a poll result would fall within 95% of the time if the poll was conducted multiple times with different samples. 2) To determine if a candidate's lead is outside the margin of error, you need to calculate the margin of error for the difference between candidates, which is about twice the individual candidate margin. 3) The size of the population being polled does not affect the needed sample size or margin of error. 4) Polls need larger samples to have adequate margins of error for subgroups like gender or age. 5) Looking at

Uploaded by

vinay kumar
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
82 views

Survey Sample Sizes and Margin of Error

This document discusses key concepts about margins of error in election polls. It makes 5 main points: 1) The margin of error represents the range that a poll result would fall within 95% of the time if the poll was conducted multiple times with different samples. 2) To determine if a candidate's lead is outside the margin of error, you need to calculate the margin of error for the difference between candidates, which is about twice the individual candidate margin. 3) The size of the population being polled does not affect the needed sample size or margin of error. 4) Polls need larger samples to have adequate margins of error for subgroups like gender or age. 5) Looking at

Uploaded by

vinay kumar
Copyright
© © All Rights Reserved
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 8

Survey Sample Sizes and Margin of

Error
The most accurate survey of a group of people is a vote: Just ask everyone to
make a decision and tally the ballots. It's 100% accurate, assuming you
counted the votes correctly.

(By the way, there's a whole other topic in math that describes the errors
people can make when they try to measure things like that. But, for now, let's
assume you can count with 100% accuracy.)

Here's the problem: Running elections costs a lot of money. It's simply not
practical to conduct a public election every time you want to test a new
product or ad campaign. So companies, campaigns and news organizations
ask a randomly selected small number of people instead. The idea is that
you're surveying a sample of people who will accurately represent the beliefs
or opinions of the entire population.

But how many people do you need to ask to get a representative sample?

The best way to figure this one is to think about it backwards. Let's say you
picked a specific number of people in the United States at random. What then
is the chance that the people you picked do not accurately represent the U.S.
population as a whole? For example, what is the chance that the percentage of
those people you picked who said their favorite color was blue does not match
the percentage of people in the entire U.S. who like blue best?

Of course, our little mental exercise here assumes you didn't do anything
sneaky like phrase your question in a way to make people more or less likely to
pick blue as their favorite color. Like, say, telling people "You know, the color
blue has been linked to cancer. Now that I've told you that, what is your
favorite color?" That's called a leading question, and it's a big no-no in
surveying.

Common sense will tell you (if you listen...) that the chance that your sample is
off the mark will decrease as you add more people to your sample. In other
words, the more people you ask, the more likely you are to get a representative
sample. This is easy so far, right?
Okay, enough with the common sense. It's time for some math. (insert smirk
here) The formula that describes the relationship I just mentioned is basically
this:

The margin of error in a sample = 1 divided by the square root of the number
of people in the sample

How did someone come up with that formula, you ask? Like most formulas in
statistics, this one can trace its roots back to pathetic gamblers who were so
desperate to hit the jackpot that they'd even stoop to mathematics for an
"edge." If you really want to know the gory details, the formula is derived from
the standard deviation of the proportion of times that a researcher gets a
sample "right," given a whole bunch of samples.

Which is mathematical jargon for..."Trust me. It works, okay?"

So a sample of just 1,600 people gives you a margin of error of 2.5 percent,
which is pretty darn good for a poll.

You've probably heard that term — "margin of error" — a lot before. Reporters
throw it around like a hot potato — like if they linger with it too long (say, by
trying to explain what it means), they'll just get burned. That's because many
reporters have no idea what a "margin of error" really represents.

I gave you the math up above. But let's talk about what that math represents.
When you do a poll or survey, you're making a very educated guess about what
the larger population thinks. If a poll has a margin of error of 2.5 percent, that
means that if you ran that poll 100 times — asking a different sample of people
each time — the overall percentage of people who responded the same way
would remain within 2.5 percent of your original result in at least 95 of those
100 polls.

(WARNING: Math Geek Stuff!)


Why 95 times out of 100? In reality, the margin of error is what statisticians
call a confidence interval. The math behind it is much like the math behind
the standard deviation. So you can think of the margin of error at the 95
percent confidence interval as being equal to two standard deviations in your
polling sample. Occasionally you will see surveys with a 99-percent confidence
interval, which would correspond to three standard deviations and a much
larger margin of error.
(End of Math Geek Stuff!)
If a poll says that 48 percent of registered voters surveyed are likely to vote for
Candidate A and 46 precent of those voters plan to cast their ballots for
Candidate B, you'll likely hear reporters saying that Candidate A has a two-
point lead. Now that's true in this poll, but given the likely margin of error, a
mathematician wouldn't say that Candidate A has a two-point lead in the
actual race. There's just too much of a chance that Candidate A's true support
is enough less than 48 percent and the Candidate B's true support is enough
higher than 46 percent that the two might actually be tied, or maybe even that
Candidate B might have a slight lead. You can't say for sure on the basis of a
single poll with a two-point gap.

If you want to get a more accurate picture of who's going to win the election,
you need to look at more polls. Just as asking more people in one poll helps
reduce your margin of error, looking at multiple polls can help you get a more
accurate view of what people really think. Analysts such as Nate
Silver and Sam Wang have created models that average multiple polls to help
predict which candidates are most likely to win elections. (Silver got his start
using baseball statistics to predict future on-field performance, which goes to
show that numbers can help you predict things other than elections.) In 2012,
Silver was 50-for-50 in predicting state results in the presidential election,
based on his model for averaging publicly available polls.

Now, remember that the size of the entire population doesn't matter when
you're measuring the accuracy of polls. You could have a nation of 250,000
people or 250 million and that won't affect how big your sample needs to be to
come within your desired margin of error. The Math Gods just don't care.

Sometimes you'll see polls with anywhere from 600 to 1,800 people, all
promising the same margin of error. That's because pollsters often want to
break down their poll results by the gender, age, race or income of the people
in the sample. To do that, the pollster needs to have enough women, for
example, in the overall sample to ensure a reasonable margin or error among
just the women. And the same goes for young adults, retirees, rich people,
poor people, etc. That means that in order to have a poll with a margin of error
of five percent among many different subgroups, a survey will need to include
many more than the minimum 400 people to get that five percent margin in
the overall sample.

Read the rest of Robert's statistics lessons for people who don't know math.
5 key things to know about the
margin of error in election polls
BY ANDREW MERCER
In presidential elections, even the smallest changes in horse-race poll results seem to become
imbued with deep meaning. But they are often overstated. Pollsters disclose a margin of error so that
consumers can have an understanding of how much precision they can reasonably expect. But cool-
headed reporting on polls is harder than it looks, because some of the better-known statistical rules
of thumb that a smart consumer might think apply are more nuanced than they seem. In other
words, as is so often true in life, it’s complicated.

Here are some tips on how to think about a poll’s margin of error and what it means for the different
kinds of things we often try to learn from survey data.

1What is the margin of error anyway?

Because surveys only talk to a sample of the population, we know that the result probably won’t
exactly match the “true” result that we would get if we interviewed everyone in the population. The
margin of sampling error describes how close we can reasonably expect a survey result to fall relative
to the true population value. A margin of error of plus or minus 3 percentage points at the 95%
confidence level means that if we fielded the same survey 100 times, we would expect the result to be
within 3 percentage points of the true population value 95 of those times.
The margin of error that
pollsters customarily report describes the amount of variability we can expect around an individual
candidate’s level of support. For example, in the accompanying graphic, a hypothetical Poll A shows
the Republican candidate with 48% support. A plus or minus 3 percentage point margin of error
would mean that 48% Republican support is within the range of what we would expect if the true
level of support in the full population lies somewhere 3 points in either direction – i.e., between 45%
and 51%.

2How do I know if a candidate’s lead is ‘outside the margin of error’?

News reports about polling will often say that a candidate’s lead is “outside the margin of error” to
indicate that a candidate’s lead is greater than what we would expect from sampling error, or that a
race is “a statistical tie” if it’s too close to call. It is not enough for one candidate to be ahead by more
than the margin of error that is reported for individual candidates (i.e., ahead by more than 3 points,
in our example). To determine whether or not the race is too close to call, we need to calculate a new
margin of error for the difference between the two candidates’ levels of support. The size of this
margin is generally about twice that of the margin for an individual candidate. The larger margin of
error is due to the fact that if the Republican share is too high by chance, it follows that the
Democratic share is likely too low, and vice versa.

For Poll A, the 3-percentage-point margin of error for each candidate individually becomes
approximately a 6-point margin of error for the difference between the two. This means that
although we have observed a 5-point lead for the Republican, we could reasonably expect their true
position relative to the Democrat to lie somewhere between –1 and +11 percentage points. The
Republican would need to be ahead by 6 percentage points or more for us to be confident that the
lead is not simply the result of sampling error.

In Poll B, which also has a 3-point margin of error for each individual candidate and a 6-point
margin for the difference, the Republican lead of 8 percentage points is large enough that it is
unlikely to be due to sampling error alone.

3How do I know if there has been a change in the race?

With new polling numbers coming out daily, it is common to see media reports that describe a
candidate’s lead as growing or shrinking from poll to poll. But how can we distinguish real change
from statistical noise? As with the difference between two candidates, the margin of error for the
difference between two polls may be larger than you think.

In the example in our graphic, the Republican candidate moves from a lead of 5 percentage points in
Poll A to a lead of 8 points in Poll B, for a net change of +3 percentage points. But taking into account
sampling variability, the margin of error for that 3-point shift is plus or minus 8 percentage points.
In other words, the shift that we have observed is statistically consistent with anything from a 5-
point decline to an 11-point increase in the Republican’s position relative to the Democrat. This is not
to say such large shifts are likely to have actually occurred (or that no change has occurred), but
rather that we cannot reliably distinguish real change from noise based on just these two surveys.
The level of observed change from one poll to the next would need to be quite large in order for us to
say with confidence that a change in the horse-race margin is due to more than sampling variability.

Even when we do see large swings in support from one poll to the next, one should exercise caution
in accepting them at face value. From Jan. 1, 2012, through the election in November, Huffpost
Pollster listed 590 national polls on the presidential contest between Barack Obama and Mitt
Romney. Using the traditional 95% threshold, we would expect 5% (about 30) of those polls to
produce estimates that differ from the true population value by more than the margin of error. Some
of these might be quite far from the truth.

Yet often these outlier polls end up receiving a great deal of attention because they imply a big
change in the state of the race and tell a dramatic story. When confronted with a particularly
surprising or dramatic result, it’s always best to be patient and see if it is replicated in subsequent
surveys. A result that is inconsistent with other polling is not necessarily wrong, but real changes in
the state of a campaign should show up in other surveys as well.

The amount of precision that can be expected for comparisons between two polls will depend on the
details of the specific polls being compared. In practice, almost any two polls on their own will prove
insufficient for reliably measuring a change in the horse race. But a series of polls showing a gradual
increase in a candidate’s lead can often be taken as evidence for a real trend, even if the difference
between individual surveys is within the margin of error. As a general rule, looking at trends and
patterns that emerge from a number of different polls can provide more confidence than looking at
only one or two.

4How does the margin of error apply to subgroups?

Generally, the reported margin of error for a poll applies to estimates that use the whole sample (e.g.,
all adults, all registered voters or all likely voters who were surveyed). But polls often report on
subgroups, such as young people, white men or Hispanics. Because survey estimates on subgroups of
the population have fewer cases, their margins of error are larger – in some cases much larger.

A simple random sample of 1,067 cases has a margin of error of plus or minus 3 percentage points
for estimates of overall support for individual candidates. For a subgroup such as Hispanics, who
make up about 15% of the U.S. adult population, the sample size would be about 160 cases if
represented proportionately. This would mean a margin of error of plus or minus 8 percentage
points for individual candidates and a margin of error of plus or minus 16 percentage points for the
difference between two candidates. In practice, some demographic subgroups such as minorities and
young people are less likely to respond to surveys and need to be “weighted up,” meaning that
estimates for these groups often rely on even smaller sample sizes. Some polling organizations,
including Pew Research Center, report margins of error for subgroups or make them available upon
request.

5What determines the amount of error in survey estimates?

Many poll watchers know that the margin of error for a survey is driven primarily by the sample size.
But there are other factors that also affect the variability of estimates. For public opinion polls, a
particularly important contributor is weighting. Without adjustment, polls tend to overrepresent
people who are easier to reach and underrepresent those types of people who are harder to interview.
In order to make their results more representative pollsters weight their data so that it matches the
population – usually based on a number of demographic measures. Weighting is a crucial step for
avoiding biased results, but it also has the effect of making the margin of error larger. Statisticians
call this increase in variability the design effect.

It is important that pollsters take the design effect into account when they report the margin of error
for a survey. If they do not, they are claiming more precision than their survey actually warrants.
Members of the American Association for Public Opinion Research’s Transparency
Initiative (including Pew Research Center) are required to disclose how their weighting was
performed and whether or not the reported margin of error accounts for the design effect.

It is also important to bear in mind that the sampling variability described by the margin of error is
only one of many possible sources of error that can affect survey estimates. Different survey firms
use different procedures or question wording that can affect the results. Certain kinds of respondents
may be less likely to be sampled or respond to some surveys (for instance, people without internet
access cannot take online surveys). Respondents might not be candid about controversial opinions
when talking to an interviewer on the phone, or might answer in ways that present themselves in a
favorable light (such as claiming to be registered to vote when they are not).

For election surveys in particular, estimates that look at “likely voters” rely on models and
predictions about who will turn out to vote that may also introduce error. Unlike sampling error,
which can be calculated, these other sorts of error are much more difficult to quantify and are rarely
reported. But they are present nonetheless, and polling consumers should keep them in mind when
interpreting survey results.

You might also like