The science of survey research is complicated, but there are a few simple terms you can learn and questions you can ask when you encounter polls in your schooling and daily life. These include:
Population. Who is the population that the survey is claiming to represent? Polls can be conducted with many different populations, so it is important to know how researchers define the population under study. For example, a survey of voters may be useful for your understanding of a particular health care issue’s importance in the election, but it might not be as useful for estimating how many people have had problems paying medical bills, since lower-income people (who may be the most likely to experience bill problems) are less likely to be voters and may be left out of the study entirely.
Sampling. How did researchers reach the participants for their poll, and was it a probability or non-probability sample? In a probability-based sample, all individuals in the population under study have a known chance of being included in the survey. Such samples allow researchers to provide population estimates (within a margin of sampling error) based on a small sample of responses from that population. Examples of probability-based sampling techniques include random digit dialing (RDD), address-based sampling (ABS), registration-based sampling (RBS), and probability-based online panels. Non-probability sampling, sometimes called convenience or opt-in sampling, has become increasingly common in recent years. While non-probability surveys have some advantages for some types of studies (particularly their much lower cost), research has shown that results obtained from non-probability samples generally have greater error than those obtained from probability-based methods, particularly for certain populations.
Data collection (survey mode). While there are many ways to design a survey sample, there are also many ways to collect the data, known as the survey mode. For many years, telephone surveys were considered the gold standard because they combined a probability-based sampling design with a live interviewer. Survey methodology is more complicated now, but it is still important to know whether the data was collected via telephone, online, on paper, or some other way. If phones were used, were responses collected by human interviewers or by an automatic system, sometimes known as interactive voice response (IVR) or a “robocall”? Or were responses collected via text message? Depending on the population represented, different approaches might make the most sense. For example, about 5% of adults in the U.S. are not online, and many others are less comfortable responding to survey questions on a computer or internet-connected device. While young adults may be comfortable responding to a survey via text message, many older adults still prefer to take surveys over the phone with a live interviewer. Some populations feel a greater sense of privacy when taking surveys on paper, while literacy challenges may make a phone survey more appropriate for other populations. Many researchers now combine multiple data collection modes in a single survey to make sure these different segments of the population can be represented.
Language. Was the survey conducted only in English, or were other languages offered? If the survey is attempting to represent a population with lower levels of English language proficiency, this may affect your confidence in the results.
Survey sponsor. Who conducted the survey and who paid for it? Understanding whether there is a political agenda, special interest, or business behind the poll could help you better determine the poll’s purpose as well as its credibility.
Timing. When was the survey conducted? If key events related to the survey topic occurred while the survey was in the field (e.g., an election or a major Supreme Court decision), that might have implications for your interpretation of the results.
Data quality checks. During and after data collection, what data quality checks were implemented to ensure the quality of the results? Most online surveys include special “attention check” questions designed to identify respondents who may have fabricated responses or rushed through the survey without paying attention to the questions being asked. Inclusion of these questions is a good sign that the researchers were following best practices for data collection.
Weighting. Were the results weighted to known population parameters such as age, race and ethnicity, education, and gender? Despite best efforts to draw a representative sample, all surveys are subject to what is known as “non-response bias” which results from the fact that some types of people are more likely to respond to surveys than others. Even the best sampling approaches usually fall short of reaching a representative sample, so researchers apply weighting adjustments to correct for these types of biases in the sample. When reading a survey methodology statement, it should be clear whether the data was weighted, and what source was used for the weighting targets (usually a survey from the Census or another high-quality, representative survey).
Sample size and margin of sampling error. The sample size of a survey (sometimes referred to as the N) is the number of respondents who were interviewed, and the margin of sampling error (MOSE) is a measure of uncertainty around the survey’s results, usually expressed in terms of percentage points. For example, if the survey finds 25% of respondents give a certain answer and the MOSE is plus or minus 3 percentage points, this means that if the survey was repeated 100 times with different samples, the result could be expected to be between 22%-28% in 95 of those samples. In general, a sample size of 1,000 respondents yields a MOSE of about 3 percentage points, while smaller sample sizes result in larger MOSEs and vice versa. Weighting can also affect the MOSE. When reading poll results, it is helpful to look at the N and MOSE not only for the total population surveyed, but for any key subgroups reported. This can help you better understand the level of uncertainty around a given survey estimate. The non-random nature of non-probability surveys makes it inappropriate to calculate a MOSE for these types of polls. Some researchers publish confidence estimates, sometimes called “credibility intervals,” to mimic MOSE as a measure of uncertainty, but they are not the same as a margin of sampling error. It’s also important to note that sampling error is only one source of error in any poll.
Questionnaire. Responses to survey questions can differ greatly based on how the question was phrased and what answer choices were offered, so paying attention to these details is important when evaluating a survey result. Read the question wording and ask yourself – do the answer options seem balanced? Does the question seem to be leading respondents toward a particular answer choice? If the question is on a topic that is less familiar to people, did the question explicitly offer respondents the chance to say they don’t know or are unsure how to answer? If the full questionnaire is available, it can be helpful to look at the questions that came before the question of interest, as information provided in these questions might “prime” respondents to answer in a certain way.
Transparency. There is no “gold seal” of approval for high-quality survey methods. However, in recent years, there has been an increasing focus on how transparent survey organizations are about their methods. The most transparent researchers will release a detailed methodology statement with each poll that answers the questions above, as well as the full questionnaire showing each question in the survey in the order they were asked. If you see a poll released with a one or two-sentence methodology statement and can’t find any additional information, that may indicate that the survey organization is not being transparent with its methods. The American Association for Public Opinion Research has a Transparency Initiative whose members agree to release a standard set of information about all of their surveys. For political polling, 538 recently added transparency as an element of their pollster ratings. Some news organizations also “vet” polls for transparency before reporting results, but many do not. This means that just because a poll or survey is reported in the news doesn’t necessarily mean it’s reliable. It’s always a good idea to hunt down the original survey report and see if you can find answers to at least some of the questions above before making judgments about the credibility of a poll.
Election polling vs. issue polling. Election polls – those designed at least in part to help predict the outcome of an election – are covered frequently in the media, and election outcomes are often used by journalists and pundits to comment on the accuracy of polling. Issue polls – those designed to understand the public’s views, experiences, and knowledge on different issues – differ from election polls in several important ways. Perhaps the most important difference is that, in addition to the methodological challenges noted above, election polls face the added challenge of predicting who will turn out to vote on election day. Most election polls include questions designed to help with this prediction, and several questions may be combined to create a “likely voter” model, but events or other factors may affect individual voter turnout in ways pollsters can’t anticipate. Election polls conducted months, weeks, or even days before the election also face the risk that voters will change their mind about how to vote between the time they answer the survey and when they fill out their actual ballot. Issue polls do not generally face these challenges, so it’s important to keep in mind that criticisms about the accuracy of election polls may not always apply to other types of polls.