The Palgrave Handbook of Survey Research
eBook - ePub

The Palgrave Handbook of Survey Research

  1. English
  2. ePUB (mobile friendly)
  3. Available on iOS & Android
eBook - ePub

The Palgrave Handbook of Survey Research

Book details
Book preview
Table of contents
Citations

About This Book

This handbook is a comprehensive reference guide for researchers, funding agencies and organizations engaged in survey research. Drawing on research from a world-class team of experts, this collection addresses the challenges facing survey-based data collection today as well as the potential opportunities presented by new approaches to survey research, including in the development of policy. It examines innovations in survey methodology and how survey scholars and practitioners should think about survey data in the context of the explosion of new digital sources of data. The Handbook is divided into four key sections: the challenges faced in conventional survey research; opportunities to expand data collection; methods of linking survey data with external sources; and, improving research transparency and data dissemination, with a focus on data curation, evaluating the usability of survey project websites, and the credibility of survey-based social science.

Chapter 23 of thisbook is open access under a CC BY 4.0 license at link.springer.com.

Frequently asked questions

Simply head over to the account section in settings and click on ā€œCancel Subscriptionā€ - itā€™s as simple as that. After you cancel, your membership will stay active for the remainder of the time youā€™ve paid for. Learn more here.
At the moment all of our mobile-responsive ePub books are available to download via the app. Most of our PDFs are also available to download and we're working on making the final remaining ones downloadable now. Learn more here.
Both plans give you full access to the library and all of Perlegoā€™s features. The only differences are the price and subscription period: With the annual plan youā€™ll save around 30% compared to 12 months on the monthly plan.
We are an online textbook subscription service, where you can get access to an entire online library for less than the price of a single book per month. With over 1 million books across 1000+ topics, weā€™ve got you covered! Learn more here.
Look out for the read-aloud symbol on your next book to see if you can listen to it. The read-aloud tool reads text aloud for you, highlighting the text as it is being read. You can pause it, speed it up and slow it down. Learn more here.
Yes, you can access The Palgrave Handbook of Survey Research by David L. Vannette, Jon A. Krosnick, David L. Vannette,Jon A. Krosnick in PDF and/or ePUB format, as well as other popular books in Politics & International Relations & Public Policy. We have over one million books available in our catalogue for you to explore.
Section 1Conventional Survey Research
Ā© The Author(s) 2018
David L. Vannette and Jon A. Krosnick (eds.)The Palgrave Handbook of Survey Research https://doi.org/10.1007/978-3-319-54395-6_1
Begin Abstract

1. Assessing the Accuracy of Survey Research

Jon A. Krosnick1
(1)
Departments of Communication, Political Science, and Psychology, Stanford University, Stanford, CA, USA
Jon A. Krosnick

Jon A. Krosnick

is the Frederic O. Glover Professor in Humanities and Social Sciences at Stanford University, Stanford, CA, USA and a University Fellow at Resources for the Future. This work was supported by National Science Foundation Award [1256359].
End Abstract
Although research on the accuracy of surveys is important, it has not received the attention it deserves. Many articles and books have focused on survey errors resulting from issues relating to coverage, sampling, non-response, and measurement, but very little work has comprehensively evaluated survey accuracy.
Research on survey accuracy may be scarce because it requires having an external measure of the ā€œtrueā€ values of a variable in order to be able to judge how well that value is measured by a survey question. For example, in the area of voting behavior, self-reports of turnout are often collected in surveys and compared with the official turnout statistics provided by the Federal Election Commission (FEC) after an election. When these sources yielded different rates, the errors have usually been assumed to be in the self-reports; the FEC numbers are assumed to document the truth.
Studies that have assessed survey accuracy have not yet been integrated into a single comprehensive review. Chang et al. (working paper) conducted such a review, the results of which constitute the first-ever meta-analysis of survey accuracy. The authors identified four principal methods for assessing the accuracy of survey results and collected published studies using each method. These studies assessed accuracy in a wide range of domains, including behaviors in the arenas of healthcare utilization, crime, voting, media use, and smoking, and measures of respondent characteristics such as demographics, height, and weight.
First, the authors identified 555 studies that matched each respondentā€™s self-report data with objective individual records of the same phenomena, resulting in a dataset of over 520,000 individual matches. This method of verification indicated that for more than 85 percent of the measurements, there was perfect agreement between the survey data and the objective records or measures. Second, the investigators found 399 studies that matched one-time aggregate survey percentages and means with available benchmarks from non-survey data. These studies involved different units of measurement, such as percentages, means in centimeters, kilograms, days, hours, drinks, etc. This assessment method indicated that survey measures matched benchmarks exactly in 8 percent of the instances, 38 percent manifested almost perfect matches (less than one-unit difference), and 73 percent manifested very close matches (less than five-unit difference). Third, the authors found 168 instances in which studies correlated individualsā€™ self-reports in surveys with secondary objective data. The results from this method indicated generally strong associations between the self-reports and the secondary data. Specific results and estimates are shown in the PowerPoint materials. The authors identified six studies that correlated trends over time in self-reports and with trends in objective benchmarks. This approach documented very strong associations between the self-report survey data and trends in the objective benchmarks. Thus, in this meta-analysis, Chang and her colleagues examined over 1000 published comparisons gauging the validated accuracy of survey data, and the vast majority of survey measurements of objective phenomena were found to be extremely accurate.
When differences do occur between survey estimates and objective benchmarks, it is important to consider exactly how these differences may have arisen, rather than immediately discounting the survey data. For example, researchers tend to assume that surveys overestimate voter turnout because of respondent lying. That is, respondents are thought to believe that voting is socially desirable, and so people who didnā€™t vote may claim to have voted in order to look presentable. However, the accumulating literature suggests instead that individual survey reports may be remarkably accurate, and the problem may be that people who participate in elections also over-participate in surveys. If so, the disagreement between aggregate rates of turnout according to surveys vs. government statistics may not be due to inaccurate respondent reporting.
These findings should give survey producers, consumers, and funding agencies considerable optimism about the continued accuracy of surveys as a method of collecting data. The findings also indicate that survey research deserves its role as one of the most used and trusted methods for data collection in the social sciences.
Ā© The Author(s) 2018
David L. Vannette and Jon A. Krosnick (eds.)The Palgrave Handbook of Survey Research https://doi.org/10.1007/978-3-319-54395-6_2
Begin Abstract

2. The Importance of Probability-Based Sampling Methods for Drawing Valid Inferences

Gary Langer1
(1)
Langer Research Associates, New York, USA
Gary Langer

Gary Langer

is a survey research practitioner. He is president of Langer Research Associates and former long-time director of polling at ABC Network News. Langer is a member of the Board of Directors of the Roper Center for Public Opinion Research, a trustee of the National Council of Public Polls, and former president of the New York Chapter of the American Association for Public Opinion Research. His work has been recognized with 2 News Emmy awards, 10 Emmy nominations, and AAPORā€™s Policy Impact Award. Langer has written and lectured widely on the measurement and meaning of public opinion.
End Abstract
Before 1936, data on populations generally were collected either via a census of the entire population or ā€œconvenienceā€ sampling, such as straw polls. The latter, while quick and inexpensive, lacked a scientific, theoretical basis that would justify generalization to a broader population. Using such methods, the Literary Digest correctly predicted presidential elections from 1916 to 1932 ā€“ but the approach collapsed in 1936. The magazine sent postcards to 10 million individuals selected from subscriptions, phone books, and automobile registration records. Through sampling and self-selection bias, the 2.4 million responses disproportionately included Republicans, and the poll predicted an easy win for the losing candidate, Alf Landon.
George Gallup used quota sampling in the same election to draw a miniature of the target population in terms of demographics and partisanship. Using a much smaller sample, Gallup correctly predicted Franklin D. Rooseveltā€™s win. This set the stage for systematic sampling methods to become standard in polling and survey research. (See, e.g., Gallup and Rae 1940.)
But quota sampling turned out not to be a panacea. The approach suffered a mortal blow in the 1948 presidential election, when Gallup and others erroneously predicted victory for Thomas Dewey over Harry Truman. While a variety of factors was responsible, close study clarified the shortcomings of quota sampling. Replicating the U.S. population in terms of cross-tabulations by ethnicity, race, education, age, region, and income, using standard categories, would require 9,600 cells, indicating a need for enormous sample sizes. Further, ā€œThe microcosm idea will rarely work in a complicated social problem because we always have additional variables that may have important consequences for the outcomeā€ (Gilbert et al. 1977). And bias can be introduced through interviewersā€™ purposive selection of respondents within each quota group.
After spirited debate, survey researchers coalesced around probability sampling as a scientifically rigorous method for efficiently and cost-effectively drawing a representative sample of the population. In this technique, each individual has a known and ideally non-zero probability of selection, placing the method on firmly within the theoretical framework of inferential statistics. As put by the sampling statistician Leslie Kish, ā€œ(1) Its measurability leads to objective statistical inference, in contrast to the subjective inference from judgment sampling, and (2) Like any scientific method, it permits cumulative improvement through the separation and objective appraisal of its sources of errorsā€ (Kish 1965).
In modern times, high-quality surveys continue to rely on probability sampling. But new non-probability methods have come forward, offering data collection via social media postings and most prominently though opt-in online samples. These often are accompanied by ill-disclosed sampling, data collection, and weighting techniques, yet also with routine claims that they produce highly accurate data. Such claims need close scrutiny, on theoretical and empirical bases alike.
Opt-in surveys typically are conducted among individuals who sign up to click through questionnaires on the Internet in exchange for points redeemable for cash or gifts. Opportunities for falsification are rife, as is the risk of a cottage industry of professional survey respondents. One study (Fulgoni 2006) found that among the 10 largest opt-in survey panels, 10 percent of panelists produced 81 percent of survey responses, and 1 percent of panelists accounted for 24 percent of responses.
An example of further challenges in opt-in online surveys is their common and generally undisclosed use of routers to maximize efficiency of administration, albeit at the cost of coverage. As an illustration, participants may be asked if they are smokers; if so, are routed to a smoking survey. If not smokers, they may be asked next if they chew gum. If yes, they are routed to a gum-chewers survey. If not, they may next be asked if they use spearmint toothpaste, and so on. Unbeknownst to sponsors of the toothpaste study, smokers and gum chewers are systematically excluded from their sample.
The approach, then, raises many questions. Who joins these poll-taking clubs, what are their characteristics, and what do we know about the reliability and validity of their responses? Are respondent identities verified? Are responses validated? What sorts of quality control measures are put in place? What survey ...

Table of contents

  1. Cover
  2. Front Matter
  3. Section 1. Conventional Survey Research
  4. Section 2. Opportunities to Expand Data Collection
  5. Section 3. Linking Survey Data with External Sources
  6. Section 4. Improving Research Transparency and Data Dissemination
  7. Section 5. Detailed Chapters
  8. Back Matter