1 / 31

The Past, Present, and Future of Social Surveys

The Past, Present, and Future of Social Surveys. Graham Kalton grahamkalton@westat.com. Context. A personal perspective on developments in survey research since the late 1950’s. Main focus: household surveys and survey statistics.

lionelc
Télécharger la présentation

The Past, Present, and Future of Social Surveys

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. The Past, Present, and Future of Social Surveys Graham Kalton grahamkalton@westat.com

  2. Context • A personal perspective on developments in survey research since the late 1950’s. • Main focus: household surveys and survey statistics. • To set the scene, I start with a brief review of the early history of survey sampling. • I then focus on the changes in the field that have taken place since the 1950’s. • I end with speculations about future directions. • “Those who don’t know history are doomed to repeat it.” ‒ Edmund Burke

  3. Early History of Survey Sampling • Kaier (1895): Representative method. • Bowley (1913): Systematic sampling of buildings in Reading (and then four other towns). Also confidence intervals. • ISI (1926): Representative method: random or purposive selection. • Neyman (1934) “On the two different aspects of the representative method: the method of stratified sampling, and the method of purposive selection.”

  4. Jerzy Neyman

  5. Neyman’s Design-Based Mode of Inference

  6. Survey Research in the 1950’s and 1960’s • Major developments in survey sampling took place after Neyman’s paper leading to books by • Yates (1949), Deming (1950), Hansen, Hurwitz & Madow (1953), Cochran (1953), and Sukhatme (1954). • Probability sampling was well established for government surveys in the U.S. and U.K. in the 1950’s. • Survey research was much simpler: • Face-to-face interviewing in general • Response rates for uncomplicated surveys were generally as high as 85% to 90%. • Few mail surveys because of response rate concerns

  7. Quota Sampling • Quota sampling was―and still is―widely used in market research (Stephan & McCarthy, 1958). • Quota controls are subgroups with known population sizes, such as age, sex and employment status. • Interviewers are instructed to obtain specified numbers of respondents in each subgroup. • Cost and speed benefits. • Sudman (1965) suggests biases of the order of 3%-5%. • Fit for purpose.

  8. Surveys Before Computers (BC)

  9. Punch Card

  10. Drivers of Major Developments since the 1960’s • Arrival of computers, the rapid and continuing advances in computing power, and the development of computer programs. • New modes of data collection. • More complex data collections. • More sophisticated user community that wants more complex estimates. • Demand for surveys of specialized subpopulations. • Declining response rates. • Rising costs. • The need for model-dependent methods.

  11. Computers and Data Collection • BC: • PAPI, with a simple questionnaire • AC: • CAPI: use of tablets for data collection • CARI for quality checking • CATI: use of laptops for telephone data collection • CASI and ACASI: Valuable for sensitive questions • Smartphones • Web • GPS for detecting fabrication

  12. Sampling Developments within the Design-Based Framework: Two Examples 1. Variance estimation • BC: Simple Taylor series for means and proportions (Keyfitz, 1957; Kish, 1957); simple replication (Mahalanobis, 1946; Deming, 1960). • AC: Linear substitute extension of the Taylor series method; balanced repeated replications and jackknife repeated replications; bootstrap. 2. Analytic techniques • BC: Totals, means, proportions. • AC: A full range of techniques, including multiple regression, chi-square tests, multi-level modeling, Cox’s proportional hazard models.

  13. Developments with a Model-Dependent Component: Compensating for Missing Data • Unit nonresponse and noncoverage • BC: Ignore (MCAR) or simple cell adjustments (MAR). Raking (Deming & Stephan, 1940). • AC: A wide range of methods including raking, propensity score weighting, calibration methods, use of CHAID, methods for MNAR situations. • Imputation • BC: Complete case analysis (MNAR). • Do not fabricate data • AC: Hot deck imputation, regression imputation methods, tree-cell imputation, cyclical imputation, multiple imputation, fractional imputation.

  14. Developments with a Model-Dependent Component: Small Area Estimation (SAE) • Early example in Hansen, Hurwitz & Madow (1953). • With Fay & Herriott (1979), SAE began to advance. • Still a reluctance to accept model-dependent estimates. • Now SAE is well-established (Rao & Molina, 2015). • Hierarchical Bayesian models can be fitted with McMC methods with modern day computing power.

  15. Nonprobability/Quasi-Probability Sample Designs Less costly: • Quota sampling • Random route sampling • WHO’s EPI methodology for childhood immunization • Web samples • Probability and nonprobability web panels. Hard-to-survey populations: • Location or venue-based sampling: men who have sex with men; illegal immigrants; homeless; nomads. • Snowball (chain-referral) sampling: IV drug users. • Respondent driven sampling: Heckathorn (1997).

  16. Internet Data Collection • Nonprobability Internet surveys (Couper, 2000) • Web surveys as entertainment, with no claim of scientific validity • Unrestricted web surveys • Volunteer opt-in panels: Volunteers sign up and, for a fee, respond to surveys from time to time. • Web scraping from such sites as Facebook and Twitter, and from searches on Google. • Weighting adjustments are used to attempt to compensate for the very unrepresentative samples.

  17. Administrative Data as a Substitute for Survey Data (Hand, 2018) • Attractions: • Less costly; data are available for all covered; data may be of higher quality and more current; may reflect what people do rather than what they say; they may provide tighter definitions. • Challenges: • Likely need extensive data cleaning; the records may not cover the full target population; and the quality of the record data may not be adequate. • When data are needed from more than one dataset: • Probabilistic record linkage methods may be needed. • Privacy and confidentiality concerns are severe. • Legislation may be needed.

  18. Administrative Data Linked to Surveys (Citro, 2014) Combined with surveys, administrative data can: • Reduce respondent burden; • Provide longitudinal data for the time before and after the survey data collection; • Provide accurate record data; • Be used for nonresponse weighting adjustments; • Be used to provide auxiliary variables in small area estimation.

  19. Classification of Survey Estimates • Descriptive estimates: • Totals, means and proportions • Analytic estimates: • Measures of association for evaluating cause-effect relationships, e.g., regression coefficients, differences in means • The role of the finite population concept is less obvious for analytic estimates.

  20. Mixed Mode of Inference for Descriptive Statistics • Some dependence on models is now inevitable when analyzing survey data. • For descriptive estimation, I view the need to use models as a response to an injury: • As a crutch to be used only to the extent that the survey data cannot fully support the desired estimates. • Dependence on models has consequences for how the quality of survey estimates should be presented to users.

  21. Describing the Uncertainty in Survey Estimates • With a perfectly executed probability sample, the precision of a survey estimate is measured by its estimated standard error or confidence interval. • The additional variance arising from nonresponse and noncoverage weighting adjustments can be captured by replicating the adjustments. • Currently this design-based approach is used. • However, under this approach the underlying MAR model is assumed true: • “All models are wrong, but some are useful.” ‒ George Box

  22. Measuring Uncertainty in Model-Dependent Estimates • Nonresponse and noncoverage adjustments are imperfect • They may reduce but they do not eliminate bias. • Weights for nonprobability samples are often adjusted so that the weighted sample counts conform to some external totals: • The biases that remain can be large and they go unmeasured. • Can better measures of uncertainty be produced? Bayesian model averaging (Lohr & Brick, 2017)?

  23. Effective Sample Size of a Nonprobability Sample

  24. Transportability of Measures of Association • For many years, internal validity was the dominant focus for randomized experiments, clinical trials, observational studies, and evaluation studies: • Do the effects apply to the study subjects? • External validity (aka generalization or transportability to other populations and subpopulations) was a lesser concern. • Transportability may be a reasonable assumption― at least approximately so―in many medical studies, but less so for social investigations. • In recent years, greater attention has been given to estimating a population average treatment effect for a specific population and to subgroup estimates.

  25. Two Large Cohort Studies • U.K. Biobank Study: 500,000 men and women aged 40-69 randomly recruited in certain areas of the U.K. between 2006 and 2010 with a 5% response rate. Electronic health records included in the follow-ups. • A “healthy volunteer” selection bias (Fry et al., 2017). • All-cause mortality half that of UK population. • Estimates of disease prevalence and incidence rates are not safely transportable to the general population (even with weighting adjustments). • With the large sample, many subpopulations can be studied. Which associations might be transportable? • The U.S. “All of Us” Program aims to recruit a million volunteers.

  26. Regression Analysis of Survey Data • Should the complex sample design be taken into account in conducting regression analyses? • Should survey weights be used and how should the variances of regression coefficients be computed? • If the regression model were correct, standard methods in general statistics should be used. • Incorporating weights simply lowers the precision of the estimates. • Since no model is correct, weights may be incorporated to estimate the best fitting (incorrect) prediction model for the finite population. • Or rather, for the superpopulation of which the finite population is a random realization.

  27. Current State of Survey Research + Ever increasing demand for social data for evidence-based policy making and for research. + An international survey research profession has emerged. + Many specialist journals and conferences • Statistics in Transition ‒ Falling response rates and increasing costs. • Adaptive designs and other methods have not had great success in counteracting these effects ‒ Competition from administrative and Internet data.

  28. Where Now? • Administrative and other data sources cannot completely replace surveys: • Some data (opinions, leisure activities, etc.) can only be obtained in person. • Surveys collect many variables that can place the responses to a particular item in context. • Surveys can produce a time series of estimates based on a common measuring instrument, whereas the rules for administrative data can change over time. • Administrative data have their own sources of error. • Researchers using administrative data need to develop metrics for evaluating the error dimensions in a way similar to the total survey error model.

  29. Nonprobability/Quasi-Probability Sampling • For cost and speed reasons, the use of these methods will likely grow, especially with Internet surveys and surveys of hard-to-survey populations. • If nonprobability or quasi-probability sampling methods is used, the data collector should provide data users with full details of the sampling method. • The reports of survey findings should be accompanied by warnings about the unknown biases in the estimates. • The user needs to be able to assess the fitness of use of the results for the intended purpose.

  30. In Summary • I do not foresee household surveys being replaced by administrative data. Rather, I expect the demand for high-quality surveys to continue to grow. • The critical problem is declining response rates. • More research on the interface between design-based and model-dependent inference is needed. • For nonresponse • For nonprobability/quasi-probability samples • With the ease of conducting Internet data collections, survey researchers need to be prepared to make the case for high-quality surveys.

  31. Concluding Remarks • The practice of survey research has shown great adaptability to changing circumstances over the past 60 years. • The profession is well equipped to adapt to the major changes now taking place.

More Related