1 / 25

Meta-analysis

Session 2.3: Assessing Quality. Funded through the ESRC’s Researcher Development Initiative. Meta-analysis. Department of Education, University of Oxford. Session 2.3: Assessing Quality. Steps in a meta-analysis. Study quality. Does it make a difference?

marlee
Télécharger la présentation

Meta-analysis

An Image/Link below is provided (as is) to download presentation Download Policy: Content on the Website is provided to you AS IS for your information and personal use and may not be sold / licensed / shared on other websites without getting consent from its author. Content is provided to you AS IS for your information and personal use only. Download presentation by click this link. While downloading, if for some reason you are not able to download a presentation, the publisher may have deleted the file from their server. During download, if you can't get a presentation, the file might be deleted by the publisher.

E N D

Presentation Transcript


  1. Session 2.3: Assessing Quality Funded through the ESRC’s Researcher Development Initiative Meta-analysis Department of Education, University of Oxford

  2. Session 2.3: Assessing Quality Steps in a meta-analysis

  3. Study quality Does it make a difference? How do we assess primary study quality? 3

  4. Study quality • The quality of evidence generated by a review depends entirely on the quality of primary studies which make up the review • Garbage-in, garbage-out! • Quality assessment helps to set apart a meta-analysis or systematic review from a narrative review

  5. What does ‘quality’ mean? • In meta-analysis, quality refers to methodological quality (the internal validity of primary studies) • High quality reduces the risk of bias by eliminating or taking into account sources of bias such as: • Selection bias • Information bias • Confounding

  6. Study quality • Increasingly, meta-analysts evaluate the quality of each study included in a meta-analysis • Sometimes this is a global holistic (subjective) rating. In this case it is important to have multiple raters to establish inter-rater agreement • Sometimes study quality is quantified in relation to objective criteria of a good study, e.g. • larger sample sizes; • more representative samples; • better measures; • use of random assignment; • appropriate control for potential bias; • double blinding, and • low attrition rates (particularly for longitudinal studies) 6

  7. In a meta-analysis of Social Science meta-analyses, Wilson & Lipsey (1993) found an effect size of .50. They evaluated how this was related to study quality: For meta-analyses providing a global (subjective) rating of the quality of each study, there was no significant difference between high and low quality studies; the average correlations between effect size and quality was almost exactly zero. Almost no difference between effect sizes based on random- and non-random assignment (effect sizes slightly larger for random assignment). The only study quality characteristic to make a difference was positively biased effects due to one-group pre/post design with no control group at all Study quality in the social sciences 7

  8. Goldring (1990) evaluated the effects of gifted education programs on achievement. She found a positive effect, but emphasised that findings were questionable because of weak studies: 21 of the 24 studies were unpublished and only one used random assignment. Effects varied with matching procedures: largest effects for achievement outcomes were for studies in which all non-equivalent groups' differences controlled by only one pretest variable. Effect sizes reduced as the number of control variables increase and disappeared altogether with random assignment. Goldring (1990, p. 324) concluded policy makers need to be aware of the limitations of the GAT literature. Study quality in the social sciences 8

  9. Schulz (1995) evaluated study quality in 250 randomized clinical trials (RCTs) from 33 meta-analyses. Poor quality studies led to positively biased estimates: lack of concealment (30-41%), lack of double-blind (17%), participants excluded after randomization (NS). Moher et al. (1998) reanalysed 127 RCTs randomized clinical trials from 11 meta-analyses for study quality. Low quality trials resulted in significantly larger effect sizes, 30-50% exaggeration in estimates of treatment efficacy. Wood et al. (2008) evaluated study quality (1,346 RCTs from 146 meta-analyses. subjective outcomes: inadequate/unclear concealment & lack of blinding resulted in substantial biases. objective outcomes: no significant effects. conclusion: systematic reviewers should assess risk of bias. Study quality in medicine 9

  10. Meta-analyses should always include subjective and/or objective indicators of study quality. In Social Sciences, there is some evidence that studies with highly inadequate control for pre-existing differences leads to inflated effect sizes. However, it is surprising that other indicators of study quality make so little difference. In medical research, studies are largely limited to RCTs where there is MUCH more control than in social science research. Here, there is evidence that inadequate concealment of assignment and lack of double-blind inflate effect sizes, but perhaps only for subjective outcomes. These issues are likely to be idiosyncratic to individual discipline areas and research questions. Study quality: Does it make a difference? 10

  11. Evaluating study quality 11

  12. Study quality • It is important to code the study quality characteristics • Juni, Witschi, Bloch, and Egger (1999): • Evaluation of scales designed to assess the quality of randomized field trials in medicine • Used an identical set of 17 studies • Applied the quality weightings dictated by 25 different scales • Seven of the scales showed that high quality trials showed an effect whereas low quality trials did not. • Six of the scales found that high quality trials showed no effect whereas low quality trials did (the reverse conclusion). • For the remaining 12 scales, effect estimates were similar across the quality levels • Overall summary quality scores were not significantly associated with treatment effects. • In summary: the scale used to evaluate study quality can determine whether a difference in quality levels is detected 12

  13. Evaluating study quality • Requires designing the code materials to include adequate questions about the study design and reporting • It requires skill and training in identifying quality characteristics • May require additional analyses: • Quality weighting (Rosenthal, 1991) • Use of kappa statistic in determining validity of quality filtering for meta-analysis (Sands & Murphy, 1996) • Regression with “quality” as a predictor of effect size (see Valentine & Cooper, 2008) 13

  14. Quality assessment • Uses of information about quality: • Narrative discussion of impact of quality on results • Display study quality and results in a tabular format • Weight the data by quality - not usually recommended because scales are not always consistent (see Juni et al., 1999; Valentine & Cooper, 2008) • Subgroup analysis by quality • Include quality as a covariate in meta-regression

  15. Valentine & Cooper (2008) • Developed an instrument for assessing study quality for inclusion in meta-analysis and systematic review • “focuses on the operational details of studies and results in a profile of scores instead of a single score to represent study quality (p. 130)” • Study Design and Implementation Assessment Device (DIAD) • Hierarchical: consists of “global”, “composite”, and “design and implementation” questions 15

  16. The four global questions of the Study DIAD • From Valentine & Cooper (2008, p. 139).

  17. Design and implementation questions Valentine & Cooper (2008) • Multiple questions within each global and composite question • Example excerpt of Table 4 (p. 144)

  18. What about non-experimental studies?? • Critical Appraisal Skills Programme (CASP) to help with the process of critically appraising articles: • Systematic Reviews (!) • Randomised Controlled Trials (RCTs) • Qualitative Research • Economic Evaluation Studies • Cohort Studies • Case Control Studies • Diagnostic Test Studies • Good start, but not comprehensive • Download from http://www.phru.nhs.uk/Pages/PHD/resources.htm

  19. Critical Appraisal Skills Programme (CASP) • Example from the CASP Cohort Studies form • http://www.phru.nhs.uk/Pages/PHD/resources.htm

  20. Also consult • Centre for Reviews and Dissemination (CRD), University of York • Report on “Study quality assessment” www.york.ac.uk/inst/crd/pdf/crdreport4_ph5.pdf • Development of quality assessment instruments • Also, quality assessment of: • effectiveness studies • accuracy studies • qualitative research • economic evaluations

  21. Example from CRD report • Some questions to consider regarding quality in case series designs (can also be used for survey data designs)

  22. Non-medical/health care guidelines?? • Evidence for Policy and Practice Information and Co-ordinating Centre (EPPI-Centre), Institute of Education, University of London • Some issues to consider when coding are listed in the Guidelines for the REPOrting of primary empirical research Studies in Education (The REPOSE Guidelines) Not as detailed/thorough (prescriptive?) as in the medical research guidelines...

  23. Work backwards • When designing your code materials, you can look at guidelines for what should be reported, and turn those into questions to evaluate quality • For example, Strengthening the Reporting of Observational Studiesin Epidemiology (STROBE) • STROBE Statement—Checklist of items that should be included in reports of cross-sectional studies • Point 13(b): Give reasons for non-participation at each stage • Could be rephrased in the coding materials as: • Did the study give a reason for non-participation? • STROBE checklists available for cohort, case-control, and cross-sectional studies at http://www.strobe-statement.org/Checklist.html

  24. Caveats & considerations... • Quality of reporting • It is often hard to separate quality of reporting from methodological quality - “Not reported” is not always “Not done” • Should code “Unspecified” as distinct from “Criteria not met” • Consult as many materials as possible when developing coding materials • There are some good references for systematic reviews that also apply to meta-analysis • Torgerson’s (2003) book • Gough’s (2007) framework • Search Cochrane Collaboration (http://www.cochrane.org/) for “assessing quality”

  25. References • Gough, D. (2007). Weight of evidence: a framework for the appraisal of the quality and relevance of evidence. In J. Furlong, A. Oancea (Eds.) Applied and Practice-based Research. Special Edition of Research Papers in Education, 22, 213-228. • Juni, P., Witschi, A., Bloch R., & Egger, M. (1999). The hazards of scoring the quality of clinical trials for meta-analysis, JAMA, 282,1054–1060 • Rosenthal, R. (1991). Quality-weighting of studies in meta-analytic research. Psychotherapy Research, 1, 25-28. • Sands, M. L., & Murphy, J. R. (1996). Use of kappa statistic in determining validity of quality filtering for meta-analysis: A case study of the health effects of electromagnetic radiation. Journal of Clinical Epidemiology, 49, 1045-1051. • Torgerson, C. (2003). Systematic reviews. UK: Continuum International. • Valentine, J. C., & Cooper, H. M. (2008). A systematic and transparent approach for assessing the methodological quality of intervention effectiveness research: The Study Design and Implementation Assessment Device (Study DIAD). Psychological Methods, 13, 130-149.

More Related