280 likes | 411 Vues
Public Health Performance Measurement Moving from Version 1.0 to 2.0. AcademyHealth Annual Meeting Boston, MA June 26, 2005. University of Kentucky Team Members. F. Douglas Scutchfield, MD, P.I. Evelyn A. Knight, PhD Ann V. Kelly, MHA Ilie Puiu Vasilescu, PhD (UVa-Wise)
E N D
Public Health Performance MeasurementMoving from Version 1.0 to 2.0 AcademyHealth Annual Meeting Boston, MA June 26, 2005
University of Kentucky Team Members • F. Douglas Scutchfield, MD, P.I. • Evelyn A. Knight, PhD • Ann V. Kelly, MHA • Ilie Puiu Vasilescu, PhD (UVa-Wise) • Michelyn Bhandari, MPH
Provide performance standards for public health systems Improve quality and accountability of public health practice Conduct systematic collection and analysis of performance data Develop a science-base for public health practice improvement Goals of the National Public Health Performance Standards Program(NPHPSP)
Aims of the Assessment Instruments • Provide a NPHPSP performance report to the public health agency and the state or local system. • Provide baseline, feedback and guidance to the public health system for continuous performance improvement.
Four Concepts Applied in NPHPS 1. Based on the 10 Essential Public Health Services (EPHS) 2. Focus on the overall public health system 3. Describe an optimal level of performance 4. Support a process of quality improvement
Goals of UKY’s Instrument Research(Local and State PHPA) • Decrease the burden on users due to instrument characteristics. • Assure content supports the goals of the performance assessment process. • Increase the usability of the results for performance improvement.
Research Phases • Preliminary interviews--determine areas of need for instrument improvement—state and local instruments. • Psychometric analysis--identify items that might be changed or eliminated in future versions—local instrument only. • Discussion groups--identify specific changes in question inclusion and/or wording which would improve instrument usability and usefulness—state and local instruments.
Preliminary Interviews—Results • Directions clear; some questions clear, some not • Understandable by public health--but not partners • Too long, instrument fatigue--but details important • Rating scale clear—but doesn’t correspond with report • Rating methodology varies—consensus, vote, etc. • Subjective scores—depend on who is present • Costly to gather information needed to answer the questions
Instrument Format Essential Service—Level 1 Indicator—Level 2 Model Standard Measure—Level 3 (First-tier stem) Measure—Level 4
Measure—Level 4 (Second-tier stem) Measure—Level 5 Summary Questions-- Not analyzed
Performance Scoring—Levels 3 & 2 • Level 3 stem scores are weighted scores based on stem and level 4 & 5 sub-questions. Stem question (3-digit) Sub-questions (4-digit) Sub-sub-questions (5 digit) • Level 2 scores (model standards) are average of Level 3 scores--yields 31 Indicator Scores.
Performance Scoring—Level 1 & Overall • Indicator Scores are averaged to obtain 10 EPHS scores—Level 1. • 10 EPHS scores are averaged to obtain one overall Public Health Performance Score.
Psychometric Analysis of Local Instrument • Determine Cronbach’sinternal reliability • Determine Item-Total Correlations (ITC) • Test effect on reliability of removing specific questions or tiers of questions • Use results to frame discussion group interviews
Table 1. Internal Reliability of Overall Performance Score Based on 10 EPHS Scores (n=228) For Overall Score = 0.915(very reliable) *Reliability of group to total increases slightly if EPHS #2 (item) is removed. Reliability is slightly higher with only 9 EPHS (0.918), than with all 10 EPHS (0.915).
Table 2: Internal Reliability of 10 EPHS Scores based on Indicator Scores (n=228)
Summary of Internal Reliability Overall & EPHS • Instrument has very high reliability overall—measuring the PH construct • EPHS-level reliability varies • EPHS 3,9 Very reliable • EPHS 2,4,5,10 Acceptable reliability • EPHS 8 Marginal reliability • EPHS 1,6,7 Not reliable
Table 3. Internal Reliability—Indicators based on Level 3 First-Tier Stem Question Scores
Summary of Internal Reliability of 31 Indicators • 14 indicators very reliable >0.8 • 10 indicators of acceptable reliability 0.7 to 0.8 • 3 become very reliable if one stem question is removed • 6 indicators not reliable <0.7
Psychometrics of Current and Simulated Instrument Versions (V.1-V.4) • Current Local Assessment Instrument • V.1—All questions as scored by CDC • Eliminate scores of drill-down details • V.2—No level 5 details—use only level 3&4 • V.3—No level 4&5 details—use only level 3 stem • V.4—No questions with high or ns correlation to stems at levels 4&5
Table 4. Comparison of Reliability across the Four Versions (v.1-4) for 10 EPHS
Table 5. Internal Reliability—10 EPHS Scores based on Indicator ScoresV.1 and simulated V.2-V.4
Moving from V1.0 to V2.0 • Instrument construction • Instrument content • Instrument context • Instrument consensus
Instrument Construction • Clarity • Time • Complexity • Subjectivity
Instrument Content • Examine unreliable EPHS and subscales • Determine key content • Fit content into a logic model with inputs—outputs--outcomes
Instrument Context • Large vs small systems • Centralized vs decentralized • Urban vs rural
Instrument Consensus • Agreed on definitions of PH practice • Purpose • Process • Strengths • Weaknesses
NPHPSP Instrument V2.0 should: • Build on current instrument and lessons learned. • Specify how users will validate their results. • Recognize and respond to agency--system issues. • Agree on purpose. • Use for purpose for which designed.