The Importance of Evaluation in Program Effectiveness and Resource Management
In this comprehensive overview, Dr. Kathleen Dowell delves into the crucial role that evaluation plays in assessing the merit and effectiveness of programs. Highlighting common barriers such as lack of time, resources, and understanding, she dispels myths surrounding evaluation as an extraneous or complicated process. By exploring how systematic inquiry can lead to better resource management, justification of funding, and overall program improvement, this presentation emphasizes evaluation not as a mere formality, but as an essential tool for demonstrating program value and ensuring accountability.
The Importance of Evaluation in Program Effectiveness and Resource Management
E N D
Presentation Transcript
&%$#@&%$!! Evaluation is NOT a Dirty Word Kathleen Dowell, Ph.D. EvalSolutions Epilepsy Foundation: Best Practices Institute September 29, 2012 Denver, Colorado
Barriers • Lack of research/statistics skills • Lack of time • Lack of resources • Other priorities • Lack of incentive • Fear • Don’t see value
What is Evaluation? The process of determining the merit, worth, or value of a program (Scriven, 1991)
What is Evaluation? Systematic inquiry that describes and explains, policies’ and programs’ operations, effects, justifications, and social implications (Mark, Henry, & Julnes, 2000)
What is Evaluation? The systematic application of social research procedures for assessing the conceptualization, design, implementation, and utility of social intervention programs (Rossi & Freeman, 1989)
In simpler terms….. Collection of information to determine the value of a program eVALUation
Evaluation is NOT…. • Auditing • Personnel assessment • Monitoring (although this can be part of an evaluation process) • Used to end or shut down programs
Evaluation Myth #1 Evaluation is an extraneous activity that generates lots of boring data with useless conclusions
Evaluation Myth #2 Evaluation is about proving the success or failure of a program
Evaluation Myth #3 Evaluation is a unique and complex process that occurs at a certain time in a certain way, and almost always includes the use of outside experts.
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources • Document program accomplishments
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources • Document program accomplishments • Justify current program funding
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources • Document program accomplishments • Justify current program funding • Support need for increased funding
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources • Document program accomplishments • Justify current program funding • Support need for increased funding • Satisfy ethical responsibility to clients to demonstrate positive and negative effects of participation
How Can Evaluation Help You? • Demonstrate program effectiveness or impacts • Better manage limited resources • Document program accomplishments • Justify current program funding • Support need for increased funding • Satisfy ethical responsibility to clients to demonstrate positive and negative effects of participation • Document program development and activities to help ensure successful replication
Ultimately… To improve program performance which leads to better value for your resources
No Evaluation Means…. • No evidence that your program is working or how it works
No Evaluation Means…. • No evidence that your program is working or how it works • Lack of justification for new or increased funding
No Evaluation Means…. • No evidence that your program is working or how it works • Lack of justification for new or increased funding • No marketing power for potential clients
No Evaluation Means…. • No evidence that your program is working or how it works • Lack of justification for new or increased funding • No marketing power for potential clients • Lack of credibility
No Evaluation Means…. • No evidence that your program is working or how it works • Lack of justification for new or increased funding • No marketing power for potential clients • Lack of credibility • Lack of political and/or social support
No Evaluation Means…. • No evidence that your program is working or how it works • Lack of justification for new or increased funding • No marketing power for potential clients • Lack of credibility • Lack of political and/or social support • No way to know how to improve
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative)
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative) • Outcomes
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative) • Outcomes • Indicators
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative) • Outcomes • Indicators • Measures
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative) • Outcomes • Indicators • Measures • Benchmarks
Basic Terminology • Types of Evaluation • Outcome (summative) • Process (formative) • Outcomes • Indicators • Measures • Benchmarks • Quantitative vs. qualitative
Engage Stakeholders • Those involved in program design, delivery, and/or funding • Those served by the program • Users of the evaluation results
Clearly Define Program • Resources, activities, outcomes • Context in which program operates • Logic model • Explicit connections between “how” and “what” • Helps with program improvement • Good for sharing program idea with others • Living, breathing model
IF THEN
IF THEN I take an aspirin
IF THEN My headache will go away I take an aspirin
IF = Inputs & Activities THEN = Outcomes
Written Evaluation Plan • Outcomes • Indicators • Tools • Timelines • Person(s) responsible (optional)
Credible Data Collection Tools • Valid and reliable tools • Valid=measures what it is intended to measure • Reliable=consistent results over time • Qualitative • Quantitative • Will answer your evaluation questions and inform decision-making
Collect Credible/Useful Data • Quantitative • Surveys • Tests • Skill assessments • Qualitative • Focus groups • Interviews • Journals • Observations
Analyze Data • Many methods • Answer evaluation questions • Engage stakeholders in interpretations • Justify conclusions and recommendations • Get help if needed!
Share/Use Results • Reporting format • Getting results into the right hands • Framing the results • Collaborative vs. confrontational approach • Keeping users “in the loop” • Debriefs and follow-up