170 likes | 190 Vues
Evaluation in a Changing World Reflections on the Australasian 2012 Evaluation International Conference www.aes.asn.au. John Owen Centre for Program Evaluation The University of Melbourne Canberra Evaluation Forum, 18 September, 2012. Reflection 1:.
E N D
Evaluation in a Changing WorldReflections on the Australasian 2012 Evaluation International Conferencewww.aes.asn.au John Owen Centre for Program Evaluation The University of Melbourne Canberra Evaluation Forum, 18 September, 2012
Reflection 1: Conference as an Evaluation Object [Intervention] Conference is an Event. But AES conference has no objectives: maybe resort to needs
Participant Needs (acceptable) • Acquire new theoretical understandings (guest speakers) • Access relevant knowledge (selected papers) • Target individuals with common interests (networking) • Learn new skills (workshops)
Participant Needs (less acceptable?) • Break from the workplace • Reward for diligence • Meeting with cronies • Social interactions away from home Conference Evaluation: Were Acceptable Needs Met? Little published literature.
Reflection 2: On Content: Evaluation/Intervention Interface Evaluation as systematic enquiry to (i)specify (ii) refine or (iii) assess the worth of an intervention.
Assessing Worth (Evaluation of Practice) Key concepts: settled intervention, objectives based, episodic. Example: John Liddle and Bronte Koop. Cost Benefit Analysis in Educational Evaluation: Is it worth it? Intervention Evaluation
Specify (Evaluation into Practice) Key concepts: evidence-based practice, meta analyses, program logic, policy development Example: Terry Buss. Policy-makers as a Constituency for Evidence-based Policy: Expanding the Partnership. Evaluation Intervention
Refinement (Evaluation in Practice) Key concepts: intervention evolving, rapid response, monitoring, strategic interactivity Example: Amanda Jupp. Designing a Results Based Management Approach within the Private Sector Intervention Intervention Intervention Evaluation Evaluation
Reflection 3: On Content: Emerging Practices • Evaluation Capacity Building • Evaluation of Complex Programs • Evaluation of Programs for the Disadvantaged/International Development • Realist Methods
Reflection 3: On Content: Emerging Practices • Use of Technology in Data Management Example: Robert Kirkpatrick. Harnessing the Power of Real-Time Data: A 21st Century Approach to Monitoring and Evaluation. www.unglobalpulse.org [EES Conference Helsinki, 2012]
Reflection 4: On Missing Topics • Evaluation Theory eg Theory of Valuing, etc • Evaluation Standards eg Feasibility, Propriety, etc • Relations with Cognate Disciplines (policy analysis, market research)
Reflection 5: On Ways Forward • Program Logic as Basis for Conference Evaluation (Harness Relevant Research) • Guest speakers: Must deliver on Latest Concepts/Theories (e.g. Patton in Adelaide) • Redress the Practice-Theory Balance www.aes.asn.au/2012-aes-conference-papers.html