CESSR Workshops in Methods Introduction to Program Evaluation September 24, 2010 Mindy Hightower King, Ph.D. Research Scientist Center for Evaluation & Education Policy Indiana University.
Download ReportTranscript CESSR Workshops in Methods Introduction to Program Evaluation September 24, 2010 Mindy Hightower King, Ph.D. Research Scientist Center for Evaluation & Education Policy Indiana University.
CESSR Workshops in Methods Introduction to Program Evaluation September 24, 2010 Mindy Hightower King, Ph.D. Research Scientist Center for Evaluation & Education Policy Indiana University 1 CEEP… •Promotes and supports rigorous program evaluation and education policy research primarily, but not exclusively, for educational, human services, and nonprofit organizations. •Takes a dynamic approach to evaluation and education policy research, using both quantitative and qualitative methodologies, including experimental designs. •Represents a merger of the Indiana Center for Evaluation and the Indiana Education Policy Center. 2 CEEP’s Mission Improve education by providing nonpartisan information, research, and evaluation on education issues to policymakers and other education stakeholders Encourage rigorous program evaluation across a variety of settings by providing evaluation expertise and services to diverse agencies, organizations, and businesses Expand knowledge of effective strategies in evaluation and policy research by developing, modeling, and disseminating innovative approaches to program evaluation and policy research 3 Current Projects CEEP researchers currently manage over 60 projects in the following areas: •Charter Schools •Professional Learning Communities’ •After School Programming •Literacy •Education Policy •Science, Technology, Engineering and Math •School Wellness •Public Health 4 Presentation Overview 1. What is Evaluation? 2. Tools of the Trade 3. A Few Words on Program Goals/Objectives 5 What is Evaluation? • The use of social science research activities directed at collecting, analyzing, interpreting, and communicating information about the workings and effectiveness of programs. • Differentiated from social science research in that: 1. Interpretation and communication of results is essential, but less standardized 2. The engagement of stakeholders and the political nature of evaluation requires additional skill sets 6 Why Conduct Evaluation? • To assess the utility of new programs or initiatives; • To satisfy accountability requirements of program sponsors; • To increase the effectiveness of program management and administration; • To aid in decisions regarding whether programs should be continued, expanded, or curtailed. 7 Who Commissions Evaluation? • Program funders • Program managers • Research organizations • Program designers • Any combination of the above 8 Some Background on the Field of Evaluation The following factors have contributed to the rise and professionalism of the field: Public health programs targeting infectious disease Post-WWII boom of federally and privately funded social programs 1960’s war on poverty More recently, the driving force behind evaluation has shifted from social science researchers to consumers of the research 9 Challenges of Evaluation • Dynamic nature of social programs Programs may change as they are implemented, and the evaluation will often need to do the same in response. • Scientific versus pragmatic concerns Challenge involves selecting techniques of the highest rigor with available resources while maintaining utility. • Diversity of perspectives and approaches There are rarely absolutes in evaluation. Most often, it involves finding the approach that best fits the situation. 10 Formative v. Summative Evaluation Formative Evaluation Summative Evaluation • Deals with program implementation • Achievement of program objectives • Geared toward program improvement • Often focus on the “bottom line • Provide feedback and advice • Can include cost-efficiency analyses • Often focused on and/or involve program managers and staff • Often requested by funders or advisory boards 11 General Steps in Conducting a Program Evaluation (The Evaluation Center, 2001) • Evaluation Assessment – Involves determination of purpose, key questions, intended use, culture/environment, research design. • Evaluation Study – Involves instrument design, data collection and analysis, report development and dissemination, utilization. 12 Evaluation Assessment 1. Who are the clients of the evaluation? 1. What are the questions and issues driving the evaluation? 2. What resources are available to do the evaluation? 1. What has been done previously? 1. What is the program all about? 13 Evaluation Assessment - Continued 6. In what kind of environment does the program operate? 7. Which research designs are desirable appropriate? 8. What information sources are available and appropriate, given the evaluation issues and environment, and the program structure? 9. Given all the issues raised in 1-8, which evaluation strategy is least problematic? 10. Should the program evaluation be undertaken? 14 Evaluation Study 1. Develop the measures and collect the data. 2. Analyze the data 3. Write the report 1. Disseminate the report / results 1. Make changes based on the data / utilize the results 15 Quick & Dirty Evaluation Design 1. Who wants to know what and why? 1. What resources are available to do the evaluation? 2. What do I need to keep in mind about the context in which the program operates? 1. What information sources are available and appropriate? 2. Which evaluation strategy is most feasible and least problematic? 16 Evaluation Tools and Strategies: Logic Models Stakeholder Interest Matrices Data Collection Plans 17 What is a Logic Model? • • • A simplified picture of a program, initiative, or intervention. Shows logical relationships among the resources that are invested, the activities that take place, and the benefits or changes that result. (This is often called program theory or the program's theory of action) It is a "plausible, sensible model of how a program is supposed to work" (Bickman, 1987). 18 Logic Model Basics INPUTS Program Investments What is invested OUTPUTS Activities What is done Participation Who is reached OUTCOMES Short Term Learning Intermediate Long Term (Impacts) Action / Performance Conditions Inputs - the resources invested that allow a program to achieve the desired outputs. Outputs - activities conducted or products created that reach targeted participants or populations. Outputs lead to outcomes. Outcomes - changes or benefits for individuals, families, groups, businesses, organizations, and communities. 19 Teaching American History Program 1. Funded by the U.S. Department of Education 2. Provides grants to schools and districts 3. Purpose of the program: increase student and teacher knowledge of American History 4. Program funds used to provide professional development to teachers, purchase materials, support collaborative efforts 20 PRACTICE EXERCISE: Developing a Logic Model: (Articulate the desired long-term outcomes and work backwards) INPUTS Program Investments STEP 3 OUTPUTS Activities OUTCOMES Participation STEP 2 Short Term Intermediate Long Term STEP 1 21 Teaching American History Logic Model INPUTS Program Investments Staff Volunteers Money Time Materials OUTPUTS Activities OUTCOMES Participation Teacher Professional Development # of teachers who attend workshops Peer Mentoring # of students impacted by trained teachers Curriculum Development Short Term Long Term Increased Teacher Knowledge in American History Increased Student Achievement in American History Technology Partners 22 Logic Models: Easy as pie…or cookies? 23 Stakeholder Interest Matrix • • • • Used to identify individuals/groups who may be interested and/or involved in an evaluation. Clarifies stakeholder interests in evaluation results, potential concerns and/or road-blocks, and opportunities to increase buy-in. Particularly useful for participatory evaluations. May also help in identifying potential evaluation resources. 24 Stakeholder Interest Matrix Stakeholder Interest in the Program Potential Potential Uses of Involvement in the Evaluation Results Evaluation 25 Stakeholder Interest Matrix Potential Potential Uses of Involvement in the Evaluation Results Evaluation Stakeholder Interest in the Program USDOE Accountability for funds; efficacy Continued program funding TA for design and implementation Students Engagement; interest Decisions to engage in learning opportunities Data source; knowledge & feedback Teachers Improved teaching skills; improved learning Decisions to engage in learning opportunities Data source; track PD information Community Partners Reaching students; Decision to partner meeting mission with schools Data source on partnerships 26 Data Collection Plans • • • Used to summarize evaluation methodology in grant proposals or to summarize for stakeholders. Illustrates the connection between project goals/objectives and data collection strategies. Particularly useful when space is limited in proposals/applications or when evaluation strategies are multi-dimensional. 27 Sample Data Collection Plan 28 A few words on goals and objectives… 29 Goals – Objectives – Performance Measures PROGRAM GOAL Project Objectives: What your project is doing to support the overall program goa?l Performance Measures: How you measure your progress toward meeting your objectives? 30 PRACTICE EXERCISE: Developing a Logic Model: (Articulate the desired long-term outcomes and work backwards) INPUTS Program Investments STEP 3 OUTPUTS Activities OUTCOMES Participation Short Term STEP 2 Intermediate Long Term STEP 1 Program Objectives Process Measures Outcome Measures 31 Performance Measures A performance measure is a measurable indicator used to determine how well objectives are being met. • How will you assess progress? • How much progress will constitute success? • How will you know if your objective or part of your objective has been achieved? 32 Relevance of Performance Measures Objective 1 Performance Measure 1a Performance Measure 1b Performance Measure 1c 33 Components of Performance Measures 1. What will change (or happen)? 2. How much change is expected? (What is the expected quantity?) 3. Who will achieve the change (or who will the events involve)? 4. When the change will take place (or happen)? 34 Performance Measure Examples Five (how much) charter schools will be developed in geographic areas with a concentration of high priority schools (as defined by state standards) (who/what) throughout the state each year between 2010 and 2012 (when). © 2010 CEEP 35 Performance Measure Examples 100% of charter school leaders and CFOs (expected quantity) will attend the Fiscal Review Workshop (what will happen/who will be involved) during years one and two of their grant period (when will it happen). 36 Objectives/Performance Measures Objective: • To encourage dissemination of best practices within charter schools to the broader public. Performance Measures: • On an annual basis, 100% of charter schools will submit their best practices to the SEA for inclusion to a catalogue of innovative methods. • During each year of the grant, at least two venues/partner organizations will disseminate collected charter school data. • Follow up surveys attending partner organization training events will show that at least 75% of those attending dissemination workshops will implement new practices based on charter school innovations. 37 For more information… Mindy Hightower King, Ph.D. Center for Evaluation and Education Policy 1900 E. Tenth Street, Room 918 Bloomington, Indiana 47401 812-855-4438 [email protected] 38 CESSR Workshops in Methods Introduction to Program Evaluation September 24, 2010 Mindy Hightower King, Ph.D. Research Scientist Center for Evaluation & Education Policy Indiana University 39