Evaluation Plans & Performance Indicators
Download
Report
Transcript Evaluation Plans & Performance Indicators
Evaluation Plans &
Performance Indicators
Office of Research, Evaluation, and Policy Studies
Marcella M. Reca Zipp
November 30, 2010
Necessity of evaluation plan
Types of evaluation plans
Components of an evaluation plan
Performance indicators
Reporting requirements
Sample evaluation plans
Purpose of Evaluation Plan
Cohesive approach to conducting
evaluation and using results
Explains what, when, how, why, who
Documents the evaluation process
Ensures implementation fidelity
Promotes a participatory approach
Source: University of Toronto
Source: University of Toronto
3 Levels of Evaluation
Project-Level Evaluation
Context
Implementation
Outcome
Cluster Evaluation
Program and Policymaking Evaluation
Project-Level Evaluation
Context
Needs, assets, and resources of community
Political atmosphere, social and environmental
strengths/weaknesses
Implementation
Critical components/activities of project
Aspects that are strengths and weaknesses
How do the components connect to goals and
outcomes
Outcome
Critical outcomes you are trying to achieve
Impact on clients, community, etc.
Unexpected impact
Cluster Evaluation
Determines how well the
collection of projects fulfills the
objective of systemic change.
Not a substitute for project-level
evaluation.
Looks across a group of projects
to identify common themes.
Information reported in
aggregate form to granting
organization.
EVALUATOR
PROJECT A
PROJECT B
PROJECT C
Program and Policy Making
Evaluation
Macro form of evaluation.
Utilizes information gathered from
both project-level and cluster
evaluation to make effective decisions
about program funding and support.
Supports communities in creating
policy change at local, state, and
federal levels.
Elements of an Evaluation
Introduction
Project Objectives
Logic Model
Partnership Roles and Responsibilities
Intervention Programming/Research
Methodology/Data Collection
Instrumentation Measures
Introduction
Provides background information for
the evaluation, identifies purpose and
goal, sets the course on the
evaluation road map.
Evaluation purpose and goals
What does the evaluation strive to achieve?
Evaluation team
Who is the evaluation coordinator?
Who are the members of the evaluation
team?
Logic Model
Graphic depiction of the program
description.
Links needs, objectives, activities,
measurements.
Provides scope of program.
Ensures systematic decisions are
made about what will be measured.
Identifies and organizes indicators.
Objectives
PIMO method
The number of objectives will be
determined by purpose (i.e.,
intervention, treatment, prevention).
Interrelated to your projected
activities (i.e., education, service,
research).
Feasible to collect, will provide
accurate results.
Partnership Roles and
Responsibilities
Project partners are expected to provide
certain, unique expertise to your project
activities, either in direct service function or
as advisory units.
Identify each partner in their role and
responsibility in terms of involvement for
your project.
Partner-cited activities must be evaluated
formatively and summatively.
Intervention Programming
Identify one or more intervention strategies
used to support project activities and
anticipated outcomes.
Cite if the program is on the federal
evidence-based initiative list: EBI
For market-available programs that require
training and certification of direct service
providers, provide a timetable for acquiring
training before intervention can be used.
Performance Indicators
Visible, measurable signs of program
performance.
Relevant, understandable and useful.
Reflect program objectives, logic
model and evaluation questions.
Define success
Reasonable expectations of program
performance.
Source: University of Toronto
Performance Indicators cont.
Other terms – industry jargon
Key Performance Indicator (KPI)
Performance metric
Performance standard
Balanced Scorecard
Quality indicators
All are different words for the same
thing: measure performance.
Data Collection
What methods will be used?
How often will data be
collected?
Who will collect the data?
Validity and reliability of data
sources
Baseline data
Outcomes-based triangulation
Quality assurance
Design (experimental, quasiexperimental, etc.)
Instrumentation Measures
Tools for data collection
Only collect the information you need
Easy to administer and use
Pilot test tools before use in the
evaluation
Human Subjects Considerations
IRB, school board approval
Data management and storage
Confidentiality and data quality
Tips & Helpful Hints
Be realistic
In your assessment of resources
In your timeline
Seek help
Use templates, tables, or guides that
may be provided in the RFP or model
after past funded proposals.
Reporting and Dissemination
Dissemination
How will you disseminate findings?
Who is responsible?
How, where, when will findings be used?
Reporting
Formative reports – quarterly, biannually
Summative reports – final report/end of
project
Project deliverables
Sample RFP Evaluation Plan
Two examples of an evaluation plan
within an RFP
General, limited specifications
Complex, very detailed
Evaluation Resources
CDC: www.cdc.gov/eval
University of Toronto: www.utoronto.ca/shp/hcu
W.K. Kellogg Foundation:
www.wkkf.org/Publications/evalhdbk
Connell, J.P., Kubisch, A.C., Schorr, L.B., Weiss, C.H.
(1995). New Approaches to Evaluating Community
Initiatives, New York, NY: Aspen Institute.
Shadish, W.R., Cook, T.D., Leviton, L.C. (1991).
Foundations of Program Evaluation. Newbury Park, CA:
Sage Publications.
Taylor-Powell, E., Steele, S., Douglas, M. (1996).
Planning a Program Evaluation. Madison, WI: University
of Wisconsin Cooperative Extension.