Karla A. Henderson, Ph.D. Professor, North Carolina State University [email protected] Framework Evaluation Inventory (handout) Needs Assessment Regarding Program Evaluation (Types of Needs Explained)
Download
Report
Transcript Karla A. Henderson, Ph.D. Professor, North Carolina State University [email protected] Framework Evaluation Inventory (handout) Needs Assessment Regarding Program Evaluation (Types of Needs Explained)
Karla A. Henderson, Ph.D.
Professor, North Carolina State University
[email protected]
Framework
Evaluation Inventory (handout)
Needs Assessment Regarding Program Evaluation
(Types of Needs Explained)
Small Group Needs Identification (6 people per group)
Importance (Very, Moderately, Slightly, Not)
5-10 topics to Discuss Together
What is Evaluation?
Evaluation= systematic collection and analysis of data to
address criteria to make judgments about the worth or
improvement of something; making decisions based on
identified criteria and supporting evidence
Assessment= the examination of some type of need that
provides a foundation for further planning
Evaluation is sometimes like solving a mystery
Types of ProgramEvaluation
Macro
Micro
(System
(Activity or
Evaluation)
Event
Evaluation)
Formal (Systematic)
Evaluation
MACRO and MICRO approaches----- Provides rigor
Systematic gathering (procedures and methods) of
evidence
Leads to decisions and action
Criteria
Evidence
Judgment
Assessments
Why assess?
eventually use the input to design programs and
objectives
generates new program ideas
gives constituents a say
helps you be responsive
What is a need? A want? An intention?
Assessments determine all three of these so YOU
can figure out how to promote what you are doing
Types of needs (Caution!)
Normative needs- what should be available
Felt needs- what individuals believe they
would like
Expressed needs- needs fulfilled through
participation
1. Why do Systematic
Evaluation
Potential Purposes:
Efficiency-How
are we doing?
(Management
/Process
Focused)
Effectiveness-
What difference
do our efforts
make?
(Impact/Outcome
Focused)
Accountability Era—DO YOU AGREE?
What gets measured gets done
If you don’t measure results, you
can’t tell success from failure
If you can’t see success, you can’t
reward it
If you can’t reward success, you’re
probably rewarding failure
Reinventing Government, Osborne and Gaebler, 1992
University of Wisconsin-Extension, Program Development and Evaluation
Accountability Era
If you can’t see success, you can’t
learn from it
If you can’t recognize failure, you
can’t correct it.
If you can demonstrate results, you
can win public support.
Reinventing Government, Osborne and Gaebler, 1992
University of Wisconsin-Extension, Program Development and Evaluation
Evaluation Process: “Begin
with the end in mind.”
Covey (1990), 7 Habits of Highly Effective People
2. How to Think Like
an Evaluator
A thinking process used by
evaluators:
(from Richard Krueger)
Reflecting
Develop a theory of action. A logical sequence that results in change.
Begin with what is supposed to happen--the results.
Listening
Share the theory of action with others.
Measuring
Determine your measurement strategies--how you're going to look at
the program
Adding value to the program
What can evaluation do to contribute to the program? How can
evaluation make the program better, more enjoyable, focused on
results, accountable, satisfying to participants and educators, etc?
Ways of Thinking:
Goal Based Thinkers - "We look for goals"
Audit Thinkers -"We investigate and find out what's
wrong"
Utilization Focused Thinkers - "We make evaluation
useful"
Empowerment Focused Thinkers - "We empower local
people"
Positivistic Thinkers - "We are scientists"
Number Thinkers - "We count--and we do it well“
Qualitative Thinkers - "We tell stories" (Richard Krueger)
Practical tips for successful
evaluation:
(Richard Krueger)
Involve others Utilization, impact and believability emerge from
involving colleagues and clientele. If you want the information used
then involve others!
Ask yourself: Do I have a program? and, Is it worthy of evaluation?
Consider your purpose for evaluating---- (see earlier slide)
Consider who wants the evaluation-Who requested it?
Use a variety of evaluation methods when possible.
Keep costs low by: Sampling strategically
Keep interest high by adding payoff to the participant.
Start with goals, but don't be unduly limited by goals
Consider "early evaluation"
Design the evaluation carefully. The evaluation should:
Enhance the program
Yield information beneficial to stakeholders
Conserve resources
3. Differences
Between
Assessment,
Evaluation, and
(Action) Research
Evaluation= systematic collection and analysis of data to
address criteria to make judgments about the worth or
improvement of something; making decisions based on
identified criteria and supporting evidence
Assessment= the examination of some type of need that
provides a foundation for further planning
Action Research-Evaluation leading to decisions/changes
4. Steps Involved in
Evaluation Process
Steps
Problem, Idea Identified
Problem Statement/Purpose Determined
Instrument/Method Chosen
Data Sources
Data Collection
Data Analysis
Conclusions/Recommendations
Evaluation Process: “Begin
with the end in mind.”
Covey (1990), 7 Habits of Highly Effective People
5. What Should be
Evaluated
Areas to Evaluate
Personnel
Places
Policies
Programs
Participant Outcomes
Potential Purposes:
Efficiency-How
are we doing?
(Mgmt/Process
Focused)
Effectiveness-
What difference
do our efforts
make?
(Impact/Outcome
Focused)
Levels of Evaluation:
END RESULTS (Impact)
PRACTICE CHANGE (Outcomes)
KASA CHANGE (Knowledge, attitudes, skills, and
aspirations)(Outcomes)
REACTIONS (SATISFACTION) (Outputs)
PEOPLE INVOLVEMENT (Outputs)
ACTIVITIES (Outputs)
INPUTS--RESOURCES
What is sampling?
A population is the theoretically specified aggregation
of study elements.
A sample represents or is representative of a
population
Types of Sampling
Probability
Non-Probability
Theoretical
Probability
Probability sampling
Samples are selected in accord with probability theory, typically
involving some random selection mechanism.
Random
Stratified Random
Systematic
Cluster
Representativeness
Quality of a sample having the same distribution of characteristics
as the population from which it was selected.
Nonprobability
Technique in which samples are selected in a way that
is not suggested by probability theory.
Purposive
Convenience
Quota
Expert
Snowball
6. Who Should
Conduct Evaluations
WHO DOES THE EVALUATIONS?
Internal
You!
Staff
Agency Evaluation Personnel
External
Consultants
University Students!
Regardless—YOU have to know your purpose, goals, and
appropriate methods!
7. When Should
Evaluataions Be
Done
Timing
Assessments (planning) – find out
where to begin based on what you
know
Formative (process)- concerned with
efficiency and effectiveness
Summative (product) – overall
performance
Approaches to Needs
Assessments
Literature/Professional Development
Advisory Groups
Structured Interviews (individual and
focus groups)
Surveys
Formative Evaluation
Evaluation in Process
Allows Changes to be Made
Immediately
Most often Focused on Inputs and
Outputs
Summative Evaluation
At the END of something
“What was?”
Recommendations for the Future
8. What Tools are
Available for
Evaluations
Data Collection Methods
MAJOR ONES:
Questionnaires/Surveys
Interviews (Individual and Focus Groups)
(Pros and Cons)
Other Methods:
Systematic Observation
Checklists
Field Observations
Unobtrusive Measures
Physical Evidence
Archives
Covert Observations
Visual Analyses
Experimental Designs
Case Studies
9. What “Cutting
Edge” Strategies
Exist for Evaluation
To consider (see below for further info)
Logic Models—”outcome focused”
Trends Analysis
Benchmarking
Proragis (NRPA)
10. CAPRA Standards
related to Evaluation
CAPRA Standards
10.1 Systematic Evaluation Program
There shall be a systematic evaluation plan to
assess outcomes and the operational
deficiency and effectiveness of the agency.
10.2 Demonstration Projects and Action Research
There shall be at least one experimental or
demonstration project or involvement in
some aspect of research, as related to any part
of parks and recreation operations, each year.
CAPRA Standards
10.3 Evaluation Personnel
There shall be personnel either on staff or a
consultant with expertise to direct the
technical evaluation/research process.
10.4 Employee Education
There shall be an in-service education program
for professional employees to enable them to
carry out quality evaluations.
11. Evaluating Inputs,
Outputs, Outcomes,
and Impacts
Evaluation Approaches—KEY POINTS!
Multiple LEVELS of evaluations:
Inputs (costs, personnel, etc.)
Outputs
Activities
People involvement
Reactions
Outcomes
KASA-knowledge, attitudes, skills, aspirations
Behavior CHANGE
Impacts
Long-term Benefits
12. Using Goals and
Objectives as Basis
for Evaluation
What are the goals of the
program?
What do we expect to
happen?
What do we want
participants to do, gain,
learn?
BEGIN WITH THE END IN MIND!!
Goals and Objectives
Goals
Broad, long-range statements that define the
programs/services that are going to be provided
Objectives
Specific statements (about the attainable parts of the
goal) that are measurable and have some dimension of
time.
Objectives
Specific
Measurable
Must be attainable and reality-based!!!
Relevant
Must be some way to determine whether or not the desired
results have been achieved
Achievable
Must be clear and concrete
Must be useful; must have worth to your organization
Time-limited/Time connected
Must specify a time frame for accomplishment
Adapted from Edginton, Hanson, & Edginton, 1980
13. Using Logic
Models as Basis for
Evaluation
Logic Model (Preview)
Goal
StrategiesActivities
Short-term
Outcomes
Long-term
Outcomes
Data
Sources &
Performance
Measures
Moving Toward Success: Framework for After-School Programs, 2005
A logic model is…
A HOT TOPIC in (Program) Evaluation
A depiction of a program showing what the
program will do and what it is to
accomplish.
A series of “if-then” relationships that, if
implemented as intended, lead to the
desired outcomes
The core of program planning and
evaluation
University of Wisconsin-Extension, Program Development and
Evaluation
Simplest Form of a Logic Model
INPUTS
OUTPUTS
OUTCOMES
54
Everyday example
H
E
A
D
A
C
H
E
Situation
Get pills
Take pills
Feel better
INPUTS
OUTPUTS
OUTCOMES
55
Where are you going?
How will you get there?
What will show that you’ve
arrived?
“If you don’t know
where you are going,
how are you gonna’
know when you get
there?”
Yogi Berra
56
How will activities lead to desired outcomes?
A series of if-then relationships
Tutoring Program Example
IF
We
invest
time and
money
then
IF
We can
provide
afterschool
programs to
50 children
then
IF
Students
will have
opportuniti
es for
active
recreation
then
IF
They will
learn and
improve
their
recreation
skills
then
IF
They will
be more
active
University of Wisconsin-Extension, Program Development and Evaluation
then
They will
become fit
and
develop
life long
skills
57
Logical chain of connections
showing what A program is to
accomplish
INPUTS
OUTPUTS
Program
investments
Activities
Participation
What we
invest
What
we do
Who we
reach
OUTCOMES
Short
Medium
Longterm
What results
58
OUTPUTS
What we do
ACTIVITIES
•Train, teach
•Deliver services
•Develop products and
resources
•Network with others
•Build partnerships
•Assess
•Facilitate
•Work with the media
•…
Who we reach
PARTICIPATION
•Participants
•Clients
•Customers
•Agencies
•Decision makers
•Policy makers
University of Wisconsin-Extension, Program
Development and Evaluation
Satisfaction
59
OUTCOMES
What results for individuals, families, communities..…
SHORT
MEDIUM
LONG-TERM
Learning
Action
Conditions
Changes in
Changes in
Changes in
• Awareness
• Knowledge
• Attitudes
• Skills
• Opinion
• Aspirations
• Motivation
• Behavioral intent
•Behavior
•Decision-making
•Policies
•Social action
Conditions
Social (well-being)
Health
Economic
Civic
Environmental
CHAIN
OF
OUTCOMES
Limitations
Logic Model…
Doesn’t address:
Are we doing the right thing?
University of Wisconsin-Extension, Program Development and Evaluation
14. Using
Benchmarking as a
Basis for Evaluation
Benchmarking
Agencies or Services compared to other agencies or
services to determine a relative value
Set goals against a “standard” of the best
Benchmarks are the statistics; Benchmarking is the
process
Information examples: budgets, salaries, fees charged,
number of participants, number of staff
Decide who you want to compare, design a
questionnaire, analyze, and write report
15. Using NRPA’s
PRORAGIS Tool for
Evaluation
16. Doing
Cost/Benefit and
Economic Impacts
Economic Impact Analyses
Determining the dollar value a
program or service has on the
economy
The net change of money in a
community from the spending of
visitors—only visitors are part of
the analyses
Cost-Benefit Analyses
Focus on Monetary Forms
Sometimes Difficult to Do
Social Benefits may outweigh
Monetary Benefits
17. Using Trends
Analysis for
Evaluation
Trends analysis
Examining the future by looking at
the past
Measuring data over time
Secret is good record keeping
18. Designing Valid,
Reliable, and
Useable
Questionnaires
Kinds of Questions
Open-ended
Close-ended
Fixed alternatives
Likert Scale
Semantic differential
Ranking
Partially close-ended
Examples
Open-ended: “What did you like best about day
camp?”
Close-ended, fixed alternatives: “What did you like
best about day camp this summer?”
Counselors
Arts
Snacks
Swimming
Sports
Another Example
Close-ended, Likert: “What did you think about each
of these parts of day camp?”
Counselors
Arts & Crafts
Snacks
Swimming
Poor Fair Good Great
1
2
3
4
1
2
3
4
1
2
3
4
1
2
3
4
Wording Advise
One idea per question
Clear, brief, and simple
Avoid leading questions
Avoid estimates if possible
Use words familiar to respondent
Avoid fancy words and jargon
Be clear about meanings of words
Avoid negative questions
Do a pilot study
State alternatives precisely-mutually exclusive
responses
Use stages if needed but give GOOD directions
Format and Layout Design
For Good Response
Give clear directions
Start with something easy & familiar (NOT demos though)
Have white space & easy font to read
Colored paper if easy to read—font size appropriate for audience
Have easy to follow directions for staged questions
Have “professional look”
Front page should include title, date or time or year, perhaps a graphic
Keep the length manageable for the info desired
Anchor all numbered responses—e.g., 5=strongly agree, 4=agree etc.
NO TYPOGRAPHICAL ERRORS
PILOT TESTING
Is the survey valid?
How much time does it take to complete?
Do the respondents feel comfortable
answering the questions?
Is the wording of the survey clear?
Are the answer choices compatible with
the respondents’ experience in the
matter?
19. Using Individual
Interviews
Types of Interviewing
Personal (in-depth)
Telephone
Part of Field Observation
Focus Groups (Group Interviews)
Interviews
Approaches:
Structured, close-ended
Standardized, open-ended
Interview guide (semi-structured)
Informal conversational (unstructured)
Asking Open-Ended Questions
NEVER allow a person to say “yes” or “no”
Purpose is to get people to talk
Start with noncontroversial questions
Demographics at end/don’t ask if you know (e.g., gender)
Avoid “WHY”
Use probes and follow-ups (“Tell me more about…?”
Use listening skills
Maintain control of the interview
Share yourself with the interviewee
Structuring Interviews
Purpose of interviews
Individual or group
If group- number of persons involved
Structured or unstructured?
Audio-taping/Note-taking
Facilitator/Recorder Roles
Rapport-building
Timing
Things to remember
Pay attention to the content of your questions
(keep your focus)
Give thought to setting up the interview
Pay attention to how you plan to record the
data
Take notes-on-notes
Conduct interviewer training
Go over possible problem areas
Training for Interviewers
Discussion of general guidelines and procedures.
Specify how to handle difficult or confusing situations.
Conduct demonstration interviews.
Conduct “real” interviews.
Guidelines for Structured Survey
Interviewing
Dress in a similar manner to the people who will be
interviewed.
Study and become familiar with the questions.
Follow question wording exactly (if it is quantitative).
Record responses exactly.
Probe for responses when necessary.
Telephone Surveys
Advantages:
Money and time.
Control over data collection.
Disadvantages:
Surveys that are really ad campaigns.
Answering machines.
20. Conducting
Focus Groups
Focus Group
A group of people are brought together in a room to
engage in guided discussion of a topic.
Pros
and cons
Pros
Socially oriented
Flexible
High face validity
Speedy
Low in cost?
Cons
Researcher has less control
Data may be difficult to
analyze
Need good moderator
Must do more than one
group-great differences
between groups
Groups difficult to assemble
Must have conducive
environment
What about group interviews?
Focus groups encourage ideas by interacting
with other individuals
Pay attention to the same issues in a personal
interview
Have a clear purpose for the group meeting
(criteria)
Have clear structured, open-ended questions
Points for Conducting a
Focus Group
Be a good moderator
Set ground rules
The beginning of the focus group process is crucial
to its success
Keep the process focused and flowing
Probe where needed
Be sure to thank participants
Other Focus Group Tips
Usually 5-10 individuals
Key is to get people to talk to each other
Usually two moderators (one to direct discussion and
another to take care of technical aspects)
Use incentives to get people to particiapte (maybe)
Pilot test just like ALL surveys
USE OPEN-ENDED questions
Be familiar with group process—getting people to talk,not
letting some take over
Ground rules-speaking one at a time, confidentiality,
breaks taken, positive and negative
Summary
Focus groups encourage ideas by interacting
with other individuals
Pay attention to the same issues in a personal
interview
Have a clear purpose for the group meeting
(criteria)
Have clear structured, open-ended questions
21. Using
Observational Tools
Types of Data
Qualitative
Quantitative
Qualitative Observations
Takes a great deal of field time
Need in-depth notes on your observations
Use of key informants is often helpful
Criteria often change and become re-defined
Tips on (Field) Observations
Only unscientific if inappropriate techniques are used
Highly reliable and valid IF systematically done
Time-consuming
Anecdotal records or critical incidents describe FACTS of situations
Field notes are collected-MANY, MANY, MANY
Researcher is data collection instrument
SEPARATE facts from interpretations
Notetaking
Record as soon as possible
Take notes in detail-running descriptions, conversations, incidents
Make copies of notes ASAP
Indicate whose language-yours, quotes, paraphrases
Rule of thumb-several (5-6 single spaced typed pages for an hour of
observation)
Can use tape-recorder and transcription
Record ALSO what you do not understand
MONITOR yourself as a data collector
Length of time spent depends on research question and theoretical
saturation
“IF it is not written down, it never happened!”
Dimensions of Approaches
Full participant observer-> Outside observer
Overt observer-> Covert observer
Single observation-> Long-term multiple obs.
Narrow focus-> Broad focus
Quantitative: Duration Sampling Checklist
Participant: _________
Activity: _____________ Time: _____ to
________
Behavior
Duration
Total
On task (appropriate)
2 sec., 3 sec., 19 sec
24 sec
Approp. Conversation
(w/ peers)
Walks around room
Argues w/ leader
Argues w/ peers
14 sec., 29 sec
112, 143 sec, 157 sec
0 sec
43 sec
43 sec
412 sec
0 sec
43 sec
Frequency Sampling Checklist
Client: _________
Activity: _____________ Time: _____ to ________
Behavior
Frequency of Behavior
Initiates a conversation
Shifts attention to new project
Asked question by peer
//
/
////
Responds appropriately when
asked question by peer
///
Rating Scales
Behavior
Excellent
Average
Poor*
How well does the
client respond to questions?
How appropriate does client
interact with peers?
*Excellent= top third Average= middle third
Poor= lowest third
22. Using On-line
Tools
(SurveyMonkey etc.)
Question Structures-SurveyMonkey
Question 1: Have you used a swimming pool in the last year?
Yes
No
Question 2: How many times in the last year have you gone to a public swimming pool?
Never
Once
Twice
More than 3 times
Question 3: How many times have you gone to a public swimming pool this past year? (Please
write in an estimate number) __________
On-Line Considerations--Pros
Same strengths as a paper version
•Better at addressing sensitive issues
•Cost efficient
•Faster delivery
•Design options
•Dynamic
•Ability to track
•Quick response time
•Easier to use for skip logic
On-Line Considerations--Cons
Spam/Privacy concerns
•Technical issues
•Submitting multiple submissions
•No interviewer present to clarify questions or issues
New Technologies and Survey
Research
CAPI - computer assisted personal interviewing.
CASI - computer assisted self interviewing.
CSAQ - computerized self-administered
questionnaires.
TDE - touchtone data entry.
VR - voice recognition.
23. “Fun” Strategies
for Doing
Evaluations
Importance-Performance Eval
How important is something?
How well did the organization perform it?
Matrix of:
Keep up the good work
Possible Overkill
Needs Work
Low Priority
Other Ideas:
Use Grading Scale A, B , C, D, F
Use Pie Chart and distribute money for preferences
Use Happy Faces
Computer Assisted
Others…
Measuring Physical Activity
SOPLAY or SOPARC
Pedometers or
Accelerometers
24. Selecting
Evaluation
Instruments
Selecting an Evaluation/Measurement Tool
Is it reliable and valid?
Does it measure what you want?
Appropriate for the participants?
Reasonable to administer and in your price range?
Directions clear, concise, unambiguous?
Easy to analyze?
Is it the best way to measure the objectives?
Is the activity reaction form put together by Cary, NC
the one you should use?
25. Contracting with
Outside Consultants
Outside:
You have to know what you want done
You have to be able to evaluate results and make
decisions
You probably need some amount of financial resources
Could save you a lot of time if you do it right
26. Developing Your
Own Evaluation Tool
Steps:
Define problem/level of evaluation
Determine contents (including criteria) and broad
questions
Identify and categorize respondents
Develop items, structure format
Write directions
Ensure response
Purpose of Data Collection
What do you want to know?
Who has the information?
What is the best approach (based on purpose,
time, money, resources, your expertise) to use?
How will you use the results?
Are you interested in outputs and/or outcomes?
Kinds of Information Sought
Behavior info
Knowledge info
Attitudes/beliefs/values info
Demographic info
**pay attention to relationship of respondent to the
question (their past, present, future)
27. Ethical
Considerations
related to Evaluation
Issues
Ethical
Political
Legal
Moral
Evaluator must be Competent
Knowledge about area to be researched
Knows about evaluation design
Knows wide range of methods
Knows how to analyze and interpret data and
relate them back to conclusions/recommendations
Developing Competencies – con’t
Knows how to use the results
Understands how to handle political, legal, and ethical
concerns encountered
Must have certain personal qualities: trustworthy,
strives for improvement, responsive to sensitive issues
“Doing the Right Thing”
Political Issues (science is touched by politics but goes on
anyway; social change is always political; values matter)
Supports/refutes views and values
Personal contacts
Value-laden definitions
Controversial findings
Pressures to produce certain findings
Know the organization’s position, don’t go beyond the data
in conclusions, have clear purpose for research
“Doing the Right Thing”
Legal Issues
Not many legal concerns except around illegal behaviors
Moral
Unintentional mistakes made by bias or mistake
Cultural & procedural biases
Letting bias, prejudice, friendships influence outcomes
Dealing with negative findings
Taking too long to get the results out
Be prepared to recognize the possibility of statistical errors and know how
to explain them
“Doing the Right Thing”
Voluntary Participation
No Harm
Anonymity and Confidentiality
Issues of Deception
Analysis and Reporting
Ethics Principles
Be open with people
Don’t promise more than you can do
Protect the rights and privacy of your subjects
Guard against coercion
Get written consent & Board approval
Guard against harm
Let participants know the results
Anonymity
The researcher cannot identify a given response with
a given respondent.
Confidentiality
Researcher can identify a given person's responses but
promises not to do so publicly.
Ethical Considerations
Avoid bias (settings, questions, populations)
Consider question content
Protect confidentiality (if appropriate)
Manage data/records
Report results of evaluation
Respect/report different opinions
28. Devising an
Evaluation Plan for
an Agency
Purposeful Planning for Evaluation
Schedule time on your yearly calendar
Involve planning Board, committees,
colleagues, staff, volunteers, etc.
Identify what you hope to achieve
(desired outcomes)
Identify your goals and objectives
Purposeful Planning for Evaluation
Identify the methods you will use to
measure whether or not your program
met objectives
Identify how often evaluation will be
undertaken
Identify WHO will do the evaluation
and how they will get the appropriate
resources
Formal (Systematic)
Evaluation--Review
MACRO and MICRO approaches----- Provides rigor
Systematic gathering (procedures and methods) of
evidence
Leads to decisions and action
Criteria
Evidence
Judgment
29. How to Use
Information from
Evaluation Projects
Using Data-KEY ADMONISHMENTS
If you collect the data, you need to analyze
the data!
Record the results of your data collection
Compare your findings to your original goals
and objectives
Share with appropriate people
Incorporate into your planning cycle