Review: Introduction
Download
Report
Transcript Review: Introduction
Review: Introduction
Define Evaluation
How do formal/informal evaluation differ?
What are two uses of evaluation in education?
What are the pros/cons of using an external
evaluator?
Alternative Approaches to
Evaluation
Dr. Suzan Ayers
Western Michigan University
(courtesy of Dr. Mary Schutten)
Alternative Approaches
Stakeholders: individuals and groups who have a
direct interest in, and may be affected by,
evaluation; should be involved early, actively &
continuously
Program: activities that are provided on a
continuing basis; typically what is evaluated
There are a variety of alternative, often conflicting,
views of what evaluation is and how it should be
carried out
Why so many alternatives?
The way one views evaluation directly impacts
the type of activities/methods used
Origins of alternative models stem from
differences in:
Philosophical & ideological beliefs
Methodological preferences
Practical choices
Philosophical & Ideological Beliefs
Epistemologies (philosophies of knowing)
Objectivism (social science base of empiricism; replicate)
Subjectivism (experientially-based; tacit knowledge)
Pros/Cons
of each?
Principles for assigning value (parallel obj/subj)
Utilitarian: focus on group gains (avg scores); greatest
good for the greatest number
Intuitionist-pluralist: value is individually-determined
Room for both or are these dichotomous?
Philosophical purists are rare (impractical?)
Choose the methods right for THAT evaluation
Understand assumptions/limitations of different
approaches
Methodological Preferences
Quantitative (numerical)
Qualitative (non-numerical)
Evaluation is a transdiscipline; crosses paradigms
“Law of the instrument” fallacy
With hammer/nails, all appears to need hammering
Identify what is useful in each evaluation approach,
use it wisely & avoid being distracted by approaches
designed to deal w/ different needs
Practical Considerations
Evaluators disagree whether/not intent of evaluation is
to render a value judgment
Decision-makers or evaluator render judgment?
Evaluators differ in views of evaluation’s political role
Authority? Responsibility? These dictate eval style
Influence of evaluators’ prior experience
Who should conduct the evaluation and nature of
expertise needed to do so
Desirability (?) of having a wide variety of evaluation
approaches
Classification Schema for
Evaluation Approaches
Conceptual approaches to evaluation, NOT techniques
Objectives-oriented: focus on goals/objectives &
degree to which they are achieved
Management-oriented: identifying and meeting
informational needs of decision makers
Consumer-oriented: generate information to guide
product/service use by consumers
Expertise-oriented: use of professional expertise to
judge quality of evaluation object
Participant-oriented: stakeholders centrally involved in
process
See figure 3.1 (p. 68)
Objectives-oriented Approach
Purposes of some activity are specified and
then evaluation focuses on the extent to which
these purposes are achieved
Ralph W. Tyler popularized this approach in
education (criterion ref test)
Tylerian models
Metfessel & Michael’s paradigm (enlarged vision of
alternative instruments to collect evaluation data)
Provus’s Discrepancy Evaluation Model
(agree on stds,
det if discrepancy exists btwn perf/std, use discrepancy info to decide to
improve, maintain, terminate program)
Logic models
Determine long-term outcomes & backtrack to today
Objectives-oriented Steps
Establish broad goals or objectives tied to
mission statement
Classify the goals or objectives
Define objectives in behavioral terms
Find situations where achievement of objectives
can be shown
Select/develop measurement techniques
Collect performance data
Compare data with behaviorally stated
objectives
Objectives-oriented Pros/Cons
Strengths: simplicity, easy to understand, follow
and implement; produces information relevant
to the mission
Weakness: can lead to tunnel vision
Ignores outcomes not covered by objectives
Neglects the value of the objectives themselves
Neglects the context in which evaluation takes
place
Goal Free Evaluation
This is the opposite of objectives-oriented
evaluation, but the two supplement one another
Purposefully avoid awareness of goals; should not
be taken as given, goals should be evaluated
Predetermined goals not allowed to narrow focus of
evaluation study
Focus on actual outcomes rather than intended
Evaluator has limited contact with program
manager and staff
Increases likelihood of seeing unintended outcomes
Management-oriented Approach
Geared to serve decision makers
Identifies decisions administrator must make
Collects data re: +/- of each decision alternative
Success based on teamwork between evaluators
and decision makers
Systems approach to education in which
decisions are made about inputs, processes,
and outputs
Decision maker is always the audience to
whom evaluation is directed
CIPP Evaluation Model
(Stufflebeam)
Context Evaluation: planning decisions
Input Evaluation: structuring decisions
Available resources, alternative strategies?
Process Evaluation: implementing decisions
Needs to address? Existing programs?
How well is plan being implemented? Barriers to
success? Revisions needed?
Product Evaluation: recycling decisions
Results? Needs reduced? What to do after program
has ‘run its course’?
CIPP Steps
Focusing the Evaluation
Collection of Information
Organization of Information
Analysis of Information
Reporting of Information
Administration of Evaluation (timeline, staffing,
budget etc…)
Context Evaluation
Table 5.1
Objective: define institutional context, target
population and assess their needs
Method: system analysis, survey, hearings,
interviews, diagnostic tests, Delphi technique (experts)
For deciding upon the setting to be served, the
goals associated with meeting needs and
objectives for solving problems
Input Evaluation
Objective: identify and assess system capabilities,
procedural designs for implementing the strategies,
budgets, schedules
Method: inventory human and material resources,
feasibility, economics via literature review, visit
exemplary programs
For selecting sources of support, solution
strategies in order to structure change activities,
provide basis to judge implementation
Process Evaluation
Objective: identify or predict defects in the process or
procedural design, record/judge procedural events
Method: monitoring potential procedural barriers,
continual interaction with and observation of the
activities of the staff
For implementing and refining the program
design and procedure (a.k.a., process control)
Product Evaluation
Objective: collect descriptions and judgments of
outcomes and relate them to CIP, interpret worth/merit
Methods: measure outcomes, collect stakeholder
information, analyses of data
For deciding to continue, terminate, modify, or
refocus an activity and to document the effects
(whether intended or unintended)
Uses of Management-oriented
Approaches to Evaluation
CIPP has been used in school districts, state
and federal government agencies
Useful guide for program improvement
Accountability
Figure 5.1 (p. 94)
Formative and summative aspects of CIPP
Management-oriented Pros/Cons
Strengths: appealing to many who like rational,
orderly approaches, gives focus to the
evaluation, allows for formative and summative
evaluation
Weaknessws: preference given to top
management, can be costly and complex,
assumes important decisions can be identified
in advance of the evaluation
REVIEW/Qs
Why are there so many alternative approaches
to evaluation?
What two conceptual approaches to evaluation
did we discuss tonight? What are their +/-?
Which, if either, of these approaches do you
think will work for your evaluation object?
Identify your most likely evaluation object