Review: Alternative Assessments I

Download Report

Transcript Review: Alternative Assessments I

Review: Alternative Assessments I

Describe the two epistemologies in ch. 3 (o/s)

Compare the two principles for assigning value
(util/int-pl)

Identify pros/cons of the two evaluation
approaches we discussed last week
Alternative Approaches to
Evaluation II
Dr. Suzan Ayers
Western Michigan University
(courtesy of Dr. Mary Schutten)
Consumer-Oriented Approach

Typically a summative evaluation approach

This approach advocates consumer education
and independent reviews of products

Scriven’s contributions based on groundswell of
federally funded educational programs in 1960s

Differentiation between formative/summative eval.
Consumer-Oriented Checklist
(Scriven, 1974, p. 102)



Need
Market
Performance









True field trials [tests in a “real” setting]
True consumer tests [tests with real users]
Critical comparisons [comparative data]
Long term [effects over the long term]
Side effects [unintended outcomes]
Process [product use fits its descriptions]
Causation [experimental study]
Statistical significance [supports product effectiveness]
Educational significance
Cost effectiveness
 Extended support [in service training]
Producer’s efforts to meet these standards
improve product effectiveness
 Key Evaluation Checklist developed to evaluate
program evaluations
 Educational Products Information Exchange
(EPIE): Independent product-reviewer service
 Curriculum Materials Analysis System (CMAS)
checklist: Describe product, analyze rationale,
consider: antecedent conditions, content,
instructional theory & teaching strategies, form
overall judgments

Uses of Consumer-Oriented
Evaluation Approach

Typically used by gov’t. agencies and consumer
advocates (i.e., EPIE)

What does one need to know about a product
before deciding whether to adopt or install it?
Process information
 Content information
 Transportability information
 Effectiveness information

Consumer-Oriented Pros/Cons

Strengths: valuable info given to those who
don’t have time to study, advance consumers’
knowledge of appropriate criteria for selection
of programs/products

Weaknesses: can increase product cost,
stringent testing may “crimp” creativity, local
initiative lessened b/c of dependency on outside
consumer services
Consumer-Oriented Qs

What educational products do you use?

How are purchasing decisions made?

What criteria seem to most important in the
selection process?

What other criteria for selection does this
approach suggest to you?
Expertise-Oriented Approach

Depends primarily upon professional expertise
to judge an institution, program, product, or
activity

This is the first view that relies heavily on
subjective expertise as the key evaluation tool

Examples: doctoral exams, board reviews,
accreditation, reappointment/tenure reviews
etc…
Expertise-Oriented Types

Formal Review Systems (accreditation)


Informal Review systems (grad S committee)


Existing structure, no standards, infrequent
schedule, experts, status usually affected
Ad hoc panel review (journal reviews)


Existing structure, standards exist, set review
schedule, experts, status usually affected by results
Multiple opinions, status sometimes affected
Ad hoc individual review (consultant)

Status sometimes affected
Expertise-Oriented Pros/Cons

Strengths: those well-versed make decisions,
standards are set, encourage improvement
through self-study

Weaknesses: whose standards? (personal
bias), expertise credentials, can this approach
be used with issues of classroom life, texts, and
other evaluation objects or only with the bigger
institutional questions?
Expertise-Oriented Qs

What outsiders review your program or
organization?

How expert are they in your program’s context,
process, and outcomes?

What are characteristics of the most/least
helpful reviewers? (list brainstorms on board)
Participant-Oriented Approach

Heretofore, the human element was missing
from program evaluation

This approach involves all relevant interests in
the evaluation

This approach encourages support for
representation of marginalized, oppressed
and/or powerless parties
Participant-Oriented Characteristics

Depend in inductive reasoning [observe, discover,
understand]

Use multiple data sources [subjective, objective,
quant, qual]

Do not follow a standard plan [process evolves as
participants gain experience in the activity]

Record multiple rather than single realities [e.g.,
focus groups]
Participant-Oriented Examples

Stake’s Countenance Framework


Description and judgment
Responsive Evaluation
Addressing stakeholders’ concerns/issues
 Case studies describe participants’ behaviors


Naturalistic Evaluation
Extensive observations, interviews, documents and
unobtrusive measures serve as both data and
reporting techniques
 Credibility vs. internal validity (x-checking, triangulation)
 Applicability vs. external validity (thick descriptions)
 Auditability vs. reliability (consistency of results)
 Confirmability vs. objectivity (neutrality of evaluation)


Participatory Evaluation


Utilization-Focused Evaluation


Collaboration between evaluators & key organizational personnel for practical problem solving
Base all decisions on how everything will affect use
Empowerment Evaluation
Advocates for societies’ disenfranchised, voiceless
minorities
 Advantages: training, facilitation, advocacy,
illumination, liberation
 Unclear how this approach is a unique participantoriented approach
 Argued in evaluation that it is not even ‘evaluation’

Participant-Oriented Pros/Cons

Strengths: emphasizes human element, gain
new insights and theories, flexibility, attention to
contextual variables, encourages multiple data
collection methods, provides rich, persuasive
information, establishes dialogue with and
empowers quiet, powerless stakeholders

Weaknesses: too complex for practitioners
(more for theorists), political element,
subjective, “loose” evaluations, labor intensive
which limits number of cases studied, cost,
potential for evaluators to lose objectivity
Participant-Oriented Qs

What current program are you involved in that
could benefit from this type of evaluation?

Who are the stakeholders?
Alternative Approaches Summary
Five cautions about collective evaluation
conceptions presented so far
1) Writings in evaluation are not models/theories

Evaluation is a transdiscipline (not yet a distinct discipline)

“Theoretical” underpinnings in evaluation lack
important characteristics of most theories

Information shared is: sets of categories, lists of
things to think about, descriptions, etc.
2) “Discipleship” to a single ‘model’ is dangerous

Use of different approaches as heuristic tools, each
appropriate for the situation, recommended
3) Calls to consolidate evaluation approaches into
a single model are unwise
These efforts based in attempts to simplify
evaluation
 Approaches are based on widely divergent
philosophical assumptions
 Development of a single omnibus model would
prematurely close a divergent phase in the field
 Just because we can does not mean we should;
would evaluation be enriched by synthesizing the
multitude of approaches into a few guidelines?

4) The choice of an evaluation approach is not
empirically based

Single most important impediment to development
of more adequate theory and models in evaluation
5) Negative metaphors underlying some
approaches can cause negative side effects

Metaphors shared in ch. 3 are predicated on
negative assumptions in two categories:
 Tacitly assume something is wrong in system
being evaluated (short-sighted indictment)
 Based on assumptions that people will lie, evade
Qs or withhold information as a matter of course
Alternative Approaches’ Contributions
Approaches shared in ch. 4-8 influence evaluation
practices in important ways
 Help evaluators think diversely
 Present & provoke new ideas/techniques
 Serve as mental checklists of things to consider,
remember, or worry about
 Alternative approaches’ heuristic value is very high,
but their prescriptive value is less so
 Avoid mixing evaluation’s philosophically incompatible
‘oil/water’ approaches; eclectic use of alternative
approaches can be advantageous to high-quality
evaluation practices
Table 9.1
Exercise

Clearly identify your evaluand
Is it a program, policy, product, service, other?
 Who does it (or should it) serve?
 Who is in charge of it?


Find a partner and explain what you have written
Does it make sense?
 Does it match what you wrote?
 Does it avoid specifying criteria?
 Is it simple enough?
 Did you avoid commenting on the merits of the
evaluand?
