Assessing Co-curricular Learning
Download
Report
Transcript Assessing Co-curricular Learning
Assessing Co-curricular
Learning
Robert Mundhenk
Visiting Scholar
The Higher Learning Commission
•Co-curricular Assessment workshop presented by Bob Mundhenk
Yellowjacket Union 203 While most institutions have devoted effort and re
Targeted Audience: Faculty and staff in non-academic units interested in assessment.
Ask Institutional Research about
graduation, retention, GPAs, and the like
Ask faculty about their teaching and the
learning it produces—but not necessarily
whether they know they’re producing
learning
Ask faculty how they know they’re
producing learning
Traditional Approaches to
Assessment of Learning
Traditional wisdom: Learning outcomes
need to be aligned at course, program,
and institutional levels
But where are many general education
goals, like “tolerance” and “teamwork”
and “the ability to function in an
increasingly diverse world” and
“inclination” taught and assessed?
Or is “taught” the right word???
The Institutional Mismatch
What happens if we substitute the word
“learned” for the word “taught”?
What are the implications of “Where are
learning outcomes learned and
assessed”?
Emphasis on student demonstration, not topiccovering
Ability to do or apply supersedes knowing
Responsibility for learning is shared
Site of learning becomes less specific, and
boundaries become more fungible
Shifting Perspectives
“Learning” is not exclusively classroombased
Many valued outcomes are not taught
exclusively in the classroom
Many valued outcomes are the result of
processes outside the classroom
“Learning” is a process based on three
interdependent student experiences:
Understanding academic content and processes
Student development
Identity formation
After Learning Reconsidered
Responsibility for “learning” exists
outside the classroom
Responsibility for “learning” doesn’t
always take the same form; some
entities on campus produce it, some
facilitate it, some support it
Responsibility for assessing learning
exists outside the classroom as well
After Learning Reconsidered
Civic Responsibility
AA: Service learning
SA: Student government, voter registration,
student judicial boards
Think and Engage as a Global Citizen
AA: Language courses, Anthropology, Sociology
SA: International experiences, culture days,
residence halls
Some Post-LR Examples of
Learning
Ability and inclination to:
◦ Think and make connections across disciplines
◦ Express oneself in multiple forms
◦ Analyze and reflect upon multiple perspectives
to arrive at a perspective of one’s own
◦ Think and engage as a global citizen
◦ Engage in evidence-based problem-solving
UW- Superior’s
Five Institutional Goals
How is ability made evident?
Where and how is ability developed?
Where is ability assessed?
What is “inclination”?
How is it made evident?
Where and how is it developed?
Where is it assessed?
Ability and Inclination
Campus Life learning outcomes:
◦ Recognizing and enhancing leadership potential
◦ Developing an appreciation of human
differences
◦ Seeking opportunities to engage in the campus
community and beyond
◦ Expanding desire for life-long learning
How Do These Outcomes Connect
with Undergraduate Learning
Outcomes?
Campus Life learning outcomes:
◦ Recognizing and enhancing leadership potential
◦ Developing an appreciation of human
differences
◦ Seeking opportunities to engage in the campus
community and beyond
◦ Expanding desire for life-long learning
What Ability or Inclination Do
They Develop?
Campus Life learning outcomes:
◦ Recognizing and enhancing leadership potential
◦ Developing an appreciation of human
differences
◦ Seeking opportunities to engage in the campus
community and beyond
◦ Expanding desire for life-long learning
How Do We Know They’re
Achieved?
Need to be intentional
Need to be planned
Need to be part of the structure of a
student’s experience
Need to be assessed
Co-curricular Outcomes
BEING INTENTIONAL
Efficiency models:
Focus on process
How well is this office/service functioning?
Focus on numbers:
Clients served
Graduation rates
Tutorial visits
Attendance at activities
Student/staff ratios
Traditional Co-curricular
Assessment
Effectiveness Models: Indirect
◦ Based on surveys and other indirect indicators,
like NSSE
◦ Often rely on student self-reporting
◦ Tend to skew positively on outcomes, if not
always on the processes that led to them
Newer Co-curricular Assessment
Effectiveness Models: Direct
◦ Focuses on student performance
◦ Can be based on observation or objective
measures
◦ Require carefully designed and consistent
measuring practices
Newer Co-curricular Assessment
Apply external standards, like CAS
Use surveys and questionnaires
Develop direct measurement strategies
All of the above
How to Assess Co-curricular
Learning
Standards for 40 functional areas
Thirteen component parts:
Mission
Program
Leadership
Organization and management
Human resources
Financial resources
Facilities, technology, and equipment
Legal responsibilities
Equity and access
Campus and community relations
Diversity
Ethics
Assessment and evaluation
CAS Standards
Knowledge acquisition, construction,
integration, and application
Cognitive complexity
Intrapersonal Development
Interpersonal competence
Humanitarianism/Civic Engagement
Practical Competence
CAS’s Six Outcome Domains
Can easily document the efficiency of
processes and organization
Can be used as well (through an
emphasis on the domains) to chart the
effectiveness of outcome development
efforts—depending on outcome definition
and quality of evidence
Using CAS Standards
Career Services: As a result of
interactions with the Career Services
Office, students and alumni will:
Identify their skills, abilities, and strengths in
order to make knowledgeable career decisions
Have the necessary resources and skills to
prepare for life-long post-graduate experiences
How are these outcomes connected to
institutional learning outcomes?
What are “resources and skills”?
Outcome Definition
College Unions: As a result of experiences in
the Yellowjacket Union, students will:
◦ Identify and utilize the opportunities and services
available to them
◦ Demonstrate a sense of ownership for the campus
community and civic involvement
◦ Interact with and value individuals from diverse
backgrounds and lifestyles
How do you know you’ve achieved the
second and third outcomes?
How do they overlap with institutional
outcomes?
Outcome Definition
Clearly defined outcomes help determine
the nature of evidence to be collected
Clearly defined outcomes focus on
student performance and development,
thus calling for both direct and indirect
forms of evidence
Outcome Definition and Evidence
DIRECT: uses performance or product,
created by students, that can be
compared to expected outcomes
--Capstone courses, projects,
internships, clinical experiences, etc.
INDIRECT: uses information that does not
directly link the learning to the outcomes
--graduation rates, grades, surveys,
“usage” rates
Direct and Indirect Measures
Is it relevant to the area’s stated mission
and function?
Does it measure what we want it to
measure?
Does it deal in some way with outcomes?
Is the information derived useful?
Can the information be used to improve
either function or learning?
Questions about Evidence
Did you accomplish what you hoped to
accomplish in your meeting with your
advisor?
How well did your experience at X prepare
you for employment?
As a result of this First-Year program, do
you feel better prepared for college?
Write a short essay in which you describe
the ways in which your attitudes and values
have changed as a result of your semester
in Argentina.
Surveys and Questionnaires:
Some Sample Questions
Traditional, indirect source of information
on effectiveness
Limitations:
Self-reporting
Unvalidated opinion
Response rates
Opportunistic data
Skewed samples
Surveys and Questionnaires:
Indirect Evidence
Kinds:
Satisfaction
Reflective
Post-experience experience (alumni and
employers)
Value:
True “customer” response
Can indicate areas for improvement and
ratification
Provides data for planning and review
Surveys and Questionnaires
Making them tools to assess learning:
Use learning outcomes as basis for at least some
questions
Validate by cross-referencing outcomes with
different populations (employers, alumni,
graduate, current students)
Emphasize the learning outcomes in design and
analysis of surveys and questionnaires
Surveys and Questionnaires
If Learning Reconsidered made the case
for cross-campus responsibility for
learning, then assessment of learning
outcomes is also a cross-campus
responsibility
Adaptation of practices and devices
already in use in academic settings
Standardized
Judgment-based
Direct Effectiveness Measures
Intentional Planning:
◦ Determine areas of responsibility: what
office/function might be a logical place to contribute
to particular learning outcomes?
◦ Plan the outcome-based purpose of the activity
◦ Aim at the appropriate level of Bloom’s taxonomy
◦ Design non-passive activities (watching a film plus
discussion; International Days as more than food,
costumes, and dance)
◦ Design outcome-focused opportunities for
processing
Developing Direct Measures of
Effectiveness
What is “inclination” and how is it
developed in co-curricular activities?
How does a student “express oneself in
multiple forms”?
What is a “global citizen”?
What does it mean to “make connections
across academic disciplines” and how
might co-curricular activities have a role
in developing this outcome?
Understanding the Outcomes
Choose one of the outcomes below and
determine how your co-curricular area
might have some responsibility for
developing it. Name specific activities that
might help develop the outcome and specify
what their effect on the student should be.
Think and make connections across academic disciplines
Express oneself in multiple forms
Analyze and reflect upon multiple perspectives to arrive at
a perspective of one’s own
Think and engage as a global citizen
Engage in evidence-based problem-solving
Being Intentional
Mapping:
◦ If the learning outcome is important, single
exposure isn’t enough
◦ How do first-year experiences differ from lastyear ones—or what difference is expected in
student response?
◦ How to assure student’s development of
outcomes from first year to last?
Planning for the Long Term
One-shot assessment produces
haphazard results that are usually
insufficient for planning improvement
Tie assessments to logical stages of
development, based on an outcome map
Be consistent in approach to assessing
Options:
Standardized instruments
Self-generated tools
Regular Assessment
Observations
Expert judgments
Student self-reflection
Employer/supervisor judgments
Self-generated Tools
Consistency across observers is crucial,
so a rubric of some kind is essential
Holistic rubrics: broad judgments
(Acceptable/Not Acceptable/Needs
Improvement)
Descriptive rubrics: defined criteria and
measures
Using Self-Generated Tools
First, determine the aspects of student
performance that would indicate he/she has
achieved an outcome (e.g., one aspect of a
“social justice” outcome might be “the student’s
writing demonstrates sensitivity to issues of
class and power”)
Second, define the specific things a student
would have to do to show he/she has mastered
that aspect (e.g., “Clear understanding of the
ways in which economic status affects
behavior.” (Criteria)
Finally, describe degrees of achievement for
each criterion (Measures)
Descriptive Rubrics
Using the outcome and the functional
area you chose earlier, develop a rubric
to measure student achievement,
defining one performative aspect of the
outcome, one criterion for measuring
that aspect, and a system (holistic,
descriptive, whatever) for describing
degrees of attainment
Developing a Rubric
Assessments done across campus should
ideally use the same rubrics or measures
When using the same tools is not
possible, it is essential that there be a
way to extract information that is usable
at the institutional level while still serving
the needs of the functional area
Institutional Assessment
To conclude:
◦ Understand the meaning of the desired
outcomes and your role in developing them
◦ Separate efficiency from effectiveness
◦ Plan experiences and assessments carefully
◦ Focus on using assessment information to
improve learning, not to justify your existence
◦ Collect information that is relevant, meaningful,
and useful
◦ Design systems that are reasonable and
manageable
ROBERT MUNDHENK
[email protected]
[email protected]