Assessment 101 Joyce Chapman Project Manager, Triangle Research Libraries Network Heads of Cataloging Interest Group ALA annual, Anaheim CA 25 June 2012
Download
Report
Transcript Assessment 101 Joyce Chapman Project Manager, Triangle Research Libraries Network Heads of Cataloging Interest Group ALA annual, Anaheim CA 25 June 2012
Assessment 101
Joyce Chapman
Project Manager, Triangle Research Libraries Network
Heads of Cataloging Interest Group
ALA annual, Anaheim CA
25 June 2012
WHAT IS ASSESSMENT?
Assessment is a continuous and cyclical process
by which we evaluate and improve services,
products, workflows, and learning.
A continuous process
"Assessment is a process whose power is
cumulative.” –University of Washington’s assessment principles
• “One-shot" assessments can be useful in the
right context
• Ongoing assessment is what fosters ongoing
improvement
Planning phase
Amazing how often people decide to skip this.
Planning is one of the most difficult phases.
• Determine your objectives
• Define the questions that need to be
answered
• Design a method to answer the questions (set
up a study, collect new data, extract existing
data from a system)
Implementation (data gathering)
• NOT numbers for the sake of numbers
• We frequently measure things that are easy to
measure, without a good reason for doing so.
This data may not help us answer meaningful
questions.
• For data collection to foster assessment, we
must first determine what it is we really care
about, then initiate data collection that will
inform meaningful analysis and outcomes.
Assessment is a continuous and cyclical process
by which we evaluate and improve services,
products, workflows, and learning.
React / refine
• The most frequent piece of the assessment
cycle that is ignored is the last: making change
based on the findings of data analysis.
• It is often inaction on the part of management
that causes the assessment loop to remain
incomplete, ending with reporting of data
analysis findings and never resulting in action.
Summary
• Continuous
• Cyclical
• Evaluate AND improve
Requires that action be taken in response to
findings
• Our environment and users are always
changing; we are always reacting.
Evidence based practice
A movement to encourage and give practitioners
the means to incorporate research into their
practice where it may have been lacking.
– Journal of Evidence Based Library and Information Practice
• Assumption that it is impossible to make good
evidence-based decisions when our evidence base is
weak; therefore, an evidence base must be built.
Evidence based practice
Stresses three aspects contributing to a practice
that is evidence-based:
1. the best available evidence is used
2. moderated by user needs and preference
3. applied to improve the quality of professional
judgments
“An approach to information science that
promotes the collection, interpretation, and
integration of valid, important and applicable
user-reported, librarian-observed, and researchderived evidence. The best available evidence,
moderated by user needs and preferences, is
applied to improve the quality of professional
judgments.” – Anne McKibbon
Context of assessment
• Libraries often talk about assessment in the
specific context of proving our institutional
value to external audiences
– Contribution to student retention, graduation, and
employment rates; student learning outcomes
• Equally valuable is assessment to improve
internal workflows and services; or
assessment of the cost and value of workflows
to contribute to knowledge in the field.
Why perform assessment?
1. Improve efficiency
2. Modify workflows to funnel staff time and
efforts where they provide the most benefit
3. Prove the value of existing or proposed
services/positions to higher administration
4. Contribute to the available data/literature in
the cataloging field so that you can work
together to implement evidence-based
practice across the nation
A CULTURE OF ASSESSMENT
What is a culture of assessment?
A culture of assessment refers to whether the
predominating attitudes and behaviors that
characterize the functioning of an institution
support assessment.
What signals a culture of assessment?
•
•
•
•
Do staff take ownership of assessment efforts?
Does administration encourage assessment?
Is there comprehensive program?
Is assessment present, do we see ongoing
assessment efforts throughout the organization?
• Are there efforts to teach staff about
assessment?
• Is there inclusion of assessment in plans and
budget?
“Establishing a Culture of Assessment” by Wendy Weiner, 2009
What signals a culture of assessment?
• Is assessment mentioned in the strategic plan?
Does the organization have an assessment plan?
• Does the organization financially support staff
members whose positions are dedicated in whole
or in part to assessment-related activities?
• Does the organization fund professional
development of staff related to assessment?
• Is the organization responsive to proposals for
new endeavors related to assessment?
“Establishing a Culture of Assessment” by Wendy Weiner, 2009
Structural difficulties
• Bottom-up: it is difficult for staff to gain
support for conducting assessment projects or
for implementing change based on findings
when upper admin is not assessment-focused.
• Top-down: an assessment-focused upper
admin can have a staff that does not support
assessment and mandates for assessment are
resented (the “defiant compliant” culture).
CONTEXT OF ASSESSMENT IN
HIGHER EDUCATION
Inputs, outputs, and outcomes
• Input measures: quantify a library's raw materials
(collection size, staff size, budget). Longest
tradition of measurement in libraries.
• Output measures: measure the actual use of
library collections and services (circulation stats,
gate counts, reference transactions)
• Outcome measures: measure the impact that
using library services, collections, and space has
on users (libraries' impact on student learning)
What drives the “assessment agenda”?
• Changing times
– Explosive growth in technologies
– Increased customer expectations for services such
as quality and responsiveness
• Shrinking budgets
– justifications for spending $ on resources,
programs, and services are now required
– Increased competition for resources
– A fight to remain relevant and prove value
Martha Kyrillidou, “Planning for Results: Making the Data Work For You.” 2008.
Why wasn’t there a focus on assessment in
libraries for so long?
Return On Investment
“A performance measure used to evaluate the
efficiency of an investment…. a way of considering
profits in relation to capital invested.”
“ROI provides a snapshot of profitability adjusted for
the size of the investment assets tied up in the
enterprise.”
Sources: Wikipedia and Investopedia
PROFITS!
Cost = $$
Value = $$
For-profits must create their own revenue.
Otherwise, they cease to exist.
Livelihood depends on cost/value assessment.
Libraries
Like most higher education, we are non-profits.
Funding sources aren’t directly tied to real value
(more “perceived value” and tradition).
Cost = ??
Value = ??
Put simply…
because higher education has historically been
given a large % of annual funding by external
powers based on perceived value, we have not
developed a culture of needing to closely prove
value, track input to output, or investment to
profit.
ASSESSING THE COST AND VALUE
OF BIBLIOGRAPHIC CONTROL
2011 LRTS article by Stalberg & Cronin
• June 2009, Heads of Technical Services in
Large Research Libraries interest Group of
ALCTS sponsored a Task Force on the
Cost/Value of Bibliographic Control.
• Members: Ann-Marie Breaux, John Chapman,
Karen Coyle, Myung-Ja Han, Jennifer O’Brien
Roper, Steven Shadle, Roberta Winjum, Chris
Cronin, and Erin Stalberg
• The task group found that the technical services
community has long struggled with making
sound, evidence-based decisions about
bibliographic control
• If technical services is to attempt to perform
cost/value assessment on bibliographic control,
one of our first problems is a lack of operational
definitions of value we must create our own
operational definitions of value with which to
work.
Fundamental questions for defining value
1. Can value be measured in ways that are
non-numeric?
2. Is discussing relative value over intrinsic
value helpful?
3. Does value equal use?
4. Is it possible to define a list of bibliographic
elements that are “high-value” and others
that are “low-value”?
While the charge was to develop measures for
value, the Task Force determined that doing so
would not be helpful until the community has a
common vocabulary for what constitutes value
and an understanding of how value is attained,
and until more user research into which
bibliographic elements result in true research
impact is conducted.
Operational definitions of value
1.
2.
3.
4.
Discovery success
Use
Display understanding
Ability of bibliographic data to operate on the
open web and interoperate with vendors and
suppliers in the bibliographic supply chain
Operational definitions of value
5. Ability to support the Functional
Requirements of Bibliographic Records
(FRBR) user tasks
6. Throughput and timeliness
7. Ability to support the library’s administrative
and management goals
“Value multipliers”
Extent to which bibliographic data:
•
•
•
•
•
are normalized
support collocation and disambiguation in discovery
use controlled terms across format and subject domains
level of granularity matches what users expect
enable a formal and functional expression of relationships
(links between resources) to find “like” items
• are accurate
• enhancements are able to proliferate to derivative records
Measuring cost
While elements contributing to the cost can be outlined,
determining whether the costs are too high is impossible
without first having a clear understanding of “value.”
•
•
•
•
Salaries & benefits multiplied by time spent on a task
Cost of cataloging tools, such as software
Time spent on database maintenance
Overhead (training, policy development,
documentation)
• Opportunity costs
COLLECTING DATA
Types of data
Quantitative methods focus on numbers and
frequencies; provide data that is easy to analyze
statistically. “Numbers.”
Analysis of log data, systems reports, time data,
web usage analytics, survey data (not free text)
Qualitative methods capture descriptive data
and focus on experience and meaning. “Words.”
Usability testing, focus groups, user interviews,
ethnographic studies
Coding qualitative data
"There's no such thing as qualitative data.
Everything is either 1 or 0.”
- Fred Kerlinger
• While qualitative data provides the important
whys and hows of user behavior, it is difficult for
us to digest large quantities of descriptive data.
• It is often useful to code quantitative data
qualitatively for analysis.
• Fear: assessment takes a lot of time
• Reality: it depends on the methodology and data
sources used
• Qualitative data gathering, coding, and analysis
usually take a lot of time
• Systems can be set up to gather quantitative data
programmatically. Such data can be analyzed
quickly, given the proper tools and skills
• Quantitative data might also be gathered
manually. Data collection will be a hassle, but
analysis will be quick
Existing data or new data?
digital exhaust data
Collect new data
• Know what questions the data needs to be
able to answer
• Data requirements; structure of the data
• Make sure you will be able to extract the data
• Make sure the data format you’ve chosen will
be interoperable with any other data you are
using in an initiative
Bad data planning
METHODOLOGIES / TECHNIQUES
A/B Testing
• Common in web-based marketing research.
Involves an online performance comparison
between a control group and a single variable
test group.
• Measures differences in web usage stats
• Could test use differences based on: presence
or absence of metadata, order of metadata
display, metadata display labels
Warning
• If more than one variable is at play during A/B
testing, it becomes difficult to know which
variables was responsible for performance
differences achieved.
• As of summer 2012, Google Analytics includes
a new A/B Testing feature.
Focus groups
• A form of qualitative research in which a group of
people are asked about their perceptions and
opinions of, or interaction with, a product or
service.
• Questions are asked in an interactive group
setting where participants are free to talk with
other group members.
• A moderator leads a small group of people who
share a common experience or characteristic
through a discussion using a pre-prepared script
of open-ended questions.
Usability testing
• Focuses on measuring a product's ability to
meet its intended purpose by gathering direct
input about how real users use the system.
• In contract to a focus group or interview,
usability testing captures a users' behavior
whether they are aware of it or not.
• Code as you go! Software such as Morae
includes special features to help with this.
Web usage analytics data
What kinds of things you can track?
• Clicks on any link (Event Tracking)
• The pages people came from
• The pages people go to next on your site
• The pages people exit your site from
Important to evaluating “value”
• A/B testing
• Focus groups
• Usability studies
• Web usage analytics data
Important to evaluating “cost”
• Time studies
Watch out for bias
• Biased goal: “To prove the value of X”
• Unbiased goal: “to prove or disprove the
value of X”
• It can be considered a serious ethical conflict
to have the people who benefit from a certain
outcome of assessment be the same people
who conduct the assessment.
Institutional Review Board (IRB)
• Charged with protecting the rights of human
research subjects
• Mandated by federal law
• Library projects are the least of their worries,
but you may need to go through the process
• Check with your Assessment Librarian or
Assessment committee; they may have
guidelines for when to go through review
Institutional Review Board (IRB)
Guidelines provided by IRB rep to UNC-CH:
Must do IRB if you are using human subjects and
plan to…
• Publish results
• Make generalizable claims
• Collect identifying or sensitive information
(SSN, sexual orientation, names)
Institutional Review Board (IRB)
• All investigators listed on IRB must do a CITI
online ethics training (~3 hours) the first time
they submit an IRB proposal ( no need to
list every single person involved in the study!)
• Most library research is eligible for expedited
approval
• Tip for quick process: list the IRB approval
number for similar library studies already
approved by the IRB.
THE ROLE OF MANAGEMENT
Earning buy-in
• Discuss/explain assessment and why it’s
important
• Considering bringing a speaker to a
department meeting to talk about assessment
practices, tools, ideas, etc.
• Show examples from other institutions about
how assessment benefited a department
• Explain that you aren’t assessing staff; you’re
assessing workflows
Dealing with fear of task timing
• It’s scary to be asked to time yourself as you
conduct your daily tasks!
• Assumption that data collection is based on a
desire to decrease the time spent on tasks, or
penalize those who take “too long”
• Explain the big picture and what you’re trying
to achieve; consider whether anonymous data
could be just as useful to you
Prove that it makes a difference
• Make sure you don’t forget the 4th step of
assessment: take ACTION based on findings
• Inform staff of how their work informed your
decision-making and helped the management
of the department: show concrete changes
• Praise staff participating in assessment librarywide; publicize the success of the assessment
project within the library or department
Provide staff with training
• It’s important to provide staff with the training
they need to do what you’re asking
• Does required reporting involve querying an
Access database? Crunching data in Excel
(writing functions or making pivot tables)? Are
you asking them to create a survey?
• These are often not listed in required job
skills, but people are later tasked with data
reporting
Thank you!
• Joyce Chapman
• [email protected]
“Digital exhaust” base matrix image credit:
http://alwaysoncommunications.com/22/advanced-online-audience-buying/
Resources
• Kyrillidou, Martha. (Feb. 2008). FLICC meeting, “Planning for Results: Making the
Data Work For You.” “Why Assess? What is assessment? What do we mean
by actionable data?” The Cato Institute, Washington, D.C.
• McKinsey Global Institute report (May 2011), “Big data: the next frontier for
innovation, competition, and productivity.”
• Stalberg, Erin & Christopher Cronin. (2011). "Assessing the Cost and Value of
Bibliographic Control.” Library Resources & Technical Services, 55(3) 124-137.
• Weiner, Wendy. (2009). “Establishing a Culture of Assessment.” Academe 95(4).
SPEAKING OF ASSESSMENT…
Please fill out the evaluation survey for the
ALCTS events that you attended at ALA
annual conference 2012 in Anaheim, CA:
http://www.surveymonkey.com/s/alctsevents2012?
GROUP DISCUSSION
Discussion questions
1. What are the kinds of things we always want
to know about ourselves (TS) but never
invest the time in assessing?
2. What are your ideas about / experiences
with how to create a culture of assessment?
3. What important assessment work do we
need to do as a community? Can we initiate
any of that collaborative work now?