MSSE Course Name - Georgia Institute of Technology
Download
Report
Transcript MSSE Course Name - Georgia Institute of Technology
Saturday March 5
Cognitive Walkthroughs Due Today
User Testing
Other Evaluation Mechanisms
Looking Ahead
1
SEng 5115
March 5, 2005
Evaluation with Users
Big investment -- big potential
Many issues
dealing with human subjects
which users? which tasks?
when in the process?
what to measure?
how to measure?
2
SEng 5115
March 5, 2005
Dealing with Human Subjects
Your responsibility to protect subjects
distress, embarrassment
remind them that you are not testing them
Informed, voluntary consent
understand that they can quit at any time
explain test in lay terms
if necessary, there is “equipment failure”
Privacy: anonymity, use of image/voice
3
SEng 5115
March 5, 2005
Which Users? Which Tasks
As close to real users as possible
if real users are scarce, try surrogates
Keep close to the real tasks
may need to shorten some for time
reasons
may need to provide users with
background information
4
SEng 5115
March 5, 2005
When in the Process?
Early is important
low investment
time to change
Mock-ups and Drawings are OK
issues in how to handle user choice
Partial prototypes when necessary
Summary: as early as possible
5
SEng 5115
March 5, 2005
What to Measure
Process data
problems, questions, reactions
what users are thinking
Bottom-line data
mostly later for usability measurement
not very useful early in design
Asking users questions?
6
problematic -- users will answer
SEng 5115
March 5, 2005
Thinking-Aloud Method
User asked to think-aloud
ask questions (though not answered)
explain decisions, identify confusion
Tester records session
avoids interfering as much as possible
• only when test would end otherwise
7
explain to subject that you won’t
answer
SEng 5115
March 5, 2005
Eye-Tracking Testing
Technology to support monitoring
where subjects are looking and for
how long
Challenge: easy to direct results
avoid thinking out loud
careful presentation of tasks
careful design to avoid distractions
8
SEng 5115
March 5, 2005
Alternative Methods
Natural testing conditions
gather performance data
video-prompted review
Two-person tests
Field studies instead of user tests
9
forces “thinking” aloud
consider deployment, logging
SEng 5115
March 5, 2005
Gathering Field Data
How to get it?
Log high-level actions
Log low-level actions
Log problems
Work products
What to get?
Detailed and statistical usage data
Example cases
10
SEng 5115
March 5, 2005
Examples
Consider a Word Processor
many alternative solutions for
commands
• toolbars, menus, keyboard shortcuts
relative frequencies of commands
co-occurrence of commands with
undo
document statistics
11
SEng 5115
March 5, 2005
Examples
Consider a Website
maps of link traversal rates
• traffic maps
hidden co-occurrence
• web usage mining
errors
apparent rates of “back” from
destinations
time on page
12
SEng 5115
March 5, 2005
Different Goals,
Different Approaches
Overall “what’s happening”
general data
possibly lots of data
Test specific questions
13
targeted data
Always consider issues of user
consent
SEng 5115
March 5, 2005
General Guidelines for
User Testing
Plan ahead of time
what data to record
what instructions to deliver
what to do if user “falls off prototype”
when to provide help, and what help
Know your objectives
14
but never lose sight of the user
SEng 5115
March 5, 2005
General Guidelines
Always have a pilot study
Get professional help for big studies
In general, it is better if you aren’t
there
too much bias
subtle clues
stay behind one-way glass
15
SEng 5115
March 5, 2005
When You Need
Bottom-Line Data
Need specific measurements
median time for task
comparison of alternatives
Work out the statistics involved
Focus on one type of test at a time
16
statistics cookbooks
can’t time and use think-aloud
SEng 5115
March 5, 2005
Putting it Together
Critique this test plan
background
• outdoor information Kiosk for 2012
Olympics in New York (we can hope!)
• used by English-speaking Americans
• will provide information on events and
schedules, locations, transportation,
results, maps, and other useful
information
17
SEng 5115
March 5, 2005
Kiosk Test Plan
use prototype kiosk
recruit test users: offer $10 for up to an hour
• post at New York employment and social service agencies
users shown kiosk and 20 minute demonstration video
each user has 3 tasks and the system times them on each
task
finally, groups of 5-10 users will be asked
• which tasks they could not accomplish
• what problems they had with the system
18
SEng 5115
March 5, 2005
Usability Goals and
Measures
Concrete, quantitative measures of usability
learning time
use time for specific tasks and users
error rates
measures of user satisfaction
Comparative usability goals
compare with prior versions or competitors
19
SEng 5115
March 5, 2005
Things to Watch
Goals should be realistic
Many goals go beyond the application UI
100% is never realistic
training, manuals
Testing goals should help improve the UI
detail--not just good/bad
20
SEng 5115
March 5, 2005
Exercise: Setting Usability
Goals
As groups, come up with three usability
goals for your project.
try to come up with markedly different goals
to give broader coverage
discuss the feasibility of testing these goals
• what is needed for the test
• when in the process?
• how much effort, user preparation/training, etc.?
21
SEng 5115
March 5, 2005
Interface Evaluation
Goals of interface evaluation
find problems
find opportunity for improvement
determine if interface is “good
enough”
22
SEng 5115
March 5, 2005
With or Without Users
Users are expensive and inconsistent
usability studies require several users
some users provide great information,
others little
Users are users
23
cannot be simulated perfectly
Best choice--Both
SEng 5115
March 5, 2005
Evaluation Without Users
Quantitative Methods
GOMS/keystroke analysis
back-of-the-envelope action analysis
Qualitative Methods
expert evaluation
cognitive walkthrough
heuristic evaluation
24
SEng 5115
March 5, 2005
GOMS/Keystroke Analysis
Formal action analysis
accurately predict task completion
time for skilled users
Break task into tiny steps
keystroke, mouse movement, refocus
gaze
retrieve item from long-term memory
25
Look up average step times
tables from large experiments
SEng 5115
March 5, 2005
GOMS/Keystroke Analysis
Primary utility: repetitive tasks
e.g., telephone operators
benefit: can be very accurate (within
20%)
may identify bottlenecks
Difficulties
challenging to decompose accurately
long/laborious process
not useful with non-experts
26
SEng 5115
March 5, 2005
Back-of-the-Envelope Action
Analysis
Coarse-grain
list basic actions (select menu item)
each action is at least 2-3 seconds
what must be learned/remembered?
what can be done easily?
documentation/training?
Goal is to find major problems
27
Example: 1950’s 35mm camera
SEng 5115
March 5, 2005
Expert Evaluation
Usability specialists are very valuable
double-specialists are even better
An inexpensive way to get a lot of
feedback
Be sure the expert is qualified in your
area
28
SEng 5115
March 5, 2005
Looking Ahead
Next week: Usability Laboratory
Location: Walter Library, room B-26
No food in the lab
Focus: user testing examples
29
external “real” client
SEng 5115
March 5, 2005
Looking Ahead
Heuristic Evaluations
individual evaluations first
• each group member should do an individual
evaluation
• you may use either the “old” or “new” heuristics
from the notes
• turn in the individual lists of problems identified
combined results
• as a group, create a merged list of issues
• turn in that list as the group work product
30
Then, revised prototypes (and user testing
plans) for March 26th
SEng 5115
March 5, 2005