Digital Library User Interface and Usability Week 9 Goals • Discover elements of good interface design for digital libraries of various sorts • Consider examples from.

Download Report

Transcript Digital Library User Interface and Usability Week 9 Goals • Discover elements of good interface design for digital libraries of various sorts • Consider examples from.

Digital Library User Interface and
Usability
Week 9
Goals
• Discover elements of good interface
design for digital libraries of various
sorts
• Consider examples from DL usability
evaluation as sources of insight.
• Look at the distinct requirements of
interfaces to libraries of video and audio
files
Caveat
• Note -– We have a whole course in User System
Interface
– Everything in that class is relevant to User
Interfaces for Digital Libraries
– One evening will not replace that course,
nor will it capture all of the relevant factors.
Note - to do later
• At the end of the class, I will ask you to
do a reflection on the points raised. You
will be asked to summarize the most
important characteristics of a welldeveloped DL interface.
• As you continue your DL projects, be
sure to apply the relevant components
of these elements.
The challenge
• A user interface for digital libraries must display
large volumes of data effectively.
– Typically the user is presented with one or more
overlapping windows that can be resized and
rearranged.
– In digital libraries, a large amount of data spread
through a number of resources necessitates intuitive
interfaces for users to query and retrieve information.
– The ability to change the user's perspective from highlevel summarization information down to a specific
paragraph of a document or scene from a film remains
a challenge to user interface researchers.
Source: http://cimic.rutgers.edu/ieee_dltf.html
Expectations of Digital Libraries
• Provide at least those services available
in traditional libraries
• … and more.
• A system is successful “only to the
degree to which the vast majority of its
intended users are able to use its
intended functionality”
Hill 97
User-centered design
• “User-centered design for a digital library must
include not only systems evaluation but also
an understanding of the process of
information seeking and use.”
• Compared to a “self-evident door handle” -once you see it, you know what it does and
how to use it. No instruction is necessary.
Hill 97
Methods of evaluation
• Surveys
• Target user groups
– Focus groups from the intended audiences
– another recommendation: faux focus groups
• When it is not practical to do a real focus group for a
while, the developers do some role playing, pretend to be
users
• What do you think of this approach?
• Ethnographic studies
– Audio/video taped sessions of users
– Analysis of feedback and comments
• Demographic analysis of beta tester
registration data
• Log analysis
Hill 97
Usability inspection of Digital
Libraries
• To produce a product with high usability
–
–
–
–
–
–
–
–
Client and user interviews
Task analysis
User class definitions
Usage scenarios
Iterative usability design
Prototyping
Unfortunately, developers often
Design walk-throughs
look at usability analysis as
Usability evaluation
something to do at the end of
the development process as a
final test, rather than as a part
of the design process.
Source: Hartson 04
Your plans
• How will you evaluate the usability of
your digital library?
– What is ideal?
– What is practical?
– What do you plan to do?
Evaluation
• Evaluation for any purpose has two major
components
– Formative
• During development, spot check how things are progressing
• Identify problems that may prevent goals from being
achieved
• Make adjustments to avoid the problems and get the
project back on track
– Summative
• After development, see how well it all came out
• Lessons learned may be applicable to future projects, but
are too late to affect the current one.
• Needed for reporting back to project sponsors on success of
the work.
Usability evaluation
• Lab-based formative evaluation
–
–
–
–
Real and representative users
Benchmark tasks
Qualitative and quantitative data
Leads to redesign where needed
• After deployment
– Real users doing real tasks in daily work
– Summative with respect to the deployed system
– Useful for later versions
Usability inspection
• Lower cost option than full lab-based testing
• Applies to early designs, well-developed
designs, and deployed systems
• Does not employ real users
• Expert based
– Usability engineering practitioners
• May be guided by typical user tasks
• Seeks to predict usability problems that users
will encounter.
Hartson 04
Inspection categories
• User classes
– Know your user
– Example from the cited study:
• Scientific researchers in computer science
• Administrators
– Do not use the regular interface, so not evaluated
• User tasks
–
–
–
–
–
Search for technical reports on a set of criteria
Browse the collection
Register
Submit
Harvest
Hartson 04
Search expanded
• Search options
– Simple search
• All bibliographic fields
• Group results by archive
• Sort
– Advanced search
• Focus on specific fields with filter options
Hartson 04
Results - 1
• Submit and Harvest tasks not evaluated
– Specialized domain requirements
– Need evaluation with real users to do meaningful
testing
• Report on Problems Found
– Usability problem types
• Wording, consistency
– Functionality
• Search and browse functionality
• Problem = anything that impacts the user’s
task performance or satisfaction.
Hartson 04
Categories of Problems
• General to most
applications, GUIs
– Wording
– Consistency
– Graphic layout and
organization
– User’s model of the
system
• Digital Library
functionality
–
–
–
–
Browsing
Filtering
Searching
Document submission
functions
Hartson 04
Wording
• About 36% of the problems in the case described in
the paper
• “Precise use of words in user interfaces is one of the
most important design considerations for usability”
• Clear, complete, correct
– Button and tab labels
– Menu choices
– Web links
• Crucial to help users learn and understand
functionality
• Easiest problems to fix if someone with right skills is
on the team.
Hartson 04
Search and Browse functionality
• Pretty basic to what a DL does!
• 18% of the problems were in that area.
• Designers consider these separate
functions
• Users see them as extremely closely
related
– Find the desired resource
– Should be designed together
Hartson 04
“Usual Suspects”
• Digital libraries prone to the same design faults as
other interactive systems
– Consistency
• In the example, “group” and “archive” were used
interchangeably
• Different labels for the same concept used in
different places
– Simple search on tab, Search all bibliographic field at
function location
• Multiple terms referring to the same concept
confuse users, slow learning
– Standardize terminology and check it carefully
Hartson 04
Usual Suspects - 2
• Problems with Feedback
– Clearly indicate where the user is in the
overall system
• Clicking a tab does not result in highlighting or
any kind of feedback about which tab is the
currently active choice.
• Selected institution (archive) highlighted when
chosen, but not maintained after some other
actions.
Hartson 04
Usual suspects - 3
• Wording
– Use of jargon or slang, or unclear or missing labels
• Challenge for users
• Example in NCSTRL
– Several dates used. The labels for the dates do not
clearly described what each represents.
– “discovery date” which is different from “accession date”
» Discovery date -- probably a developers term, and
not likely to be of interest to the user.
– Use terms that are meaningful to users without
explanation whenever possible. Resist presenting data
that is not useful for user purposes.
Hartson 04
Usual suspects - 4
• Wording, continued
– Example: “Submit to CoRR” tab
• Could be “Submit Technical Report(s) to CoRR
– Example: Search all bibliographic fields
• Could be “Simple Search: Search all bibliographic
fields in selected archive (or for selected
institution)”
– Other examples of unclear labels
• Archive’s Set - technical term from OAI-PMH
• DateStamp
• Discovery Date
• Label for the user, not the developer
Hartson 04
Usual Suspects - 5
• Incorrect or inappropriate wording
– “Search results” label for browsing results
– hits (1-n) or total xxx hits displayed
• Not search results, just reports available for
browsing
– Apparent use of combined code for browse
and search.
• Label results appropriately, even
scrupulously, for their real meaning.
Hartson 04
Usual suspects - 6
• Appropriate terms
– Use of “hits” for individual search (or
browse) results
• Commonly used
• Inappropriate slang, according to usability
experts
• Considered unattractive, even slightly offensive
• Recommended: something like “Matches with
search term”
– Cosmetic consideration can have a positive
affect on user’s impression of the site.
Hartson 04
Layout and design
• The whole point of a graphical user interface is
to convey more information to the user in a
short time.
– The GUI must support the user needs
• Example problems in the NCSTRL evaluation
– Menu choices - no logical order
– Reorganize by task or functionality
• Organize task interfaces by categories to
present a structured system model and reduce
cognitive workload.
Hartson 04
Layout example
• Instead of randomly ordered tabs, group them
by
– Information links
•
•
•
•
About NCSTRL
About CoRR
OAI
Help
– User tasks
•
•
•
•
•
Simple search
Advanced search
Browse
Register
Submit technical reports to CoRR
Hartson 04
Graphical design
• Proximity of elements suggests
associations and relatedness
– Search button very close to OR radio box
– Applies equally to all parts of the dialog
• Consider the implications of placement
and association of graphical elements.
Hartson 04
Start off right
• Any application should have a home
page that explains what the site is about
and gives the user a sense of the overall
site capability and use.
• NCSTRL starts with the Simple Search
page, with no introduction.
DL specific problems
• Searching, filtering, browsing
– User view: all are aspects of finding a needed
resource
– Developer view: differences based on what
must go into an index to support searching,
how filtering is combined with searching to
form a new query, etc.
• Usability suggestion: combine search,
browse, filter into one selection and
navigation facility.
– Give users the power to combine these
elements to serve their needs.
Hartson 04
Iterative search
• Search is often implemented as a one-shot
function.
– Users want to iterate on their query string to
improve results
• NCSTRL does not show the query that produced the
given results.
– Users want to prune the result set by applying
a subsequent query to just those results
• Not available in NCSTRL
• Examples where it is available?
Hartson 04
Browsing
• NCSTRL allows browsing only by institutions
(archive)
– Other possibilities
• Date
• Author
• Subject
• Allow user activity that will serve user needs.
Try to find out what users want before making
decisions about services offered.
Portal
• “A portal <is> a single point of access to
distributed systems that provides
services to support user needs to search,
browse, and contribute content, often
linking to shared existing functionality at
other sites.”
• Portal pass through problem
– Does the portal add service, or just provide
a link to a collection of other sites?
Hartson 04
Portal - submission
• NCSTRL - submission to CoRR
• Link opens to another page, not directly
offering the opportunity to submit.
• Disconnect for the user between the
original page and the action promised.
• Link directly to the service offered
without any intermediate pages unless
needed in support of the service.
Hartson 04
Summary for NCSTRL case
• System exhibited many typical problems
with user interfaces
• Investigation illuminated some issues
specific to digital libraries or other
systems for retrieving information.
Ensemble – an early page design Your initial thoughts?
First serious revision
Thoughts?
Mock up - 1
Mockup - 2
Mockup – 2a
Mockup - 3
Recommendations
• Based on our earlier review of usability
characteristics, what is your advice to
the Ensemble team?
Video Digital Libraries
• Video digital libraries offer more
challenges for interface design
– Information attributes are more complex
• Visual, audio, other media
– Indicators and controlling widgets
• Start, stop, reverse, jump to beginning/end,
seek a particular frame or a frame with a
specified characteristic
Source: Lee 02
Video Interface Features
• Cataloging
– Semi-automatic tool
– Manual tool
– Threshold adjustable
before automatic
segmentation
• Textual Query
– Natural language (or
keyword)
– Category or keyword list
browsing
– Audio information for
indexing, browsing
• Intelligent frame
selection
Source: Lee 02
• Video browsing
–
–
–
–
–
–
–
–
–
–
–
–
Text description
Transcript
Single keyframe
Storyboard
Option re granularity of keyframe
set
Interactive hierarchical keyframe
browser
Keyframe slide show
Video summary playing
Playback
Transcript + playback synch
Keyframe + playback synch
Text search + playback and/or
keyframe synch
Common features for Video DLs
• Most systems use a textual querying interface and few
systems provide any form of visual query interface,
probably indicating the need for further development in
this area;
• Most systems use keyframe(s) as their video browsing
method;
• Playback is provided in all listed systems, indicating that
playback is regarded as a most important interface feature;
• Whereas most systems provide more than one video
browsing method (often transcript + playback and/or
keyframe + playback), browsing aids such as
synchronization between different browsing methods are
not often facilitated.
Source: Lee 02
Stages of Information seeking in
Video Digital Libraries
• Browsing and then selecting video programs
(as a collection)
• Querying within a video program (content
querying)
• Browsing the content of a video program
• Watching (part of) a video program
• Re-querying the video digital library and/or
within a video program
Source: Lee 02
Summarizing
stages of
information
seeking and
the interface
elements that
support them
as described
in four
researchers’
work.
Source: Lee 02
Granularity in Video Browsing
• Abstraction
– Reducing the information available to a
manageable, usable subset
• Traditional video & audio browsing
– Sequential, single access point, linear nature of the
medium
– Constrained by time
– Fast forward
• Difficult to see the content
• Need to return to the beginning to repeat search
Source: Lee 02
A scenario
• Directory with 100 (or 1000 or…) video files.
• No information except the file name.
– Maybe reasonable name, but not very descriptive
• You want to find a particular clip from a party
or a ceremony or some other event.
• What are your options?
• What would you like to have available?
Spend a bit of time now talking about this.
Video Abstraction
• Levels to present: (from Shneiderman 98)
– Overview first
– Zoom and Filter
– Details on Demand
• Example levels (from Christel 97)
–
–
–
–
Title: text format, very high level overview
Poster frame: single frame taken from the video
Filmstrip: a set of frames taken from the video
Skim: multiple significant bits of video sequences
• Time reference
– Significant in video
– Options include simple timeline, text specification of time of the
current frame, depth of browsing unit
Source: Lee 02
Informedia-II Digital Video Library
• Carnegie Mellon project from the DLI program
• Capture audio and video from news programs
and documentaries
• Full-content search and retrieval
• Retrieval of video paragraphs based on spoken
or typed query.
• Words extracted from soundtrack, closedcaptioning or text on the screen.
Source: http://www.informedia.cs.cmu.edu/dli2/index.html
Informedia interface
Another view of query results from Informedia-II
What do we think of this interface? What is here?
How do we use it?
Keyframe browsing
• Extract a set of frames from the video
– Display each as a still image
– Link each to play the video from that point
• Selection is not random
– Video analysis allows recognition
• Sudden change of camera shot
• Scenes with motion or largely stationary
– Video indexing based on frame-by-frame image
comparison
• Similar to thumbnail browsing of image collections
Source: Lee 02
Keyframe
extraction
for display
on
browsing
interface
Source: Lee 02
Keyframe extraction
• Manual
– Owner or editor explicitly selects the frames to be
used as index elements
• Automatic
– Subsampling - select from regular intervals
• Easy, but may not be the best representation
– Automatic segmentation - break the video into
meaningful chunks and sample each
• Shot boundary detection - note switch from one camera
to another, or distinct events from one camera
Source: Lee 02
Displaying the frames
• Once the key frames are selected, display
them for effective user interaction
– Storyboard
• Miniaturized keyframes in chronological order
• Aka keyframe list or filmstrip
– Slide show
• Keyframes displayed one at a time
– Hierarchically arranged
• Good when content is structured
More detail
• For much more detail about Video
browsing and presentation, see Lee 02.
Summary
• Much of digital library user interface design and
usability analysis is the same as that of other web
services
–
–
–
–
Keep the user central in the design phase
Be careful about word use
Organize the graphics and layout carefully
Think about the user experience
• Some special considerations about DL usability
have to do with DL services
– Search, filter, browse
– Connections with other collections to which this is a
portal
References
Adam, N., Holowczak, R., Halem, M., Lal, N., and Yesha, Y. “Digital Lbrary
Technical Committee” cimic.rutgers.edu/ieee_dltf.html
Christel 97: Christel, M., Winkler, D. and Taylor, C. (1997) “Multimedia
abstractions for a digital video library” Proceedings of the 2nd ACM
International Conference on Digital Libraries (DL ‘97), Philadelphia, PA
July, pp 21-29
Hartson, H. R., Shivakmar, P, and Perez-Quiñones (2004) “Usability
inspection of digital libraries: a case study” International Journal of
Digital Libraries 4: 108-123
Hill, L., Dolin, R., Frew, J., Kemp, R., Larsgaard, M., Montello, D., Rae, M.,
and Simpson, J. “user Evaluation: Summary of the Methodologies and
Results for the Alexandria Digital Library, University of California at
Santa Barbara. www.asis.org/annual-97/alexia.htm
Lee 02: Lee, H., and Smeaton, A. (2002) “Designing the User Interface for
the Físchlár Digital Video Library” Journal of Digital Information, Volume
2, Issue 4 May 2002 http://jodi.tamu.edu/Articles/v02/i04/Lee/#2
Shneiderman 98: Shneiderman, B (1998) Designing the user interface:
strategies for effective human-computer interaction, 3rd edition
(Addison Wesley Longman)
Wactlar, Howard, “Informedia II Digital Video Library: Auto Summarization
and Visualization Across Multiple Video Documents and Libraries”
http://www.informedia.cs.cmu.edu/dli2/index.html