Presentation on theme: "Digital Library User Interface and Usability Week 8."— Presentation transcript:
Digital Library User Interface and Usability Week 8
Goals Discover elements of good interface design for digital libraries of various sorts Consider examples from DL usability evaluation as sources of insight. Look at the distinct requirements of interfaces to libraries of video and audio files
Caveat Note -- –We have a whole class in User System Interface –Everything in that class is relevant to User Interfaces for Digital Libraries –One evening will not replace that course, nor will it capture all of the relevant factors.
Note - to do later At the end of the class, I will ask you to do a reflection on the points raised. You will be asked to summarize the most important characteristics of a well- developed DL interface. As you continue your DL projects, be sure to apply the relevant components of these elements.
The challenge A user interface for digital libraries must display large volumes of data effectively. –Typically the user is presented with one or more overlapping windows that can be resized and rearranged. –In digital libraries, a large amount of data spread through a number of resources necessitates intuitive interfaces for users to query and retrieve information. –The ability to smoothly change the user's perspective from high-level summarization information down to a specific paragraph of a document or scene from a film remains a challenge to user interface researchers. Source: http://cimic.rutgers.edu/ieee_dltf.htmlhttp://cimic.rutgers.edu/ieee_dltf.html
Expectations of Digital Libraries Provide at least those services available in traditional libraries … and more. A system is successful “only to the degree to which the vast majority of its intended users are able to use its intended functionality” Hill 97
User-centered design “User-centered design for a digital library must include not only systems evaluation but also an understanding of the process of information seeking and use.” Compared to a “self-evident door handle” -- once you see it, you know what it does and how to use it. No instruction is necessary. Hill 97
Methods of evaluation Surveys Target user groups –Focus groups from the intended audiences Ethnographic studies –Audio/video taped sessions of users –Analysis of feedback and comments Demographic analysis of beta tester registration data Log analysis –We will consider in more detail next week as we look at quality measures Hill 97
Usability inspection of Digital Libraries To produce a product with high usability –Client and user interviews –Task analysis –User class definitions –Usage scenarios –Iterative usability design –Prototyping –Design walk-throughs –Usability evaluation Source: Hartson 04 Unfortunately, most developers look at usability analysis as something to do at the end of the development process as a final test, rather than as a part of the design process.
Evaluation Evaulation for any purpose has two major components –Formative During development, spot check how things are progressing Identify problems that may prevent goals from being achieved Make adjustments to avoid the problems and get the project back on track –Summative After development, see how well it all came out Lessons learned may be applicable to future projects, but are too late to affect the current one. Needed for reporting back to project sponsors on success of the work.
Usability evaluation Lab-based formative evaluation –Real and representative users –Benchmark tasks –Qualitative and quantitative data –Leads to redesign where needed After deployment –Real users doing real tasks in daily work –Summative with respect to the deployed system –Useful for later versions
Usability inspection Lower cost option than full lab-based testing Applies to early designs, well-developed designs, and deployed systems Does not employ real users Expert based –Usability engineering practitioners May be guided by typical user tasks Seeks to predict usability problems that users will encounter. Hartson 04
Inspection categories User classes –Know your user –Example from the cited study: Scientific researchers in computer science Administrators –Do not use the regular interface, so not evaluated User tasks –Search for technical reports on a set of criteria –Browse the collection –Register –Submit –Harvest Hartson 04
Search expanded Search options –Simple search All bibliographic fields Group results by archive Sort –Advanced search Focus on specific fields with filter options Hartson 04
Results - 1 Submit and Harvest tasks not evaluated –Specialized domain requirements –Need evaluation with real users to do meaningful testing Report on Problems Found –Usability problem types Wording, consistency –Functionality Search and browse functionality Problem = anything that impacts the user’s task performance or satisfaction. Hartson 04
Categories of Problems General to most applications, GUIs –Wording –Consistency –Graphic layout and organization –User’s model of the system Digital Library functionality –Browsing –Filtering –Searching –Document submission functions Hartson 04
Wording About 36% of the problems in the case “Precise use of words in user interfaces is one of the most important design considerations for usability” Clear, complete, correct –Button and tab labels –Menu choices –Web links Crucial to help users learn and understand functionality Easiest problems to fix if someone with right skills is on the team. Hartson 04
Search and Browse functionality Pretty basic to what a DL does! 18% of the problems were in that area. Designers consider these separate functions Users see them as extremely closely related –Find the desired resource –Should be designed together Hartson 04
“Usual Suspects” Digital libraries prone to the same design faults as other interactive systems –Consistency “group” and “archive” used interchangeably Different labels for the same concept used in different places –Simple search on tab, Search all bibliographic field at function location Multiple terms referring to the same concept confuse users, slow learning –Standardize terminology and check it carefully Hartson 04
Usual Suspects - 2 Problems with Feedback –Clearly indicate where the user is in the overall system Clicking a tab does not result in highlighting or any kind of feedback about which tab is the currently active choice. Selected institution (archive) highlighted when chosen, but not maintained after some other actions. Hartson 04
Usual suspects - 3 Wording –Use of jargon or slang, or unclear or missing labels Challenge for users Example in NCSTRL –Several dates used. The labels for the dates do not clearly described what each represents. –“discovery date” which is different from “accession date” »Discovery date -- probably a developers term, and not likely to be of interest to the user. –Use terms that are meaningful to users without explanation whenever possible. Resist presenting data that is not useful for user purposes. Hartson 04
Usual suspects - 4 Wording, continued –Example: Submit to CoRR tab Could be “Submit Technical Report(s) to CoRR –Example: Search all bibliographic fields Could be “Simple Search: Search all bibliograhic fields in selected archive (or for selected institution” –Other examples of unclear labels Archive’s Set - technical term from OAI-PMH DateStamp Discovery Date Label for the user, not the developer Hartson 04
Usual Suspects - 5 Incorrect or inappropriate wording –“Search results” label for browsing results –hits (1-n) or total xxx hits displayed Not search results, just reports available for browsing –Apparent use of combined code for browse and search. Label results appropriately, even scrupulously, for their real meaning. Hartson 04
Usual suspects - 6 Appropriate terms –Use of “hits” for individual search (or browse) results Commonly used Inappropriate slang, according to usability experts Considered unattractive, even slightly offensive Recommended: something like “Matches with search term” –Cosmetic consideration can have a positive affect on user’s impression of the site. Hartson 04
Layout and design The whole point of a graphical user interface is to convey more information to the user in a short time. –The GUI must support the user needs Example problems in the NCSTRL evaluation –Menu choices - no logical order –Reorganize by task or functionality Organize task interfaces by categories to present a structured system model and reduce cognitive workload. Hartson 04
Layout example Instead of randomly ordered tabs, group them by –Information links About NCSTRL About CoRR OAI Help –User tasks Simple search Advanced search Browse Register Submit technical reports to CoRR Hartson 04
Graphical design Proximity of elements suggests associations and relatedness –Search button very close to OR radio box –Applies equally to all parts of the dialog Consider the implications of placement and association of graphical elements. Hartson 04
Start off right Any application should have a home page that explains what the site is about and gives the user a sense of the overall site capability and use. NCSTRL starts with the Simple Search page, with no introduction.
DL specific problems Searching, filtering, browsing –User view: all are aspects of finding a needed resource –Developer view: differences based on what must go in an index to support searching, how filtering is combined with searching to form a new query, etc. Usability suggestion: combine search, browse, filter into one selection and navigation facility. –Give users the power to combine these elements to serve their needs. Hartson 04
Iterative search Search is often implemented as a one-shot function. –Users want to iterate on their query string to improve results NCSTRL does not show the query that produced the given results. –Users want to prune the result set by applying a subsequent query to just those results Not available in NCSTRL Examples where it is available? Hartson 04
Browsing NCSTRL allows browsing only by institutions (archive) –Other possibilities Date Author Subject Allow user activity that will serve user needs. Try to find out what users want before making decisions about services offered.
Portal “A portal a single point of access to distributed systems that provides services to support user needs to search, browse, and contribute content, often linking to shared existing functionality at other sites.” Portal pass through problem –Does the portal add service, or just provide a link to a collection of other sites? Hartson 04
Portal - submission NCSTRL - submission to CoRR Link opens to another page, not directly offering the opportunity to submit. Disconnect for the user between the original page and the action promised. Link directly to the service offered without any intermediate pages unless needed in support of the service. Hartson 04
Summary for NCSTRL case System exhibited many typical problems with user interfaces Investigation illuminated some issues specific to digital libraries or other systems for retrieving information.
Another Case - CITIDEL CITIDEL Databases -- Resource metadata -- Users and Accounts MySQL Database Perl Page Generatio n Modules Apache HTTP server Custom HTTP Handler Dispatch routines Open Archives Initiative (OAI) Interface CITIDEL XSL Templates HTML pages End Users -- Educators -- Students -- Researchers -- Professionals Other Digital Libraries and metadata indices INTERNETINTERNET Custom Search Engine (ESSEX)
CITIDEL Practical issue –What would be the results of applying a usability review to CITIDEL, similar to that applied to NCSTRL? A few extra notes that come up in examining CITIDEL –No way to submit a resource (only accepts metadata for resources located elsewhere) Is that an issue? Why or why not? –Design of the front page Cluttered, confusing What is really essential? What is useful? How should it be organized?
Video Digital Libraries Video digital libraries offer more challenges for interface design –Information attributes are more complex Visual, audio, other media –Indicators and controlling widgets Start, stop, reverse, jump to beginning/end, seek a particular frame or a frame with a specified characteristic Source: Lee 02
Video Interface Features Cataloging –Semi-automatic tool –Manual tool –Threshold adjustable before automatic segmentation Textual Query –Natural language (or keyword) –Category or keyword list browsing –Audio information for indexing, browsing Intelligent frame selection Video browsing –Text description –Transcript –Single keyframe –Storyboard –Option re granularity of keyframe set –Interactive hierarchical keyframe browser –Keyframe slide show –Video summary playing –Playback –Transcript + playback synch –Keyframe + playback synch –Text search + playback and/or keyframe synch Source: Lee 02
Common features for Video DLs Most systems use a textual querying interface and few systems provide any form of visual query interface, probably indicating the need for further development in this area; Most systems use keyframe(s) as their video browsing method; Playback is provided in all listed systems, indicating that playback is regarded as a most important interface feature; Whereas most systems provide more than one video browsing method (often transcript + playback and/or keyframe + playback), browsing aids such as synchronisation between different browsing methods are not often facilitated. Source: Lee 02
Stages of Information seeking in Video Digital Libraries Browsing and then selecting video programs (as a collection) Querying within a video program (content querying) Browsing the content of a video program Watching (part of) a video program Re-querying the video digital library and/or within a video program Source: Lee 02
Summarizing stages of information seeking and the interface elements that support them as described in four researchers’ work. Source: Lee 02
Granularity in Video Browsing Abstraction –Reducing the information available to a manageable, usable subset Traditional video & audio browsing –One point of access –Sequential –Fast forward Difficult to see the content Need to return to the beginning to repeat search Source: Lee 02
Video Abstraction Levels to present: (from Shneiderman 98) –Overview first –Zoom and Filter –Details on Demand Example levels (from Christel 97) –Title: text format, very high level overview –Poster frame: single frame taken from the video –Filmstrip: a set of frames taken from the video –Skim: multiple significant bits of video sequences Time reference –Significant in video –Options include simple timeline, text specification of time of the current frame, depth of browsing unit Source: Lee 02
Keyframe browsing Extract a set of frames from the video –Display each as a still image –Link each to play the video from that point Selection is not random –Video analysis allows recognition Sudden change of camera shot Scenes with motion or largely stationary –Video indexing based on frame-by-frame image comparison Similar to thumbnail browsing of image collections Source: Lee 02
Keyframe extraction for display on browsing interface
Keyframe extraction Manual –Owner or editor explicitly selects the frames to be used as index elements Automatic –Subsampling - select from regular intervals Easy, but may not be the best representation –Automatic segmentation - break the video into meaningful chunks and sample each Shot boundary detection - note switch from one camera to another, or distinct events from one camera Source: Lee 02
Displaying the frames Once the key frames are selected, display them for effective user interaction –Storyboard Miniaturized keyframes in chronological order Aka keyframe list or filmstrip –Slide show Keyframes displayed one at a time –Hierarchically arranged Good when content is structured
More detail For much more detail about Video browsing and presentation, see Lee 02.
Summary Much of digital library user interface design and usability analysis is the same as that of other web services –Keep the user central in the design phase –Be careful about word use –Organize the graphics and layout carefully –Think about the user experience Some special considerations about DL usability have to do with DL services –Search, filter, browse –Connections with other collections to which this is a portal
References Adam, N., Holowczak, R., Halem, M., Lal, N., and Yesha, Y. “Digital Lbrary Technical Committee” cimic.rutgers.edu/ieee_dltf.html Christel 97: Christel, M., Winkler, D. and Taylor, C. (1997) “Multimedia abstractions for a digital video library” Proceedings of the 2nd ACM International Conference on Digital Libraries (DL ‘97), Philadelphia, PA July, pp 21-29 Hartson, H. R., Shivakmar, P, and Perez-Quiñones (2004) “Usability inspection of digital libraries: a case study” International Joural of Digital Libraries 4: 108- 123 Hill, L., Dolin, R., Frew, J., Kemp, R., Larsgaard, M., Montello, D., Rae, M., and Simpson, J. “user Evaluation: Summary of the Metholologies and Results for the Alexandria Digital Library, University of California at Santa Barbara. www.asis.org/annual-97/alexia.htm Lee 02: Lee, H., and Smeaton, A. (2002) “Designing the User Interface for the Físchlár Digital Video Library” Journal of Digital Information, Volume 2, Issue 4 May 2002 http://jodi.tamu.edu/Articles/v02/i04/Lee/#2 Shneiderman 98: Shneiderman, B (1998) Designing the user interface: strategies for effective human-computer interaction, 3rd edition (Addison Wesley Longman)