1 SIMS 247: Information Visualization and Presentation Marti Hearst Nov 30, 2005.

Slides:



Advertisements
Similar presentations
Agile Usability Testing Methods
Advertisements

Web E’s goal is for you to understand how to create an initial interaction design and how to evaluate that design by studying a sample. Web F’s goal is.
HCI 특론 (2007 Fall) User Testing. 2 Hall of Fame or Hall of Shame? frys.com.
The Scientific Method.
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
USABILITY AND EVALUATION Motivations and Methods.
Semester in review. The Final May 7, 6:30pm – 9:45 pm Closed book, ONE PAGE OF NOTES Cumulative Similar format to midterm (probably about 25% longer)
Group Project. Don’t make me think Steve Krug (2006)
HCI Methods for Pathway Visualization Tools Purvi Saraiya, Chris North, Karen Duca* Virginia Tech Dept. of Computer Science, Center for Human-Computer.
1 User Testing. 2 Hall of Fame or Hall of Shame? frys.com.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
User Interface Testing. Hall of Fame or Hall of Shame?  java.sun.com.
Energy Information Administration Usability Testing Colleen Blessing.
1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.
ICS 463, Intro to Human Computer Interaction Design: 8. Evaluation and Data Dan Suthers.
HCI revision lecture. Main points Understanding Applying knowledge Knowing key points Knowing relationship between things If you’ve done the group project.
Chapter 14: Usability testing and field studies
SBD: Usability Evaluation
류 현 정류 현 정 Human Computer Interaction Introducing evaluation.
Conducting Usability Tests ITSW 1410 Presentation Media Software Instructor: Glenda H. Easter.
Evaluation: Controlled Experiments Chris North cs3724: HCI.
Predictive Evaluation
Copyright © 2010 Pearson Education, Inc. Chapter 13 Experiments and Observational Studies.
Experiments and Observational Studies. Observational Studies In an observational study, researchers don’t assign choices; they simply observe them. look.
1 SWE 513: Software Engineering Usability II. 2 Usability and Cost Good usability may be expensive in hardware or special software development User interface.
Design and Evaluation. Overview Formal Evaluations of Visualization Techniques Design (review) Evaluation/Critique of Visualizations (what we’ve been.
Involving Users in Interface Evaluation Marti Hearst (UCB SIMS) SIMS 213, UI Design & Development April 8, 1999.
Formative Evaluation cs3724: HCI. Problem scenarios summative evaluation Information scenarios claims about current practice analysis of stakeholders,
Human Computer Interaction
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
UI Style and Usability, User Experience Niteen Borge.
Usability Testing Chapter 6. Reliability Can you repeat the test?
1 Usability Studies. 2 Evaluate Usability Run a usability study to judge how an interface facilitates tasks with respect to the aspects of usability mentioned.
Usability Testing Chris North cs3724: HCI. Presentations karen molye, steve kovalak Vote: UI Hall of Fame/Shame?
Software Engineering User Interface Design Slide 1 User Interface Design.
User Interface Design & Usability for the Web Card Sorting You should now have a basic idea as to content requirements, functional requirements and user.
Chapter 8 Usability Specification Techniques Hix & Hartson.
The product of evaluation is knowledge. This could be knowledge about a design, knowledge about the user or knowledge about the task.
LInfoVis Winter 2011 Chris Culy Evaluation of visualizations.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
SBD: Usability Evaluation Chris North cs3724: HCI.
Usability Evaluation, part 2. REVIEW: A Test Plan Checklist, 1 Goal of the test? Specific questions you want to answer? Who will be the experimenter?
Prof. James A. Landay University of Washington Autumn 2004 User Testing December 2, 2004.
Design and Evaluation. Design Use Field Guide to identify information relevant to planning visualization.Field Guide Formally plan visualization using.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
SBD: Usability Evaluation Chris North CS 3724: HCI.
Prof. James A. Landay University of Washington Autumn 2006 User Testing November 30, 2006.
User Testing. CSE490f - Autumn 2006User Interface Design, Prototyping, & Evaluation2 Hall of Fame or Hall of Shame? frys.com.
Empirical Evaluation Chris North cs5984: Information Visualization.
Lesson Overview Lesson Overview What Is Science?.
Usability Evaluation or, “I can’t figure this out...do I still get the donuts?”
How do we know if our UI is good or bad?.
Trouble? Can’t type: F11 Can’t hear & speakers okay or can’t see slide? Cntrl R or Go out & come back in 1 Sridhar Rajappan.
11/10/981 User Testing CS 160, Fall ‘98 Professor James Landay November 10, 1998.
Evaluation / Usability. ImplementDesignAnalysisEvaluateDevelop ADDIE.
User Interface Evaluation
Evaluation through user participation
Qualitative vs. Quantitative
SBD: Usability Evaluation
Usability Evaluation, part 2
Title of your science project
Data Collection and Analysis
Professor John Canny Spring 2003
SY DE 542 User Testing March 7, 2005 R. Chow
based on notes by James Landay
User Testing November 27, 2007.
Professor John Canny Spring 2004
Title of your experimental design
Formative Evaluation cs3724: HCI.
Professor John Canny Fall 2004
Presentation transcript:

1 SIMS 247: Information Visualization and Presentation Marti Hearst Nov 30, 2005

2 Today Evaluating Infoviz –What are the goals of evaluation? –How to do usability evaluations Two examples: –Graph viz comparison (nodes & links vs matrix) –Tree navigation (explorer vs. hyperbolic tree) Focus on conflating factors Ideas for evaluating class projects

3 Why do Evaluation? To tell how good or bad a visualization is –People must use it to evaluate it –To compare against the status quo –Something that looks useful to the designer might be too complex or superfluous for real users For iterative design –Interface might be almost right but require adjustments –The interactive components might have problems To advance our knowledge of how people understand and use technology

4 Slide adapted from Chris North's Usability Test vs. Controlled Experiment Usability test: Formative: helps guide design Single UI, early in design process Few participants Identif y usability problems, incidents Qualitative feedback from users Controlled experiment: Summative: measure final result Compare multiple UIs Many participants, strict protocol Independent & dependent variables Quantitative results, statistical significance

5 Slide adapted from Chris North's Controlled Experiments

6 Slide adapted from Chris North's Scientific Method 1.Form Hypothesis 2.Collect data 3.Analyze 4.Accept/reject hypothesis How to “prove” a hypothesis in science? –Easier to disprove things, by counterexample –Null hypothesis = opposite of hypothesis –Disprove null hypothesis –Hence, hypothesis is proved

7 Slide adapted from Chris North's Empirical Experiment Typical question: Which visualization is better in which situations? Spotfirevs.TableLens

8 Slide adapted from Chris North's Cause and Effect Goal: determine “cause and effect” Cause = visualization tool (Spotfire vs. TableLens) Effect = user performance time on task T Procedure: Vary cause Measure effect Problem: random variation Cause = vis tool OR random variation?

9 Slide adapted from Chris North's Variables Independent Variables (what you vary), and treatments (the variable values): –Visualization tool Spotfire, TableLens, Excel –Task type Find, count, pattern, compare –Data size (# of items) 100, 1000, Dependent Variables (what you measure) –User performance time –Errors –Subjective satisfaction (survey)

10 Slide adapted from Chris North's Example: 2 x 3 design n users per cell Task1Task2Task3 Spot-fire Table- Lens Ind Var 1: Vis. Tool Ind Var 2: Task Type Measured user performance times (dep var)

11 Slide adapted from Chris North's Groups “Between subjects” variable –1 group of participants for each variable treatment Group 1: 20 participants, Spotfire Group 2: 20 participants, TableLens Total: 40 participants, 20 per cell “With-in subjects” (repeated) variable –All participants perform all treatments –Counter-balancing order effect Group 1: 20 participants, Spotfire then TableLens Group 2: 20 participants, TableLens then Spotfire Total: 40 users, 40 per cell

12 Selecting Tasks Should reflect what real tasks will be like –may need to shorten if they take too long require background that test user won’t have Be sure tasks measure something directly related to your design But don’t bias the tasks so that only your design can win –should be a realistic task in order to avoid this Don’t choose tasks that are too fragmented

13 Special Considerations for Evaluating Visualizations Be careful about what is being compared Example of how to do it wrong: –One study compared a web path history visualization that had thumbnails fisheye properties hierarchical layout –against the Netscape textual history list Problem: –too many variables changed at once! –can’t tell which of the novel properties caused the effects

14 Important Factors Perceptual abilities –spatial abilities tests –colorblindness –handedness (lefthanded vs. righthanded)

15 Slide adapted from Chris North's Procedure For each participant: Sign legal forms Pre-Survey: demographics Instructions Training runs Actual runs –Give task, measure performance Post-Survey: subjective measures

16 Slide adapted from Chris North's Usability Testing

17 Slide adapted from Chris North's Usability Specification Table Scenario task Worst casePlanned Target Best case (expert) Observed Find most expensive house for sale? 1 min.10 sec.3 sec.??? sec …

18 Slide adapted from Chris North's Usability Test Setup Set of benchmark tasks –Easy to hard, specific to open-ended –Coverage of different UI features –E.g. “find the 5 most expensive houses for sale” Experimenters: –Facilitator: instructs user –Observers: take notes, collect data, video tape screen –Executor: run the prototype if low-fi Participants –3-5; quality not quantity

19 “Think Aloud” Method This is for usability testing, not formal Need to know what users are thinking, not just what they are doing Ask participants to talk while performing tasks –tell us what they are thinking –tell us what they are trying to do –tell us questions that arise as they work –tell us things they read Make a recording or take good notes –make sure you can tell what they were doing

20 Thinking Aloud (cont.) Prompt the user to keep talking –“tell me what you are thinking” Try to only help on things you have pre- decided to help with –keep track of anything you do give help on –if participant stuck or frustrated, then end the task (gracefully) or help them

21 Pilot Study Goal: –help fix problems with the study –make sure you are measuring what you mean to be Procedure: –do twice, first with colleagues then with real users –usually end up making changes both times

22 Slide adapted from Chris North's Usability Test Procedure Goal: mimic real life –Do not cheat by showing them how to use the UI! Initial instructions –“We are evaluating the system, not you.” Repeat: –Give participant a task –Ask participant to “think aloud” –Observe, note mistakes and problems –Avoid interfering, hint only if completely stuck Interview –Verbal feedback –Questionnaire ~1 hour / participant (max)

23 Slide adapted from Chris North's Data Note taking –E.g. user keeps clicking on the wrong button…” Verbal protocol: think aloud –E.g. user expects that button to do something else… Rough quantitative measures –e.g. task completion time,.. Interview feedback and surveys Video-tape screen & mouse

24 Slide adapted from Chris North's Analyze Initial reaction: –“stupid user!”, “that’s developer X’s fault!” Mature reaction: –“how can we redesign UI to solve that usability problem?” –the user is always right Identify usability problems –Learning issues: e.g. can’t figure out or didn’t notice feature –Performance issues: e.g. arduous, tiring to solve tasks –Subjective issues: e.g. annoying, ugly Problem severity: critical vs. minor

25 Slide adapted from Chris North's Cost-Importance Analysis Importance 1-5: (task effect, frequency) 5 = critical, major impact on user, frequent occurance 3 = user can complete task, but with difficulty 1 = minor problem, small speed bump, infrequent Ratio = importance / cost Sort by this 3 categories: Must fix, next version, ignored ProblemImportanceSolutionsCostRatio I/C

26 Slide adapted from Chris North's Refine UI Simple solutions vs. major redesigns Solve problems in order of: importance/cost Example: –Problem: user didn’t know he could zoom in to see more… –Potential solutions: Better zoom button icon, tooltip Add a zoom bar slider (like moosburg) Icons for different zoom levels: boundaries, roads, buildings NOT: more more “help” documentation!!! You can do better. Iterate –Test, refine, test, refine, test, refine, … –Until? Meets usability specification

27 Study Example Mohammad Ghoniem, Jean-Daniel Fekete, Philippe Castagliola. A Comparison of the Readability of Graphs Using Node-Link and Matrix- Based Representations. InfoVis 2004, Austin, TX, Oct IEEE

28 Nodes & Links vs. Matrix Initial studies led to a design modification –Added in color highlighting of moused-over and selected nodes –Needed this because participants became frustrated initial study without them Looked at a set of typical graph operations Varied graphs by # of nodes and connectivity Found that matrix better for all tasks except path following –Better here means faster and higher accuracy on average

29 Study Example Peter Pirolli, Stuart K. Card, and Mija Van Der Wege. The Effect of Information Scent on Searching Information Visualizations of Large Tree Structures. AVI 2000, Palermo, Italy.

30 Main Conclusions Results of the Browse-Off were a function of the types of tasks, the structure of the information, and the skills of the participants –NOT the difference in the browsers The Hyperbolic Browser can be faster when –The clues about where to go next are clear –The tasks are relatively simple

31 Lessons for Study Design This paper contains a wealth of good ideas about how to –Isolate what’s really going on –Assess and understand the data

32 Assessing Infoviz Projects How does all this apply to you?