Design and Evaluation. Overview Formal Evaluations of Visualization Techniques Design (review) Evaluation/Critique of Visualizations (what we’ve been.

Slides:



Advertisements
Similar presentations
Critical Reading Strategies: Overview of Research Process
Advertisements

Lesson Overview 1.1 What Is Science?.
Copyright © 2014 by The University of Kansas Collecting and Analyzing Data.
1 i247: Information Visualization and Presentation Marti Hearst How to Critique a Visualization.
Information Retrieval: Human-Computer Interfaces and Information Access Process.
Department of Industrial Management Engineering 1.Introduction ○Usability evaluation primarily summative ○Informal intuitive evaluations by designers even.
Data analysis and interpretation. Agenda Part 2 comments – Average score: 87 Part 3: due in 2 weeks Data analysis.
Social Science Research and
1 SIMS 247: Information Visualization and Presentation Marti Hearst Nov 30, 2005.
USABILITY AND EVALUATION Motivations and Methods.
Semester in review. The Final May 7, 6:30pm – 9:45 pm Closed book, ONE PAGE OF NOTES Cumulative Similar format to midterm (probably about 25% longer)
©N. Hari Narayanan Computer Science & Software Engineering Auburn University 1 COMP 7620 Evaluation Chapter 9.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Information Retrieval: Human-Computer Interfaces and Information Access Process.
CS 5764 Information Visualization Dr. Chris North.
1 Validation and Verification of Simulation Models.
1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.
© Associate Professor Dr. Jamil Bojei, 2007
Copyright © 2010 SAS Institute Inc. All rights reserved. Effective Data Visualization Design for Dashboards Lisa Whitman TriUPA May 25, 2011.
SBD: Usability Evaluation
Evaluation: Controlled Experiments Chris North cs3724: HCI.
Chapter 1: Introduction to Statistics
Predictive Evaluation
Information Design and Visualization
 For the IB Diploma Programme psychology course, the experimental method is defined as requiring: 1. The manipulation of one independent variable while.
Term 2, 2011 Week 1. CONTENTS Types and purposes of graphic representations Spreadsheet software – Producing graphs from numerical data Mathematical functions.
Analyzing Reliability and Validity in Outcomes Assessment (Part 1) Robert W. Lingard and Deborah K. van Alphen California State University, Northridge.
Tutor: Prof. A. Taleb-Bendiab Contact: Telephone: +44 (0) CMPDLLM002 Research Methods Lecture 8: Quantitative.
T tests comparing two means t tests comparing two means.
Object-Oriented Software Engineering Practical Software Development using UML and Java Chapter 7: Focusing on Users and Their Tasks.
Writing the “Results” & “Discussion” sections Awatif Alam Professor Community Medicine Medical College/ KSU.
Today: Assignment 2 misconceptions Our process True Experiments: Single-Factor Design Assignment 3 Q&A Mid-term: format, coverage.
How to read a scientific paper
Evaluation of User Interface Design 4. Predictive Evaluation continued Different kinds of predictive evaluation: 1.Inspection methods 2.Usage simulations.
Where did plants and animals come from? How did I come to be?
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
1.1 Statistical Analysis. Learning Goals: Basic Statistics Data is best demonstrated visually in a graph form with clearly labeled axes and a concise.
Info Vis: Multi-Dimensional Data Chris North cs3724: HCI.
EDG Evaluation Phase 1 12/08/03 Xin Chen Javier Tuason Dennis Frezzo.
Scientific Methods and Terminology. Scientific methods are The most reliable means to ensure that experiments produce reliable information in response.
Question paper 1997.
Presenting and Analysing your Data CSCI 6620 Spring 2014 Thesis Projects: Chapter 10 CSCI 6620 Spring 2014 Thesis Projects: Chapter 10.
SBD: Usability Evaluation Chris North cs3724: HCI.
Copyright 2010, The World Bank Group. All Rights Reserved. COMMUNICATION AND DISSEMINATION, PART 2 DEVELOPING DISSEMINATION PRODUCTS 1.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
C82MST Statistical Methods 2 - Lecture 1 1 Overview of Course Lecturers Dr Peter Bibby Prof Eamonn Ferguson Course Part I - Anova and related methods (Semester.
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
Design and Evaluation. Design Use Field Guide to identify information relevant to planning visualization.Field Guide Formally plan visualization using.
Lesson Overview Lesson Overview What Is Science? Lesson Overview 1.1 What Is Science?
Some Alternative Approaches Two Samples. Outline Scales of measurement may narrow down our options, but the choice of final analysis is up to the researcher.
SBD: Usability Evaluation Chris North CS 3724: HCI.
Empirical Evaluation Chris North cs5984: Information Visualization.
Research Word has a broad spectrum of meanings –“Research this topic on ….” –“Years of research has produced a new ….”
Lesson Overview Lesson Overview What Is Science?.
mQ OBJECTIVES The student should be able to: 1.list and describe the steps of the scientific method 2.define.
URBDP 591 A Lecture 16: Research Validity and Replication Objectives Guidelines for Writing Final Paper Statistical Conclusion Validity Montecarlo Simulation/Randomization.
Welcome! Seminar – Monday 6:00 EST HS Seminar Unit 1 Prof. Jocelyn Ramos.
1 Collecting and Interpreting Quantitative Data Deborah K. van Alphen and Robert W. Lingard California State University, Northridge.
Framework and Models. Framework To help understand the field To develop a system that will allow us to ▫ Develop good designs ▫ Test ▫ Evaluate We need.
Design Evaluation Overview Introduction Model for Interface Design Evaluation Types of Evaluation –Conceptual Design –Usability –Learning Outcome.
Visualization Design Principles cs5984: Information Visualization Chris North.
DATA VISUALIZATION BOB MARSHALL, MD MPH MISM FAAFP FACULTY, DOD CLINICAL INFORMATICS FELLOWSHIP.
Integrating Graphics, Illustrations, Figures, Charts.
THE SCIENTIFIC METHOD Murtaugh 1A Living Environment.
The Scientific Method & Experimental Design
SBD: Usability Evaluation
Data Collection and Analysis
Evaluation techniques
HCI Evaluation Techniques
The Scientific Method & Experimental Design
Presentation transcript:

Design and Evaluation

Overview Formal Evaluations of Visualization Techniques Design (review) Evaluation/Critique of Visualizations (what we’ve been doing all semester)

Formal Evaluations of Visualization Techniques

Evaluating Visualizations Empirical: ▫ Usability Test  Observation, problem identification ▫ Controlled Experiment  Formal controlled scientific experiment  Comparisons, statistical analysis Analytic: ▫ Expert Review  Examination by visualization expert ▫ Heuristic Evaluation  Principles, Guidelines  Algorithmic

Usability test vs. Controlled Experiment Usability test:  Formative: helps guide design  Single UI, early in design process  Few users  Usability problems, incidents  Qualitative feedback from users Controlled experiment:  Summative: measure final result  Compare multiple UIs  Many users, strict protocol  Independent & dependent variables  Quantitative results, statistical significance

Controlled Experiments (a refresher)

What is Science? Measurement Modeling

Scientific Method 1. Form Hypothesis 2. Collect data 3. Analyze 4. Accept/reject hypothesis How to “prove” a hypothesis in science?  Easier to disprove things, by counterexample  Null hypothesis = opposite of hypothesis  Disprove null hypothesis  Hence, hypothesis is proved

Empirical Experiment Typical question:  Which visualization is better in which situations? Spotfirevs.TableLens

Cause and Effect Goal: determine “cause and effect”  Cause = visualization tool (Spotfire vs. TableLens)  Effect = user performance time on task T Procedure:  Vary cause  Measure effect Problem: random variation  Cause = vis tool OR random variation? Real world Collected data random variation uncertain conclusions

Stats to the Rescue Goal:  Measured effect unlikely to result by random variation Hypothesis:  Cause = visualization tool (e.g. Spotfire ≠ TableLens) Null hypothesis:  Visualization tool has no effect (e.g. Spotfire = TableLens)  Hence: Cause = random variation Stats:  If null hypothesis true, then measured effect occurs with probability < 5%  But measured effect did occur! (e.g. measured effect >> random variation) Hence:  Null hypothesis unlikely to be true  Hence, hypothesis likely to be true

Variables Independent Variables (what you vary), and treatments (the variable values):  Visualization tool ▫ Spotfire, TableLens, Excel  Task type ▫ Find, count, pattern, compare  Data size (# of items) ▫ 100, 1000, Dependent Variables (what you measure)  User performance time  Accuracy, Errors  Subjective satisfaction (survey)  HCI metrics

Example: 2 x 3 design n users per cell Task1Task2Task3 Spot- fire Table- Lens Ind Var 1: Vis. Tool Ind Var 2: Task Type Measured user performance times (dep var)

Step 1: Visualize it Dig out interesting facts Qualitative conclusions Guide stats Guide future experiments

Step 2: Stats Task1Task2Task3 Spot- fire Table- Lens Ind Var 1: Vis. Tool Ind Var 2: Task Type Average user performance times (dep var)

TableLens better than Spotfire? Problem with Averages ? Spotfire TableLens Avg Perf time (secs)

TableLens better than Spotfire? Problem with Averages: lossy  Compares only 2 numbers  What about the 40 data values? (Show me the data!) Spotfire TableLens Avg Perf time (secs)

The real picture Need stats that compare all data Spotfire TableLens Avg Perf time (secs)

p < 0.05 Woohoo! Found a “statistically significant” difference Averages determine which is ‘better’ Conclusion:  Cause = visualization tool (e.g. Spotfire ≠ TableLens)  Vis Tool has an effect on user performance for task T …  “95% confident that TableLens better than Spotfire …”  NOT “TableLens beats Spotfire 95% of time”  5% chance of being wrong!  Be careful about generalizing

p > 0.05 Hence, no difference?  Vis Tool has no effect on user performance for task T…?  Spotfire = TableLens ?

p > 0.05 Hence, no difference?  Vis Tool has no effect on user performance for task T…?  Spotfire = TableLens ? NOT!  Did not detect a difference, but could still be different  Potential real effect did not overcome random variation  Provides evidence for Spotfire = TableLens, but not proof  Boring, basically found nothing How?  Not enough users  Need better tasks, data, …

Class Exercise—Formal Evaluation Usability Analysis Controlled Experiment

Design (revisited)

Design Use Field Guide to identify information relevant to planning visualization.Field Guide Formally plan visualization using CUT-AD-DDV. Develop and try some visualizations. Mockup by hand, or by powerpoint. Test with users. Then develop real one. Don’t be afraid to modify/change. Do testing with users to get feedback. Do formal evaluation if needed.

Full Framework: CUTT-AD-DDV Given, and should be identified by designer Context User Task Data Types Generally known Human Abilities (perception, memory, cognition) Design Principles These you have some control over Data Model Display Visualization Techniques

CUTT-AD-DDV Visualization Framework Data Model Display(s) Visualization Techniques Design Process Iterative design Design studies Evaluation Design Principles Visual display Interaction Context User Tasks Data Types Human Abilities Visual perception Cognition Memory Motor skills Imply Constrain design Inform design Given Chosen

Evaluation and Critique

Evaluation Critiquing Formal user studies

Critiquing a Visualization First, consider your reaction to the initial (Message) presentation ▫ 1-2 seconds (Yes/No, Good/Bad) ▫ 5-7 seconds (understandable, interesting, makes me want to ask questions, want to explore) Second, consider your reaction to the exploration part of the visualization ▫ Study the visualization in more detail to evaluate it. ▫ If there is an interactive exploration capability explore that

Critiquing a Visualization Third, identify the purpose, content and choices made in this visualization using our CUTT-AD- DDV model ▫ Identify the purpose of the visualization and who the intended audience is (Context/User/Task from our model). ▫ Identify the Data Types available, and what was chosen for the Data Model ▫ Identify display type(s) utilized (or expected). ▫ Identify Display techniques utilized.

Questions to Ask Is the design visually appealing/aesthetically pleasing? Is it immediately understandable? If not, is it understandable after a short period of study? Does it provide insight or understanding that was not obtainable with the original representation (text, table, etc)? Was the choice of data model appropriate? Does it transform nominal, interval, and quantitative information properly?

Questions to Ask Was the choice of display appropriate? Does it use visual components properly? That is, does it properly represent the data using lines, color, position, etc? Does the visualization provide insight or understanding better than some alternative visualization would? What kind of visualization technique might have been better? Does it use labels and legends appropriately?

How Successful is the Visualization? Does the visualization reveal trends, patterns, gaps, and/or outliers? Does the visualization successfully highlight important information, while providing context for that information? Does it distort the information? If it transforms it in some way, is this misleading or helpfully simplifying? Does it omit important information? Does it engage you, make you interested, want to explore?

Example from NY Times on housing priceshousing prices ▫ Message  1-2 second understanding?  5-7 second understanding  Engage you? (could they have done more?) ▫ Highlighting of areas of interest, accompanying narrative text ▫ Dual axis appropriate to data values

Presentation in Time and Space Static vs. Dynamic Information. ▫ Is there a good initial message presentation? ▫ Is there an obvious way to explore the data? ▫ Did the visualization appropriately divide information into these two parts? Make use of screen space (not all space is created equal). What should be highlighted (Don’t highlight too much!)

Data-to-Pixel Ratios Examples of wasteful pixels: ▫ Decorative graphics ▫ Meaningless variations in color ▫ Borders to separate when white space will suffice ▫ Distracting backgrounds and fill colors ▫ Unnecessary grid lines in tables and graphs ▫ 3D when not needed (nearly always not needed) ▫ Ornate/physical looking widgets

Data-to-Pixel Ratios How to make non-data pixels work well: ▫ Use light, low-saturated colors  Hard to see when printed though! ▫ Regularize the appearance of non-data visual elements.

Few on “Eloquence Through Simplicity” Well-designed dashboards are: ▫ Well-organized ▫ Condensed  Summaries and exceptions ▫ Concise, communicate the data and its message in the clearest and most direct way possible. ▫ Can be assimilated quickly ▫ Provide a broad, high-level overview ▫ Address specific information needs

More General Guidelines Simplify, simplify, simplify! Must be the best way to show this data.

User Studies Discussion of Readings on user studies. Example(s)

Class Exercise—Evaluation Putting it all together Browser Market Share Movie Ebb & Flow Risky Roads AOL Revenue Stream Divide in groups of 4; each group works on one problem using Field Guide and CUTT-AD- DDV, and proposes evaluation.