Evaluation Methods April 20, 2005 Tara Matthews CS 160.

Slides:



Advertisements
Similar presentations
Chapter 15: Analytical evaluation
Advertisements

Chapter 14: Usability testing and field studies
Evaluation of User Interface Design
Human Computer Interaction
Human Computer Interaction
CS305: HCI in SW Development Evaluation (Return to…)
IS214 Recap. IS214 Understanding Users and Their Work –User and task analysis –Ethnographic methods –Site visits: observation, interviews –Contextual.
Chapter 14: Usability testing and field studies. 2 FJK User-Centered Design and Development Instructor: Franz J. Kurfess Computer Science Dept.
©N. Hari Narayanan Computer Science & Software Engineering Auburn University 1 COMP 7620 Evaluation Chapter 9.
Methodology Overview Dr. Saul Greenberg John Kelleher.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check.
Empirical Methods in Human- Computer Interaction.
Evaluation Methodologies
Design and Evaluation of Iterative Systems n For most interactive systems, the ‘design it right first’ approach is not useful. n The 3 basic steps in the.
Usable Privacy and Security Carnegie Mellon University Spring 2008 Lorrie Cranor 1 Designing user studies February.
Usability 2004 J T Burns1 Usability & Usability Engineering.
Testing and Modeling Users Kristina Winbladh & Ramzi Nasr.
1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.
Evaluation How do we test the interaction design? Several Dimensions
From Controlled to Natural Settings
ICS 463, Intro to Human Computer Interaction Design: 8. Evaluation and Data Dan Suthers.
©2011 1www.id-book.com Analytical evaluation Chapter 15.
Predictive Evaluation
Action Research March 12, 2012 Data Collection. Qualities of Data Collection  Generalizability – not necessary; goal is to improve school or classroom.
Ch 14. Testing & modeling users
Evaluation Techniques Material from Authors of Human Computer Interaction Alan Dix, et al.
Scientific Process ► 1) Developing a research idea and hypothesis ► 2) Choosing a research design (correlational vs. experimental) ► 3) Choosing subjects.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
Human Computer Interaction
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
The Research Enterprise in Psychology
Evaluation Techniques zEvaluation ytests usability and functionality of system yoccurs in laboratory, field and/or in collaboration with users yevaluates.
Testing & modeling users. The aims Describe how to do user testing. Discuss the differences between user testing, usability testing and research experiments.
Assumes that events are governed by some lawful order
Experimentation in Computer Science (Part 1). Outline  Empirical Strategies  Measurement  Experiment Process.
New Media Research Methods- Part 1 Planning stage: Using Methods, Data and Tools to explore User’s Experience. How methods relate to research questions?
GOMs and Action Analysis and more. 1.GOMS 2.Action Analysis.
CS2003 Usability Engineering Usability Evaluation Dr Steve Love.
Chapter 15: Analytical evaluation. Inspections Heuristic evaluation Walkthroughs.
Evaluation Techniques Evaluation –tests usability and functionality of system –occurs in laboratory, field and/or in collaboration with users –evaluates.
Software Engineering User Interface Design Slide 1 User Interface Design.
Chapter 8 Usability Specification Techniques Hix & Hartson.
Usability Engineering Dr. Dania Bilal IS 582 Spring 2006.
SBD: Analyzing Requirements Chris North CS 3724: HCI.
Fall 2002CS/PSY Empirical Evaluation Data collection: Subjective data Questionnaires, interviews Gathering data, cont’d Subjective Data Quantitative.
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
The effects of Peer Pressure, Living Standards and Gender on Underage Drinking Psychologist- Kanari zukoshi.
Chapter 15: Analytical evaluation. Aims: Describe inspection methods. Show how heuristic evaluation can be adapted to evaluate different products. Explain.
Paper III Qualitative research methodology.  Qualitative research is designed to reveal a specific target audience’s range of behavior and the perceptions.
Oct 211 The next two weeks Oct 21 & 23: Lectures on user interface evaluation Oct 28: Lecture by Dr. Maurice Masliah No office hours (out of town) Oct.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
Prototyping Creation of concrete but partial implementations of a system design to explore usability issues.
Human Computer Interaction
User Interface Evaluation
SIE 515 Design Evaluation Lecture 7.
Chapter 2 Sociological Research Methods
Methods Choices Overall Approach/Design
Introducing Evaluation
Evaluation techniques
Model based design.
Evaluation.
HCI Evaluation Techniques
CSM18 Usability Engineering
Testing & modeling users
Evaluation Techniques
Experimental Evaluation
Empirical Evaluation Data Collection: Techniques, methods, tricks Objective data IRB Clarification All research done outside the class (i.e., with non-class.
Presentation transcript:

Evaluation Methods April 20, 2005 Tara Matthews CS 160

In 160 We’ve Covered… Task Analysis & Contextual Inquiry Cognitive Walkthrough Heuristic Evaluation WOZ usability study w/ paper prototypes

There are many more methods… Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

How to chose a method? Stage of study –formative, iterative, summative Pros & cons Metrics –depends on what you want to measure Qualitative vs. quantitative Research perspective –CS vs. psychology vs. sociology

Pros & Cons Realism Precision Generalizability Time & cost Researcher expertise

Methods Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Survey Online / paper questionnaires distributed to target audience Can be used to –tabulate quantitative data –gather qualitative feedback (opinions, feelings, etc.) Useful at any time in study

Survey Pros –Easy to get a large number of responses. –Quick and easy to conduct. –Highly generalizable. Cons –Self-selection. –Participants often only offer enough information to answer the question. –Can miss details. –Low in realism and precision.

Interview Evaluators formulate questions on the issues of interest. Interview representative users, asking them these questions in order to gather information desired. Interviewer reads questions to user, who replies verbally; interviewer records responses.

Interview Pros –Quick and easy to conduct. –Gives designer quick feedback on a range of ideas. –Can get a person’s initial reaction to an idea. –Can get detailed information from a person. Cons –Often takes place away from natural setting. –Question wording or interviewer “body language” can bias answers. –High probability of false positives and missed problems (e.g., users may not have a clear idea of how an app will be used). –Can miss details if interviewer doesn’t know what issues to draw out.

Controlled Lab Experiment In lab, manipulate one feature of a system to assess the causal effects of the difference in that manipulated feature on other behaviors of the system. Example: –in lab, show users 4 versions of a website: blue, yellow, red, and black text –measure time to find specific words –compare

Controlled Lab Experiment Pros –Provides precise, quantifiable data. –Easier to draw inferences from data. –Relatively quick. –Can get a medium-sized number of participants. Cons –Short duration of a lab experiment may not be enough to allow users to become accustomed to an app. –Not a natural setting – interaction may not be normal.

In-lab Observation Participants come to lab to "use" an interface Given sample tasks to complete with it Evaluators observe and possibly audio- or videotape Participants may "think out loud" Can use lo-fi prototype (for a project in the design stage) to an almost-complete interface Evaluators note participants’ –emotions, exclamations, facial expressions, and other "qualitative" data –take note of quantitative data such as time to complete a task or number of errors

In-lab Observation Pros –Relatively quick. –Can get a medium-sized number of participants. Cons –Observations are subjective and error prone. –Short duration of lab observation is not enough time for user to get accustomed to using the interface. –Not a natural setting – interaction may not be normal.

Controlled Field Experiment In natural setting, manipulate one feature of a system to assess the causal effects of the difference in that manipulated feature on other behaviors of the system. Example: –Participants use 3 different input devices in their own office: mouse with 1, 2, or 3 buttons –Perform a set of tasks –Measure differences

Controlled Field Experiment Pros –Less intrusive than most other evaluation methods. –Provides more precise data than field observation. –Can observe natural behavior of user (though some part of the system will be controlled/unnatural). Cons –More intrusive than field observation. –Less natural than field observation.

Field Observation Study Evaluator makes direct observations of “natural” systems Takes care to not intrude on / disturb those systems A.K.A. “ethnography”

Field Observation Study Pros –Only way to observe natural behavior of user & interaction between user & tools. Cons –Difficult and time consuming. –Hard to get permission to observe people. –Observations are subjective and error prone. –Cannot make strong interpretations from observations. –Not very generalizable.

Heuristic Evaluation Pros –Quick and easy. Cons –Nielson’s heuristics may not be as relevant to non-GUIs. –Results in false positives in missed problems, especially when experts are not part of target audience.

Cognitive Walkthrough Pros –Quick and easy. Cons –Results in false positives and missed problems when evaluator is different from target audience.

Automate Observation Study Techniques include –video or audio recording of user –pop-up screens –screen shots –time logging –log users actions (collecting statistics about detailed system use)

Automate Observation Study Pros –Eases burden on observers for data collection & analysis. Cons –Setup is often more time-consuming to complete. –Harder to get approved if it involves analysis of videotape or audiotape. –May miss nuanced/interpretive details.

Experimental Simulation In-lab experiment that is as much like some real situation as possible. Example: –ground-based flight simulator –behaves as closely as possible to a real flight –still under researcher control

Experimental Simulation Pros –Still fairly precise. –More realistic than in-lab experiment. Cons (same as lab exp.) –Short duration of a lab experiment may not be enough to allow users to become accustomed to an app. –Not a natural setting – interaction may not be normal.

Claims Analysis Claim = statement that a certain aspect (button, scrollbar) of a design has psychological implications reflected in how capable a user is in using that design UI artifacts are listed along with their design features & pros/cons Helps –select among alternative designs –clarify questions to be analyzed through user testing by stating how the design should work (in claims)

GOMS A method to describe user tasks and how a user performs those tasks with a specific interface design Views humans as information processors –Small number of cognitive, perceptual, and motor operators characterize user behavior To apply GOMS: –Analyze task to identify user goals (hierarchical) –Identify operators to achieve goals –Sum operator times to predict performance GOMS = –Goals: What a user wants to accomplish –Operators: Cognitive or physical actions that change the state of the user or the system –Methods: Groups of goals and operators –Selection rules: Determine which method to apply

GOMS Pros –Predict human performance before committing to a specific design in code or running user studies –Many studies have validated the model (it works) Cons –Assumes error-free, skilled user behavior –No formal recipe for how to perform analysis –Significant time investment

Computer Simulation Creating a complete & closed system that models the operation of the concrete system without users. Example: –geophysical process going on in connection with the eruption of Mount St. Helens

Computer Simulation Pros –Supposedly high in realism (depends on accuracy of data/system replication) Cons –Low in precision & generalizability

Formal Theory Formulating general relations (propositions, hypothesis, or postulates) among a number of variables of interest. Pros –Relatively generalizable Cons –Not realistic or precise

How to chose a method? Stage of study Pros & cons –Realism –Precision –Generalizability –Time & cost Researcher expertise Metrics Qualitative vs. quantitative Research perspective

Methods Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Early Stage Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Early Stage Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Iterative & Summative Stages Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Iterative & Summative Stages Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Realism Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Realism Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Precision Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Precision Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Generalizability Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Generalizability Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Time & Cost Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Time & Cost Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Researcher Perspective Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Metrics: examples Traditional GUIs: –efficiency (time to complete task) –accuracy (# of errors) –simplicity Peripheral Displays: –awareness (recall) –distraction (dual-task behavior) –aesthetics

Peripheral Displays Survey Interview Controlled-lab experiment In-lab observation Controlled field experiment Field observation study Heuristic Evaluation Cognitive Walkthrough Contextual Inquiry Automated observation user study Experimental simulation Claims analysis GOMS Computer simulation Formal theory

Questions?