ISE554 The WWW 3.4 Evaluation Methods. Evaluating Interfaces with Users Why evaluation is crucial to interface design General approaches and tradeoffs.

Slides:



Advertisements
Similar presentations
Internal Assessment Your overall IB mark (the one sent to universities after the IB test) in any IB science course is based upon two kinds of assessments.
Advertisements

Evaluating interfaces with users Why evaluation is crucial Quickly debug prototypes by observing people use them Methods reveal what a person is thinking.
Validity (cont.)/Control RMS – October 7. Validity Experimental validity – the soundness of the experimental design – Not the same as measurement validity.
Copyright © 2005, Pearson Education, Inc. An Instructor’s Outline of Designing the User Interface 4th Edition by Ben Shneiderman & Catherine Plaisant Slides.
6.811 / PPAT: Principles and Practice of Assistive Technology Wednesday, 16 October 2013 Prof. Rob Miller Today: User Testing.
CS305: HCI in SW Development Evaluation (Return to…)
The Ways and Means of Psychology STUFF YOU SHOULD ALREADY KNOW BY NOW IF YOU PLAN TO GRADUATE.
Conducting a User Study Human-Computer Interaction.
Methodology Overview Why do we evaluate in HCI? Why should we use different methods? How can we compare methods? What methods are there? see
Welcome to Survey of Research Methods and Statistics! PS 510.
Methodology Overview Dr. Saul Greenberg John Kelleher.
1 User Centered Design and Evaluation. 2 Overview Why involve users at all? What is a user-centered approach? Evaluation strategies Examples from “Snap-Together.
Doing Social Psychology Research
6.S196 / PPAT: Principles and Practice of Assistive Technology Monday, 28 Nov Prof. Rob Miller Today: User Testing & Ethics.
PSYCO 105: Individual and Social Behaviour Lecture 1: The Ways and Means of Psychology.
James Tam Qualitative Evaluation Techniques Why evaluation is crucial to interface design General approaches and tradeoffs with the different approaches.
Psych 231: Research Methods in Psychology
ICS 463, Intro to Human Computer Interaction Design: 9. Experiments Dan Suthers.
1 User Centered Design and Evaluation. 2 Overview My evaluation experience Why involve users at all? What is a user-centered approach? Evaluation strategies.
Saul Greenberg Evaluating Interfaces With Users Why evaluation is crucial to interface design General approaches and tradeoffs in evaluation The role of.
Sampling & Experimental Control Psych 231: Research Methods in Psychology.
James Tam Evaluating Interfaces With Users Why evaluation is crucial to interface design General approaches and tradeoffs in evaluation The role of ethics.
Methods of Psychology Hypothesis: A tentative statement about how or why something happens. e.g. non experienced teachers use corporal punishment more.
Chapter 8 Experimental Research
Empirical Evaluation Assessing usability (with users)
류 현 정류 현 정 Human Computer Interaction Introducing evaluation.
Identify and List… Theory behind the study Aim of study.
Conducting a User Study Human-Computer Interaction.
Chapter 11: An Evaluation Framework Group 4: Tony Masi, Sam Esswein, Brian Rood, & Chris Troisi.
Presentation: Techniques for user involvement ITAPC1.
Research Methods Irving Goffman People play parts/ roles
Human Computer Interaction
Conducting a User Study Human-Computer Interaction.
Chapter 2 AP Psychology Outline
Module 4 Notes Research Methods. Let’s Discuss! Why is Research Important?
© 2004 John Wiley & Sons, Inc. Huffman: PSYCHOLOGY IN ACTION, 7E Psychological Research and Ethics.
Conducting a User Study Human-Computer Interaction.
Usability Testing Chapter 6. Reliability Can you repeat the test?
Midterm Stats Min: 16/38 (42%) Max: 36.5/38 (96%) Average: 29.5/36 (78%)
 Descriptive Methods ◦ Observation ◦ Survey Research  Experimental Methods ◦ Independent Groups Designs ◦ Repeated Measures Designs ◦ Complex Designs.
Experiment Basics: Variables Psych 231: Research Methods in Psychology.
Human-Computer Interaction. Overview What is a study? Empirically testing a hypothesis Evaluate interfaces Why run a study? Determine ‘truth’ Evaluate.
Research Strategies. Why is Research Important? Answer in complete sentences in your bell work spiral. Discuss the consequences of good or poor research.
Ethics in Evaluation Why ethics? What you have to do Slide deck by Saul Greenberg. Permission is granted to use this for non-commercial purposes as long.
G063 – Prototyping. Learning Objective: At the end of this topic you should be able to: describe prototyping as a software development methodology.
Methodology: How Social Psychologists Do Research
Oct 211 The next two weeks Oct 21 & 23: Lectures on user interface evaluation Oct 28: Lecture by Dr. Maurice Masliah No office hours (out of town) Oct.
Spring /6.831 User Interface Design and Implementation1 Lecture 13: User Testing next 3 lectures required for 6.831G only nanoquizzes on those.
Usability Evaluation or, “I can’t figure this out...do I still get the donuts?”
How Psychologists Do Research Chapter 2. How Psychologists Do Research What makes psychological research scientific? Research Methods Descriptive studies.
How do we know if our UI is good or bad?.
Sociology 12 Acad. New Unit: Sociological Research Methods.
Chapter 2: The Research Enterprise in Psychology.
School of Engineering and Information and Communication Technology KIT305/607 Mobile Application Development Week 7: Usability (think-alouds) Dr. Rainer.
Research in Psychology. Quantitative Methods  Quantitative: experiments and studies gathering data with questionnaires and analyzing results with correlations.
Engineering Design Process. First, let’s review the “Scientific Method” 1.Ask a question 2.Research 3.Procedure/Method 4.Data Observation 5.Conclusion.
What Do We Mean by Usability?
Approaches to social research Lerum
1.2 Research Methods AP Psychology.
Evaluating interfaces with users
Methodology Overview 2 basics in user studies Lecture /slide deck produced by Saul Greenberg, University of Calgary, Canada Notice: some material in this.
Conducting a User Study
Evaluating interfaces with users
Experimental Design.
Experimental Design.
ABAB Design Ethical considerations
IS THE RESEARCH MEASURING WHAT IT AIMED TO MEASURE?
How psychologist know what they know
AS Psychology Research Methods
Presentation transcript:

ISE554 The WWW 3.4 Evaluation Methods

Evaluating Interfaces with Users Why evaluation is crucial to interface design General approaches and tradeoffs in evaluation The role of ethics

Why Bother? Tied to the usability engineering lifecycle Pre-design -investing in new expensive system requires proof of viability Initial design stages -develop and evaluate initial design ideas with the user Iterative design -does system behaviour match the user’s task requirements? -are there specific problems with the design? -can users provide feedback to modify design? Acceptance testing -verify that human/computer system meets expected performance criteria ease of learning, usability, user’s attitude, performance criteria e.g., a first time user will take 1-3 minutes to learn how to withdraw $50. from the automatic teller design implementationevaluation

Approaches: Naturalistic Naturalistic: describes an ongoing process as it evolves over time observation occurs in realistic setting -ecologically valid “real life” External validity degree to which research results applies to real situations

Approaches: Experimental Experimental study relations by manipulating one or more independent variables -experimenter controls all environmental factors observe effect on one or more dependent variables Internal validity confidence that we have in our explanation of experimental results Trade-off: Natural vs Experimental precision and direct control over experimental design versus desire for maximum generalizability in real life situations

Approaches: Reliability Concerns Would the same results be achieved if the test were repeated? Problem: individual differences: best user 10x faster than slowest best 25% of users ~2x faster than slowest 25% Partial Solution reasonable number and range of users tested statistics provide confidence intervals of test results -95% confident that mean time to perform task X is 4.5+/-0.2 minutes means 95% chance true mean is between 4.3 and 4.7, 5% chance its outside that

Approaches: Validity Concerns Does the test measure something of relevance to usability of real products in real use outside of lab? Some typical reliability problems of testing vs real use -non-typical users tested -tasks are not typical tasks -physical environment different quiet lab vs very noisy open offices vs interruptions -social influences different motivation towards experimenter vs motivation towards boss Partial Solution use real users tasks from task-centered system design environment similar to real situation

Ethics... and to think that you want me to test it!!!

Ethics Testing can be a distressing experience pressure to perform, errors inevitable feelings of inadequacy competition with other subjects Golden rule subjects should always be treated with respect

Managing subjects in an ethical manner Before the test don’t waste the user’s time -use pilot tests to debug experiments, questionnaires etc -have everything ready before the user shows up make users feel comfortable -emphasize that it is the system that is being tested, not the user -acknowledge that the software may have problems -let users know they can stop at any time maintain privacy -tell user that individual test results will be kept completely confidential inform the user -explain any monitoring that is being used -answer all user’s questions (but avoid bias) only use volunteers -user must sign an informed consent form

Managing subjects in an ethical manner During the test don’t waste the user’s time -never have the user perform unnecessary tasks make users comfortable -try to give user an early success experience -keep a relaxed atmosphere in the room -coffee, breaks, etc -hand out test tasks one at a time -never indicate displeasure with the user’s performance -avoid disruptions -stop the test if it becomes too unpleasant maintain privacy -do not allow the user’s management to observe the test

Managing subjects in an ethical manner After the test make the users feel comfortable -state that the user has helped you find areas of improvement inform the user -answer particular questions about the experiment that could have biased the results before maintain privacy -never report results in a way that individual users can be identified -only show videotapes outside the research group with the user’s permission

You know now Evaluation is crucial for designing, debugging, and verifying interfaces There is a tradeoff in naturalistic vs experimental approaches internal and external validity reliability precision generalizability Subjects must be treated with respect ethical rules of behaviour