STAT – User testing and evaluation 1 User testing and evaluation Lecturer: Smilen Dimitrov Applied statistics for testing and evaluation – MED4.

Slides:



Advertisements
Similar presentations
Chapter 14: Usability testing and field studies
Advertisements

Evaluation of User Interface Design
Human Computer Interaction
March 19, 2002 Internet Librarian International Darlene Fichter Data Coordinator, University of Saskatchewan Libraries
6.811 / PPAT: Principles and Practice of Assistive Technology Wednesday, 16 October 2013 Prof. Rob Miller Today: User Testing.
CS305: HCI in SW Development Evaluation (Return to…)
Department of Industrial Management Engineering 1.Introduction ○Usability evaluation primarily summative ○Informal intuitive evaluations by designers even.
©N. Hari Narayanan Computer Science & Software Engineering Auburn University 1 COMP 7620 Evaluation Chapter 9.
Methodology Overview Dr. Saul Greenberg John Kelleher.
Usability presented by the OSU Libraries’ u-team.
Evaluation. formative 4 There are many times throughout the lifecycle of a software development that a designer needs answers to questions that check.
Empirical Methods in Human- Computer Interaction.
Design and Evaluation of Iterative Systems n For most interactive systems, the ‘design it right first’ approach is not useful. n The 3 basic steps in the.
Usable Privacy and Security Carnegie Mellon University Spring 2008 Lorrie Cranor 1 Designing user studies February.
An evaluation framework
HCI revision lecture. Main points Understanding Applying knowledge Knowing key points Knowing relationship between things If you’ve done the group project.
Usability Evaluation Methods Computer-Mediated Communication Lab week 10; Tuesday 7/11/06.
Mid-Term Exam Review IS 485, Professor Matt Thatcher.
Damian Gordon.  Summary and Relevance of topic paper  Definition of Usability Testing ◦ Formal vs. Informal methods of testing  Testing Basics ◦ Five.
1. Learning Outcomes At the end of this lecture, you should be able to: –Define the term “Usability Engineering” –Describe the various steps involved.
Chapter 14: Usability testing and field studies
McGraw-Hill © 2006 The McGraw-Hill Companies, Inc. All rights reserved. The Nature of Research Chapter One.
Dr. Engr. Sami ur Rahman Assistant Professor Department of Computer Science University of Malakand Research Methods in Computer Science Lecture: Research.
Predictive Evaluation
User Interface Evaluation Usability Testing Methods.
Evaluation Framework Prevention vs. Intervention CHONG POH WAN 21 JUNE 2011.
Chapter 11: An Evaluation Framework Group 4: Tony Masi, Sam Esswein, Brian Rood, & Chris Troisi.
Part 1-Intro; Part 2- Req; Part 3- Design  Chapter 20 Why evaluate the usability of user interface designs?  Chapter 21 Deciding on what you need to.
Multimedia Specification Design and Production 2013 / Semester 1 / week 9 Lecturer: Dr. Nikos Gazepidis
User Study Evaluation Human-Computer Interaction.
This slide presentation is Confidential and Proprietary to ProtoTest LLC.
Human Computer Interaction
Usability testing. Goals & questions focus on how well users perform tasks with the product. – typical users – doing typical tasks. Comparison of products.
Usability Evaluation June 8, Why do we need to do usability evaluation?
What is Usability? Usability Is a measure of how easy it is to use something: –How easy will the use of the software be for a typical user to understand,
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley The Resonant Interface HCI Foundations for Interaction Design First Edition.
Evaluation approaches Text p Text p
Usability Testing CS774 Human Computer Interaction Spring 2004.
Testing & modeling users. The aims Describe how to do user testing. Discuss the differences between user testing, usability testing and research experiments.
1 ISE 412 Usability Testing Purpose of usability testing:  evaluate users’ experience with the interface  identify specific problems in the interface.
Usability Assessment Methods beyond Testing Chapter 7 Evaluating without users.
Evaluation Techniques Evaluation –tests usability and functionality of system –occurs in laboratory, field and/or in collaboration with users –evaluates.
Copyright  2003 by Dr. Gallimore, Wright State University Department of Biomedical, Industrial Engineering & Human Factors Engineering Human Factors Research.
Determination of Sample Size: A Review of Statistical Theory
Chapter 8 Usability Specification Techniques Hix & Hartson.
AMSc Research Methods Research approach IV: Experimental [1] Jane Reid
Usability Engineering Dr. Dania Bilal IS 582 Spring 2006.
1 Notes from
EVALUATION PROfessional network of Master’s degrees in Informatics as a Second Competence – PROMIS ( TEMPUS FR-TEMPUS-JPCR)
OSU Libraries presented by the u-team.
Usability Engineering Dr. Dania Bilal IS 592 Spring 2005.
Evaluation Methods - Summary. How to chose a method? Stage of study – formative, iterative, summative Pros & cons Metrics – depends on what you want to.
June 5, 2007Mohamad Eid Usability Testing Chapter 8.
Introduction to Evaluation without Users. Where are you at with readings? Should have read –TCUID, Chapter 4 For Next Week –Two Papers on Heuristics from.
Chapter 27 Variations and more complex evaluations.
Fall 2002CS/PSY Predictive Evaluation (Evaluation Without Users) Gathering data about usability of a design by a specified group of users for a particular.
User Interface Evaluation Introduction Lecture #15.
STAT03 - Descriptive statistics (cont.) - variability 1 Descriptive statistics (cont.) - variability Lecturer: Smilen Dimitrov Applied statistics for testing.
Usability Engineering Dr. Dania Bilal IS 582 Spring 2007.
Usability Engineering Dr. Dania Bilal IS 587 Fall 2007.
Evaluation / Usability. ImplementDesignAnalysisEvaluateDevelop ADDIE.
Day 8 Usability testing.
User Interface Evaluation
SIE 515 Design Evaluation Lecture 7.
Human Computer Interaction Lecture 15 Usability Evaluation
Evaluation Techniques 1
Evaluation techniques
HCI Evaluation Techniques
Testing & modeling users
Presentation transcript:

STAT – User testing and evaluation 1 User testing and evaluation Lecturer: Smilen Dimitrov Applied statistics for testing and evaluation – MED4

STAT – User testing and evaluation 2 Introduction We previously discussed measures of central tendency and dispersion, as descriptive statistics parameters, and we did a couple of statistic inferential tests. We will now look at a bit more about usability and user testing, and how statistics is applied there.

STAT – User testing and evaluation 3 Usability testing Historical developments: Taylor study of 'which is the best way to do a job?' and 'what should constitute a day's work' to determine time standards for basic tasks Frank (an Industrial Engineer) and Lillian (a Psychologist) Gobreth, studied the motions involved in bricklaying - reduced motions from 18 to 5, the development of the 'therblig' unit of motion Late 1940, psychologist at Write Patterson Air Force Base studies crashes to determine the cause, it was not what was expected A study on the effect of a redundant, high-centered taillight on rear end car crashes

STAT – User testing and evaluation 4 Usability testing Usability testing is a means for measuring how well people can use some human-made object (such as a web page, a computer interface, a document, or a device) for its intended purpose, i.e. usability testing measures the usability of the object. –Usability testing is the common name for user-based system evaluation. Popularized in the media by Jakob Neilson and usually thought of as related to web site design in the 1990's, Usability is defined as "the ability of a specific group of users to perform a specific set of activities within a specific environment with effectiveness, efficiency, and satisfaction (ISO 9241 standard) Systems are made up of users performing some activity within a context. [We] can't redesign users but we can design equipment, so our goal as designers is to 'design equipment to optimize system performance‘

STAT – User testing and evaluation 5 Usability testing If usability testing uncovers difficulties, such as people having difficulty understanding instructions, manipulating parts, or interpreting feedback, then developers should improve the design and test it again. During usability testing, the aim is to observe people using the product in as realistic a situation as possible, to discover errors and areas of improvement. In respect to usability, we look at the following factors: –Functional Suitability - does the product contain the functionality required by the user? –Ease-of-learning - can the user figure out how to exercise the functionality provided –Ease-of-use - can the user exercise the functionality accurately and efficiently (includes accessibility issues) –Ease-of-recall - can the knowledge of operation be easily maintained over time? –Subjective Preference

STAT – User testing and evaluation 6 Usability testing Four principles of usability –Functional 'Visibility' through an obvious 'visible' structure and adequate feedback –Good (conceptual) model to predict the effects of our actions (examples: Real World Metaphors - Trash Can icon, Shopping Carts) –Design for the Intended user (and not for yourself) –Design for Errors and Slips (or 'Don't Blame the User')

STAT – User testing and evaluation 7 Usability testing Setting up a usability test involves –carefully creating a scenario, or realistic situation, wherein the person performs a list of tasks using the product being tested while observers watch and take notes. –Several other test instruments such as scripted instructions, paper prototypes, and pre- and post-test questionnaires are also used to gather feedback on the product being tested. For example, to test the attachment function of an program, a scenario would describe a situation where a person needs to send an attachment, and ask him or her to undertake this task. The aim is to observe how people function in a realistic manner, so that developers can see problem areas, and what people like. Techniques popularly used to gather data during a usability test include think aloud protocol and eye tracking.

STAT – User testing and evaluation 8 Usability testing Two types of usability testing during design: Non-User Based –Expert Review –Compliance Reviews –Heuristic Evaluations –Cognitive Walkthroughs User-based –User Surveys –Ethnographic Observation –Performance-based –Think Aloud –Co-Discover

STAT – User testing and evaluation 9 User-based usability testing Several types of user based tests: –User surveys, which are inexpensive and can be conducted remotely, [and] can provide trend data –Ethnographic observation can be understood as contextual inquiery, which "places methodological emphasis on the rigorous description of the ways in which situated action is produced everyday –Performance-based testing sets a goal for the users to perform, and measures their performance, such as time taken for completion of the task thus it provides an 'objective' measure - good for comparative evaluations but also it is not a complete picture of usability, possibly misleading –Think-aloud protocols are probably the most common form of usability testing in use today, designed to capture participant's understanding –Co-discovery protocols are variation on Think Out Loud protocol, with a multiple participant perspective

STAT – User testing and evaluation 10 User-based usability testing Subjective assessment tells the evaluator how the users feel about the software being tested. This is distinct from how efficiently or effectively they perform with the software. The usual method of assessment is to used a standardized opinion questionnaire to avoid criticisms of subjectivity. Subjective measures part of user-based evaluation: –Self-reported ease-of-use measures (summative evaluations) SUS QUIS SUMI –Aesthetic value –User preferences Standard questionnaires

STAT – User testing and evaluation 11 User-based usability testing Subjective assessment tells the evaluator how the users feel about the software being tested. This is distinct from how efficiently or effectively they perform with the software. The usual method of assessment is to used a standardized opinion questionnaire to avoid criticisms of subjectivity. Subjective measures part of user-based evaluation: –Self-reported ease-of-use measures (summative evaluations) SUS QUIS SUMI –Aesthetic value –User preferences Standard questionnaires

STAT – User testing and evaluation 12 User-based usability testing Usability metrics or benchmarks - formal measurements that are used as guides to the level of usability of a product. Metrics include: –how fast a user can perform a task, –number of errors made on a task, –learning time, and –subjective ratings. Criterion testing - user testing that measures user performance to determine whether a target performance level has been reached. –A criterion test measures whether the game meets the requirements of the test. –For example: 'starting from the options button, the average player will be able to locate the screen resolution option within 10 seconds'.

STAT – User testing and evaluation 13 User-based usability testing Another division of user-based testing: –Formative evaluation, which is a user test performed during iterative design with the goal of finding usability problems to fix on the next design iteration. The purpose of formative evaluation is finding usability problems in order to fix them in the next design iteration. Formative evaluation doesn't need a full working implementation, but can be done on a variety of prototypes. –Field study - Running a test in a lab environment on tasks of your invention may not tell you enough about how well your interface will work in a real context on real tasks. A field study can answer these questions, –Controlled experiment - goal is to test a quantifiable hypothesis about one or more interfaces. Controlled experiments happen under carefully controlled conditions using carefully-designed tasks - often more carefully chosen than formative evaluation tasks. Hypotheses can only be tested by quantitative measurements of usability, like time elapsed, number of errors, or subjective satisfaction. Tests a hypothesis (e.g., interface X is faster than interface Y) Evaluates working implementation, in controlled lab environment, on chosen tasks Mostly quantitative observations (time, error rate, satisfaction)

STAT – User testing and evaluation 14 Formative evaluation The basic steps: –(1) find some representative users (should be representative of the target user class[es], based on user analysis); –(2) give each user some representative tasks (should be representative of important tasks, based on task analysis); and –(3) watch the user do the tasks. There are three roles in a formative evaluation test: –a user, –a facilitator, and –some observers.

STAT – User testing and evaluation 15 Review Arithmetic mean Median Mode Range Variance Standard deviation Quantiles, Interquartile range Probability distributions – uniform, normal (Gaussian) and T Standard error – inference about unreliability Confidence interval Single sample t-test Two sample F-test and t-test. Measures of Central tendency (location) Measure of Statistical variability (dispersion - spread) Descriptive statistics Inferential statistics