The Effects of Interface Design on Telephone Dialing Performance Masters thesis in Computer Science Andrew R. Freed 4/30/2003.

Slides:



Advertisements
Similar presentations
Recuperação de Informação B Cap. 10: User Interfaces and Visualization , , , November 29, 1999.
Advertisements

Chapter 15: Analytical evaluation
Pat Langley Computational Learning Laboratory Center for the Study of Language and Information Stanford University, Stanford, California
Introduction to Computer Input Devices and Their Evaluation Shumin Zhai IBM Almaden Research Center.
The world leader in serving science TQ ANALYST SOFTWARE Putting your applications on target.
Making sense out of recorded user-system interaction Dr Willem-Paul Brinkman Lecturer Department of Information Systems and Computing Brunel University.
User Interfaces 4 BTECH: IT WIKI PAGE:
G063 - The Model Human Processor. Learning Objective: describe the user interface designers tool known as the ‘Model Human Processor', describe how the.
QUASID – Measuring Interaction Techniques Karin Nieuwenhuizen.
Copyright 2008 by User Interface Lab Industrial Engineering Dept. of Industrial Systems & Information Engineering Korea University Serial Modules in Parallel.
TAUCHI – Tampere Unit for Computer-Human Interaction Manual Text Entry: Experiments, Models, and Systems Poika Isokoski Tampere Unit for Computer-Human.
Korea Univ. Division Information Management Engineering UI Lab. Korea Univ. Division Information Management Engineering UI Lab. Human Interface PERCEPTUAL-MOTOR.
Tracking multiple independent targets: Evidence for a parallel tracking mechanism Zenon Pylyshyn and Ron Storm presented by Nick Howe.
Topic: Fitts' Law Lawrence Fyfe CPSC 681. Fitts' Law Formula: ID (index of difficulty) = log 2 (D/W +1) Soukoreff R.W., MacKenzie I.S., Towards.
Component-specific usability testing Dr Willem-Paul Brinkman Lecturer Department of Information Systems and Computing Brunel University
People & Devices: (Inputs & Outputs) Startlingly small child using computer History of human-computer interaction Another history video.
Models of Human Performance Dr. Chris Baber. 2 Objectives Introduce theory-based models for predicting human performance Introduce competence-based models.
Objectives Define predictive and descriptive models and explain why they are useful. Describe Fitts’ Law and explain its implications for interface design.
Predictive Evaluation Predicting performance. Predictive Models Translate empirical evidence into theories and models that can influence design. Performance.
Some questions of hypermedia and CHI Josep Blat Universitat Pompeu Fabra.
Research Methods for HCI: Cognitive Modelling BCS HCI Tutorial 1 st September, 2008.
Predictive Evaluation Simple models of human performance.
Discussion Silvia Lindtner INF 132 April 07. Fitts’ law - recap A predictive model of time to point at an object Help decide the location and size of.
User Interface Design Chapter 11. Objectives  Understand several fundamental user interface (UI) design principles.  Understand the process of UI design.
Human Factors for Input Devices CSE 510 Richard Anderson Ken Fishkin.
Division of Information Management Engineering User Interface Laboratory Model-Based Evaluation of Expert Cell Phone Menu Interaction ROBERT ST. AMANT.
Frank E. Ritter Urmila Kukreja Robert St. Amant 1 Including a Model of Visual Processing With a Cognitive Architecture to Model a Simple Teleoperation.
Engineering Design Centre Project Updates –August 2014.
Chapter 5 Models and theories 1. Cognitive modeling If we can build a model of how a user works, then we can predict how s/he will interact with the interface.
Evaluation of digital collections' user interfaces Radovan Vrana Faculty of Humanities and Social Sciences Zagreb, Croatia
User Models Predicting a user’s behaviour. Fitts’ Law.
Eye Tracking in the Design and Evaluation of Digital Libraries
Evaluation of Adaptive Web Sites 3954 Doctoral Seminar 1 Evaluation of Adaptive Web Sites Elizabeth LaRue by.
UNDERSTANDING USERS: MODELING TASKS AND LOW- LEVEL INTERACTION Human-Computer Interaction
Slides based on those by Paul Cairns, York ( users.cs.york.ac.uk/~pcairns/) + ID3 book slides + slides from: courses.ischool.berkeley.edu/i213/s08/lectures/i ppthttp://www-
1 Brief Review of Research Model / Hypothesis. 2 Research is Argument.
Ch 14. Testing & modeling users
Gary MarsdenSlide 1University of Cape Town Case Study - Nokia 5110 We will try to put together what we have learnt to date by looking at a cell- phone,
Keystroke-Level Model IST 331 Gaurav Dubey Based on ‘The ABCs of Users’, Ritter et al 2011.
Testing & modeling users. The aims Describe how to do user testing. Discuss the differences between user testing, usability testing and research experiments.
Identifying needs and establishing requirements
GOMs and Action Analysis and more. 1.GOMS 2.Action Analysis.
Human Factors Issues Chapter 9. Human Factors = ergonomics WWII based – military significance… … a necessary part of medical device design…
Copyright © 2008 Pearson Education, Inc. Publishing as Pearson Addison-Wesley Prof Jim Warren with reference to sections 7.4 and 7.6 of The Resonant Interface.
Analytical evaluation Prepared by Dr. Nor Azman Ismail Department of Computer Graphics and Multimedia Faculty of Computer Science & Information System.
Modeling Visual Search Time for Soft Keyboards Lecture #14.
Visual Thinking and the Objects of Visual Attention Colin Ware University of New Hampshire.
ITM 734 Introduction to Human Factors in Information Systems
The Psychology of Human-Computer Interaction
1 A Learning Model of a Long, Non-iterative Spreadsheet Task Frank E. Ritter, Jong W. Kim, and Jaehyon Paik College of IST, Penn State Presented at the.
Evaluation Using Modeling. Testing Methods Same as Formative Surveys/questionnaires Interviews Observation Documentation Automatic data recording/tracking.
Technology and Aging Eileen Wood. Why should we be talking about computers and aging? Social connections Independence Cognitive Skills.
1 Cognitive Modeling GOMS, Keystroke Model Getting some details right!
Cognitive Models Lecture # March, 2008Human Computer Intercation Spring 2008, Lecture #10 2 Agenda Cognitive models –KLM –GOMS –Fitt’s Law –Applications.
Dynamic Decision Making Laboratory Carnegie Mellon University 1 Social and Decision Sciences Department ACT-R models of training Cleotilde Gonzalez and.
Chapter 15: Analytical evaluation. Aims: Describe inspection methods. Show how heuristic evaluation can be adapted to evaluate different products. Explain.
Korea University User Interface Lab Copyright 2008 by User Interface Lab Human Action Laws in Electronic Virtual Worlds – An Empirical Study of Path Steering.
1 1 ITM 734 Introduction to Human Factors in Information Systems Cindy Corritore This material has been developed by Georgia Tech HCI.
6.S196 / PPAT: Principles and Practice of Assistive Technology Wed, 19 Sept Prof. Rob Miller Today: User-Centered Design [C&H Ch. 4]
GOMS as a Simulation of Cognition Frank Ritter, Olivier Georgeon 28 oct 2014.
Cognitive Modeling Cogs 4961, Cogs 6967 Psyc 4510 CSCI 4960 Mike Schoelles
Human Computer Interaction
CSc4730/6730 Scientific Visualization
Identifying Confusion from Eye-Tracking Data
Bryan Stearns University of Michigan Soar Workshop - May 2018
GOMS as a Simulation of Cognition
Human Computer Interaction
Fitts’s Law Incredibly professional presentation by Thomas Gin, someone please hire me.
Testing & modeling users
Presentation transcript:

The Effects of Interface Design on Telephone Dialing Performance Masters thesis in Computer Science Andrew R. Freed 4/30/2003

The Effects of Interface Design on Telephone Dialing Performance n Towards automatic interface evaluation n Methods of evaluation n Experiment design n Three analyses n Comparison of analyses n Further work

Towards automatic interface evaluation n Why not test with actual users instead? n It takes too much time and money! n Automatic evaluation has been useful in the past (Project Ernestine - Gray et al 1992) to the tune of $2.4M savings/year n Several proposed tools will make this type of evaluation easier

Towards automatic interface evaluation n Motivation: –Eye-tracking studies by Byrne (1999, 2001) and Hornof (1997) –Cognitive models as surrogate users (Ritter 2001)

Towards automatic interface evaluation n 100 phones to choose from n Selected 10 for analysis

Towards automatic interface evaluation n 10 tasks (Ritter 2000) – 1. Call home (*) – 2. Call work (*) – 3. Redial last number (*) – 4. Call directory inquiries – 5. Call mother (*) – 6. Conference call work and home (*) – 7. Conference call work (flash) then home – 8. Forward call to another number (*) – 9. Forward call (flash) to another number –10. Hang up

Towards automatic interface evaluation n 10 telephone numbers – – – – – n and 3 other tasks –Forward, redial, conference call

Methods of evaluation n Possible tools n Cognitive architectures n ACT-R/PM n Generic Simulated Eyes and Hands n Focused analysis methods

Possible tools n Ivorys tools to evaluate websites (2001) n Apex (M. Freed 1998) and iGen (Emmerson 2000) model complex tasks n Glean (Kieras et al 1995) evaluates Lisp interfaces n Shortcomings: no learning, no visual search, tied to a specific interface format, no cognitive theory

Cognitive architectures n Unified theory of cognition (Newell 1990) n Simulate human behavior n Perceptual and motor capability (simulated eyes and hands) n Can do visual search, click buttons, sometimes learn

Cognitive architectures (examples) n EPIC (Kieras and Meyer 1997) - has visual search and perceptual/motor skills… but only evaluates Common Lisp interfaces n Soar (Newell 1990) - also has visual search, perceptual motor skills, plus learning… but only evaluates Tcl/Tk interfaces (or requires a socket connection) n ACT-R/PM (Anderson & Lebiere 1998, Byrne 2001) - nearly identical benefits and limitations as EPIC, plus has learning

ACT-R/PM n Why did we choose ACT-R/PM? n Well-accepted cognitive architecture n Used in past to evaluate interfaces n Can overcome the Lisp interface-only problem with generic eyes and hands

Generic Simulated Eyes and Hands n Segman (St. Amant & Riedl 2001) can parse a Windows screen capture and determine the interface components n Can use interfaces written in Lisp, Tcl/Tk, HTML, Visual C++,... n Segman can be connected to ACT- R/PM

Focus of analysis n A - Analytical model (Fitts Law) n B - Cognitive model (ACT-R/PM) n C - Human data

General experiment design n Analytical model, cognitive model, and human users interact with same interfaces n Analytical model dials each number once on each phone, does not do other tasks n Cognitive model: Dialed each phone number 50 times on each phone, performed other phone tasks 50 times on each phone. n Human users (N=9): Dialed each phone number on each phone, performed other phone tasks once on each phone

n Experimental software General experiment design

n Cognitive model and users –Timing and mouse-click logging –Eye-tracking –Users can control pace of trials, model does not care n Analytical model –Does not need to see telephones –Mathematical formula with pixel-level input yields reaction times

A. Fitts Law analysis n What is Fitts Law? n Numerical analysis n Simple conclusions and problems

What is Fitts Law? n Fitts Law (two possible forms): –MT = a + b * LOG 2 (2 * D/W) (Fitts 1954) –MT = max(t m, k * LOG 2 [0.5 + D/W]) (Card et al, 1983) n MT is mouse movement time n D is distance to target, W is target width n a, b, k are constants n t m is minimum movement time

Numerical analysis n Collected pixel-level input about telephones (size and location of buttons) n Dialing a phone requires 10 movements n Total the times from the 10 movements and a base dialing time is established (with no visual search!)

Numerical analysis n Validating our choice of sample telephone numbers (R 2 = 0.96)

Simple conclusions and problems n Fitts Law analysis is fast (it is just an equation!) n Does not consider many factors n Not affected by any aspect of interface design other than button sizing and spacing

B. ACT-R/PM model analysis n Description of model n Visual search predictions n ACT-R/PM makes different reaction time conclusions

Description of ACT-R/PM model n Model has three main components that can operate in parallel: –retrieve a phone digit from memory –visually search for the digit –move the mouse/click on a digit (governed by Fitts Law) n Composed of 71 production rules (mostly for visual search)

Description of ACT-R/PM model n Visual search strategy: random or systematic n One production for random search n Find-random-target IFthe goal is to find a phone target THENfind a visual object of type text which has not been attended lately

Description of ACT-R/PM model n Sixty productions for systematic search n Systematic-search-from-target IFa digit x is in the visual buffer ANDthe goal is to find a target y ANDy is in direction z from x THENfind a visual object of type text in direction z from target x which is within the bounds of the keypad

Visual search predictions n Count fixations and note fixation locations n Search for the keypad is random n Search within the keypad is systematic n The telephones do not generally require a statistically significant different number of fixations to dial (about 16) n (The telephone numbers are significantly different)

Visual search predictions n Model trace

Visual search predictions Phone 4 Phone 9 Whats wrong with this picture?

Visual search predictions n Two phones are predicted to have abnormally long visual searches n These phones require approximately sixty fixations (average on others was sixteen) n Phone 4 has an upside-down keypad -- the systematic search fails! n Phone 9 contains extra information on the buttons… distracts the visual search n We will see the model takes much longer than humans to dial these phones

ACT-R/PM makes different reaction time conclusions n This is no surprise - more factors are being considered n Phones 4 and 9 pay a large visual search penalty n Fitts Law still a factor - phones with Fitts Law violations still perform worse

ACT-R/PM makes different reaction time conclusions

n The phones are often shown to have different dialing times (T-test, p<.05) n The significance level of the differences depends on the telephone number being dialed n On average, approximately 8.7 seconds to dial a telephone. n Never faster than six seconds n No errors!

ACT-R/PM makes different reaction time conclusions n Model is able to perform additional tasks (redial, forward, conference) with a random search n Model does not always succeed but never gives up n Will attend the same visual target several times

C. User data analysis n Where and how users look (eye- tracking) n Humans make errors n Summary of user reaction times

Where and how users look n Fast random search for keypad n Systematic search within keypad

Where and how users look n User trace

Where and how users look n Users require approximately the same number of fixations per telephone as the model did (also true for telephone numbers) n User able to cope with phones 4 and 9 by changing search strategy –Phone 4: Up is down, down is up –Phone 9: Ignore ABCs on the keypad

Where and how users look n Fixation comparison across numbers (R 2 = 0.11)

Where and how users look n Fixation comparison across 8 phones (R 2 = 0.34)

Humans make errors n Errors not predicted by the automatic analyses n Depend on several factors –Number being dialed –Dialing speed (weak correlation) –Interface being used

Errors dependent on interface n Most errors on Fitts Law violators n Least errors when large and adjacent buttons n Users will move mouse while clicking (ACT-R/PM will not), this can cause errors n Possible to estimate number of errors with Fitts index of difficulty?

Summary of reaction times n User on average more than one second faster than model n This probably due to efficient pipelining of motor tasks (room for ACT-R/PM improvement) n Users can dial as fast as 3.5 seconds (average is seven seconds)

Summary of reaction times n Model (R 2 = 0.41), Fitts (R 2 = 0.85), user dial time across phones

Summary of reaction times n Users can do other phone tasks faster than ACT-R/PM n Users can find the target under varied conditions n Users try more strategies to find target n Users will give up if they cant succeed!

Summary of reaction times n Model vs user on extra tasks (R 2 = 0.60, 0.26, 0.11)

Summary of reaction times n User data also shows that the interfaces are often significantly different (p <.05), though less often than the model says n User time differences also depend on the number being dialed n Theory: users less affected by additional interface objects than ACT- R/PM

Comparison of analyses n Analytical model is not enough n Visual search differences between ACT- R/PM and users n ACT-R/PM and Segman need better representation of interfaces n Cognitive models can make more complicated predictions n ACT-R/PM model is generally slower than users

Further work n Cellular phones –This analysis does not work out of the box for cellular phones –These phones have different tasks! (Golightly 2003) n Hutchinson 3G UK phone task (Golightly 2003) –Analysis of menu controls for cellular phone menus, included analytical model –Interface became easier to use when more directional controls were provided

Further work n Analyzing ten additional designs –Easy if you use existing automatic models! Fifteen minutes for Fitts Law analysis Forty-five minutes for 500 model runs –Hard if you test with actual users! Can take weeks to get scheduled Humans miss appointments

Further work n This analysis is generalizable –The same procedures and techniques can be done with other types of interfaces –Automatic models provide fast, easy analysis that mirrors human performance –Must do task analysis first, otherwise you will test for wrong tasks –The hard work (Fitts Law, ACT-R/PM, Segman) has already been done –Cognitive models are available freely as open source

Thank you! n Any questions?

Why is this Computer Science? n Interfaces affect how computers are used (Project Ernestine) n Cognitive modeling is an inter- disciplinary effort n Automatic analysis similar to SPICE n Analysis of visual search algorithms –Random search: O(10*n) –Systematic search: O(10+n >0,<1 )

References n Anderson, J. R., & Lebiere, C. (1998). The atomic components of thought. Mahwah, NJ: Lawrence Erlbaum. n Byrne, M. D. (1999). ACT-R Perceptual-Motor (ACT-R/PM) version 1.0b5: A users manual. Houston, TX: Psychology Department, Rice University. n Byrne, M. D. (2001). ACT-R/PM and menu selection: Applying a cognitive architecture to HCI. International Journal of Human-Computer Studies, 55, n Card, S., Moran, T., & Newell, A. (1983). The psychology of human-computer interaction. Hillsdale, NJ: Lawrence Erlbaum Associates, Inc. n Emmerson, P. (2000). Review of iGEN software. Ergonomics in Design, n Fitts, P. M. (1954). The information capacity of the human motor system in controlling the amplitude of movement. Journal of Experimental Psychology, 47, n Freed, M. A. (1998). Simulating performance in complex, dynamic environments. Northwestern, Evanston, IL. n Golightly, D. (2003). Personal communication. n Gray, W. D., John, B. E., & Atwood, M. E. (1992). The precis of Project Ernestine or An overview of a validation of GOMS. Proceedings of the CHI92 Conference on Human Factors in Computer Systems. n Hornof, A. J., & Kieras, D. E. (1997). Cognitive modeling reveals menu search is both random and systematic. Proceedings of the CHI97 Conference on Human Factors in Computer Systems, New York, NY.

References n Ivory, M. Y., & Hearst, M. A. (2001). The state of the art in automating usability evaluation of user interfaces. ACM Computing Surveys, 33(4), n Kieras, D. E., & Meyer, D. E. (1997). An overview of the EPIC architecture for cognition and performance with application to human-computer interaction. Human- Computer Interaction, 12, n Kieras, D. E., Wood, S. D., Abotel, K., & Hornof, A. (1995). GLEAN: A computer- based tool for rapid GOMS model usability evaluation of user interface designs. Proceedings of the ACM Symposium on User Interface Software and Technology (UIST'95), New York, NY. n Newell, A. (1990). Unified theories of cognition. Cambridge, MA: Harvard University Press. n Ritter, F. E. (2000). A role for cognitive architectures: Guiding user interface design. Seventh Annual ACT-R Workshop, Department of Psychology, Carnegie-Mellon University. n Ritter, F. E., & Young, R. M. (2001). Embodied models as simulated users: Introduction to this special issue on using cognitive models to improve interface design. International Journal of Human-Computer Studies, 55, n St. Amant, R., & Riedl, M. O. (2001). A perception/action substrate for cognitive modeling in HCI. International Journal of Human-Computer Studies, 55,