Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Multi-tasking, Memory Failures, and.

Slides:



Advertisements
Similar presentations
Robert L. Helmreich, Ph.D. FRAes The University of Texas
Advertisements

Pilots & Weather Considerations Federal Aviation Administration 1.
Key Dismukes, PhD Chief Scientist for Aerospace Human Factors
The pilot and airline operator’s perspective on runway incursion hazards and mitigation options Session 3 Presentation 1.
Flight Safety Foundation 21 st Annual European Air Safety Seminar Cyprus, March, 2009 The Hidden Complexity of Cockpit Operations Loukia D. Loukopoulos.
HUMAN FACTORS Modified by Lt Colonel Fred Blundell TX-129 Fort Worth Senior Squadron For Local Training Rev Jan-2014.
Downloaded from Approach and Landing Accidents Reducing the Risk Presented by Bryan W. Neville Aviation Safety Inspector.
Incident Review Meeting Guidance Material & Presentation Template
The Importance of Synergy Between Flight Deck and Cabin Crews
Lecture 9: Ground Proximity Warning System (GPWS)
Best Practices for Taxi Operations at Towered & Non-Towered Airports
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 The Limits of Expertise: The Misunderstood.
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 The Limits of Expertise: The Misunderstood.
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Why Do Good Pilots Make Errors That.
Cognetics and the Locus of Attention
User interface design Designing effective interfaces for software systems Objectives To suggest some general design principles for user interface design.
24 October 2002 ICAO NAM/CAR/SAM RUNWAY SAFETY/INCURSION CONFERENCE 1 Retrospective Human Factors Analysis of US Runway Incursions (Focus: Air Traffic.
Executive Briefing This briefing is designed as a stand alone briefing for Airline Senior Executives / CEOs. Minor text amendments following review at.
Why do people make mistakes? Learning Lite
Best Practices for Taxi Operations at Towered & Non-Towered Airports
Key Dismukes and Ben Berman Human Systems Integration Division NASA Ames Research Center 6 March 2007 Human Factors in Aviation Conference Checklists and.
Incident Review Meeting Example  The next slides are an example of how to complete the template and identify latent conditions, threats, errors, UAS and.
V-1 Module V ______________________________________________________ Providing Positive Behavioral Interventions and Supports.
University of Palestine software engineering department Testing of Software Systems Fundamentals of testing instructor: Tasneem Darwish.
Error Management OGHFA 1_HP_06_VIS_Error Management 1.
Human-Centered Information Visualization Jiajie Zhang, Kathy Johnson, Jack Smith University of Texas at Houston Jane Malin NASA Johnson Space Center July.
Attention as a Limited Capacity Resource
The Field Guide to Human Error Investigations- The Old View (Chapters 1 – 6) By Dekker AST 425.
Topic 5 Understanding and learning from error. LEARNING OBJECTIVE Understand the nature of error and how health care can learn from error to improve patient.
Downloaded from 9/14/2015 Aeronautical Decision Making - The Theory Behind the Practice 1 AERONAUTICAL DECISION MAKING The Theory Behind The.
Office of Aviation Safety Pilot Professionalism Human Performance presentation.
1 Runway Incursion Causal Analysis Section 3. 2 PILOTS and VEHICLE OPERATORS must taxi and maneuver their aircraft/vehicles on taxiways and runways in.
Griffith Aerospace Safety Centre Training Interventions for Managing Startle During Unexpected Critical Events Wayne Martin Griffith University Aerospace.
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Human Error or System Error: Are.
CREW RESOURCE MANAGEMENT
Health Profession Education for Patient Safety” Blink or Think? Pat Croskerry MD, PhD The Safety Competencies Enhancing Patient Safety Across the Health.
1 IE 590D Applied Ergonomics Lecture 26 – Ergonomics in Manufacturing & Automation Vincent G. Duffy Associate Prof. School of IE and ABE Thursday April.
Key Dismukes Chief Scientist for Aerospace Human Factors Human Factors Research and Technology Division NASA-Ames Research Center 20 August 2003 ALPA Air.
Key Dismukes Human Factors Research and Technology Division 16 December 2003 Prospective Memory: Remembrance of Things Future.
Hazard Identification
Understanding Users Cognition & Cognitive Frameworks
Applying Error and Threat Management Concepts to Soaring Operations
February 15, 2004 Software Risk Management Copyright © , Dennis J. Frailey, All Rights Reserved Simple Steps for Effective Software Risk Management.
Presented to: By: Date: Federal Aviation Administration North Florida FSDO Runway Safety Program- Pilot Deviations Runway Incursions Instructor and Student.
Runway Incursion Causal Categories OPERATIONAL ERROR (OE) - A human error caused by a tower controller. There are over 8000 tower controllers in the U.S.
Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Multi-tasking, Memory Failures, and.
“ A location on an aerodrome movement area with a history or potential risk of collision or runway incursion, and where heightened attention by pilots/drivers.
Click to edit Master title style Click to edit Master text styles Second level TOOLBOX TALK MANAGING AIRSIDE SAFETY.
Threat and Error Management. What is TEM?  “Defensive flying”
Trainings to Avoid Deviations and Incursions. Trainings to Avoid Deviations and Incursions A Big Training Challenge Looking for a Solution.
Downloaded from HOW TO AVOID A GEAR-UP LANDING Presented By Larry Enlow Aviation Safety Inspector Federal Aviation Administration Orlando.
Human Factors An Overview
Development of Expertise. Expertise We are very good (perhaps even expert) at many things: - driving - reading - writing - talking What are some other.
Effective leadership in airline crews. 2 Good teamwork in aircraft crews (Ginnett) & hospital teams (K-P) Aircraft crewsHospital teams.
Crew Resources Management Lecture 7: Workload Management.
Crew Resources Management
HOW TO AVOID A GEAR-UP LANDING Presented By Larry Enlow Aviation Safety Inspector Federal Aviation Administration Orlando Flight Standards District Office.
Lecture 9: Ground Proximity Warning System (GPWS).
The air traffic controller’s perspective on runway incursion hazards and mitigation options Session 2 Presentation 1.
How Did That Happen? Analyzing Prevention Resources Available.
6. (supplemental) User Interface Design. User Interface Design System users often judge a system by its interface rather than its functionality A poorly.
Complacency Some people have their head in the sand and ignore what is happening around them, what’s changed, what NEW hazards are there…
Office of Aviation Safety Air Traffic Control: Human Performance William J. Bramble, Jr., Ph.D. Senior Human Performance Investigator.
Strategic Team Decision Making Florida Reliability Coordinating
Human Error Definition Broad Classes of Human Errors
The air traffic controller’s perspective on runway incursion hazards and mitigation options Session 2 Presentation 1.
The air traffic controller’s perspective on runway incursion hazards and mitigation options Session 3 Presentation 3.
Situation Monitoring Know the plan, share the plan, review the risks.
Presentation transcript:

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 1 Multi-tasking, Memory Failures, and a Perspective on Human Error Key Dismukes, Ph.D. Chief Scientist for Aerospace Human Factors NASA Ames Research Center FAA Northwest Region Runway Summit July 2008 Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 2 Overview  Most runway incursions are attributed to human error, especially pilot error  Human factors perspective on why highly skilled experts make errors performing routine tasks  Runway incursions occur for same underlying reasons as errors made in other phases of flight  Will focus on one particular type of error as an illustration often contributes to incursions contributed to many other accidents Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 3 Two Representative Accidents LAX, 1991 Tower cleared commuter to position & hold on 24L Delay to cross other aircraft on far end of runway Poor visibility: twilight, haze, & glare from lights Controller forgot commuter not departed or confused with another commuter Cleared B737 to land on 24L Both aircraft destroyed; 34 killed LaGuardia, 1994 Captain inadvertently forgot to turn on pitot heat Pitot probe froze Captain rejected takeoff because of anomalous airspeed indications Aircraft ran off end of runway, destroying aircraft; 30 minor injuries

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 4 Why?  Carelessness?  Lack of skill or proficiency? Why would highly experienced controllers and pilots forget to perform routine tasks?  Overwhelmed by workload?  More subtle issues?

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 5 A NASA Study: The Limits of Expertise Key Dismukes, Ben Berman, and Loukia Loukopoulos  Re-examined all major U.S. airline accidents attributed to crew error  Why might any crew in position of accident crew, and knowing only what accident crew knew, be vulnerable to same errors?  Traced interaction of task demands, equipment features, events, and organizational factors with human cognitive processes Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 6 Understanding Accidents  Highly diverse: each accident is unique in terms of surface features  Countermeasures developed after accident prevent recurrence But would that accident happen again anyway?  Single-point failures rarely cause accidents in airline operations Multiple happenstance factors combine to defeat defenses  To maintain/improve safety must look beyond surface features What underlying features cut across accidents? Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 7 Conditions (e.g., weather) Events Equipment and interface design Task Demands Inherent characteristics and limitations of human perception and cognition Individual Factors: goals technical & interpersonal skills experience and currency physiological state attitudes Organizational/Industry Factors: goals – production vs. safety training policy procedures regulations norms for actual operations Individual / Team Performance

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 8 Two Fallacies About Human Error  MYTH: Experts who make errors performing a familiar task reveal lack of skill, vigilance, or conscientiousness  FACT: Skill, vigilance, and conscientiousness are essential but not sufficient to prevent error     MYTH: If experts can normally perform a task without difficulty, they should always be able to perform that task correctly  FACT: Exerts periodically make errors as consequence of subtle variations in task demands, information available, and cognitive processing Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 9 Some Argue Solution to Human Error is Automation  This perspective ignores the nature of work of pilots and controllers  Humans do what computers cannot: Interpret incomplete or ambiguous information Consider implications Make appropriate value judgments and decisions Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 10 A Modern System Safety Perspective  Experts make errors for three kinds of reasons: Lack of information to be certain of outcome of competing choices Task demands and equipment not well matched to human information processing characteristics Competing organizational goals must be balanced (e.g., production vs. safety)  Mathematically impossible to simultaneously maximize two or more variables  Accidents result from interactions among components of complex systems that are not anticipated and controlled Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 11 Confluence of Factors in a CFIT Accident (Bradley, 1995) Non-precision approach ≥ 250 foot terrain clearance Are most pilots aware of this? Weather conditions Airline’s use of QFE altimetry Strong crosswind Autopilot would not hold PF selected Heading Select Additional workload Increased vulnerability to error Crew error (70 feet) in altimeter setting Altimeter update not available 170 foot error in altimeter reading Tower closed Tower window broke Rapid change in barometric pressure Approach controller failed to update altimeter setting Altitude Hold may allow altitude sag 130 feet in turbulence PF used Altitude Hold to capture MDA PM used non-standard callouts to alert PF Training & Standardization issues? Aircraft struck trees 310 feet below MDA ? ? Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 12 To Prevent Runway Incursions…  Start with thorough analysis of large volume of incidents from operational and human factors perspective  See work by Kim Cardosi and colleagues at Volpe Transportation Center  My research examines two issues identified by Cardosi, et al.: Human Systems Integration Division Multi-tasking Prospective memory failures

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 13 Multi-tasking and Prospective Memory  Multi-tasking demands are heavy in work of controllers and pilots  Prospective memory: Individual must remember to perform a task that cannot be performed when the intention to act is formed Cardosi: forgetting was most common form of controller error Did not forget content (e.g., call sign) Forgot to act or forgot implications of current situation for what they should or should not do later  Examples: LAX, 1991; LaGuardia, 1994 Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 14 WHY???  Importance of task does not protect against forgetting at crucial moments Surgical teams forget to remove instruments Parents forget infants sleeping in the back seat of the car  Workload? Sometimes high but, more often than not, normal LAX and LaGuardia accidents were within typical workload range Human Systems Integration Division How could experienced operators forget to perform simple tasks with monumental consequences?

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 15 Our Research: Forget to Act in Six Prototypical Situations  Non-habitual tasks that must be deferred e.g., “Report passing through 10,000 feet” Human Systems Integration Division  Interruptions e.g., Controller interrupted before turning aircraft onto final  Attention switching among multiple concurrent tasks e.g., First officer re-programming the FMC during taxi  Habitual tasks with normal trigger cues removed e.g., “Go to tower at final approach fix”  Habitual tasks performed out of the normal sequence e.g., Setting flaps delayed because of slush on taxiway  Habit capture (atypical action must be substituted for habitual action) e.g., Modified standard instrument departure

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 16 A Quick Sketch of the Science  Six prototypical situations appear diverse but share underlying features: Pilots or controllers were juggling multiple tasks concurrently Had to remember to perform deferred task or perform task out of normal sequence Human Systems Integration Division  Individuals forget to act because characteristics of these situations interact with the way the human brain processes information

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 17 Human Brain: Two Ways to Process Information  “Controlled” processing (fully attentive): Required when learning new tasks, performing tasks with novel aspects and tasks that are unusually dangerous or difficult, and when solving problems Corresponds roughly to conscious awareness Slow, serial, effortful (narrow-bandwidth, low capacity)  Automatic processing: Takes over as we master specific task Fast, high-capacity, requires minimal conscious supervision Essential for much of experts’ work Drawback: powerful but dumb; unreliable in certain situations Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 18 Our Research: Forget to Act in Six Prototypical Situations  Non-habitual tasks that must be deferred e.g., “Report passing through 10,000 feet” Human Systems Integration Division  Interruptions e.g., Controller interrupted before turning aircraft onto final  Attention switching among multiple concurrent tasks e.g., First officer re-programming the FMC during taxi  Habitual tasks with normal trigger cues removed e.g., “Go to tower at final approach fix”  Habitual tasks performed out of the normal sequence e.g., Setting flaps delayed because of slush on taxiway  Habit capture (atypical action must be substituted for habitual action) e.g., Modified standard instrument departure

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 19 Non-Habitual Situations: An Everyday Example  Deferred intention: Pick up milk on way home from work  Intention cannot be held in conscious awareness throughout the day – moves to memory  How is intention retrieved from memory back into awareness? Requires noticing salient cue to remind of intention (get milk) Cue must occur at time action required (driving home)  What makes a good reminder cue? Placed where it will be noted when needed (e.g., on car dashboard) Clearly related to deferred intention (e.g., empty milk carton = get milk) Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 20 Non-habitual Situations: A Cockpit Example  “Report passing through 10,000 feet”  Crew must attend to other tasks for several minutes before reaching 10,000 feet Intention to report moves from conscious awareness to memory Cannot monitor altimeter continuously  Crew may notice altimeter during scan, but it is a mediocre cue Altimeter associated with many items in memory, not just reporting 10,000 feet Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 21 Non-habitual Tasks  With extensive repetition, no longer have to think what to do next – automatic e.g., complete After Start checklist call for flaps  Most tasks consist of a series of subtasks  Normally highly reliable but vulnerable if cueing is disrupted Subtasks performed out of sequence (e.g., deferred subtask) Interruptions (e.g., controller distracted by emergency) Human Systems Integration Division A B C D… environmental cues

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 22 Other Factors Affecting Vulnerability to Forgetting to Act  Organizational / industry factors e.g., rushing to make slot time; trying to beat T-storms to airport  Design of procedures e.g., running checklists when both pilots should be heads-up  Recent study: The Myth of Multi-tasking: Managing Complexity in Real-World Operations (Loukopoulos, Dismukes, & Barshi) Flight operations manuals present idealized picture: tasks are linear, predictable, and under moment-to-moment control of the crew Reality: execution of procedures is frequently perturbed by interruptions, situations change dynamically, and tasks must be performed concurrently Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 23 Practical Countermeasures to Reduce Error Vulnerability 1) Discard “blame and punish” mentality when experts make mistakes 2) Periodically analyze SOPs to indentify aspects that contribute to vulnerability 3) Use training to explain why expert pilots and controllers are vulnerable to error Evaluate and share personal techniques to reduce vulnerability to error 4) Treat monitoring as essential rather than secondary task 5) Don’t underestimate subtle effects of fatigue on cognitive performance 6) Do the research! Procedures, training, and equipment design must be based on science Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 24 Additional Information  Can download papers from:  Dismukes, Berman, & Loukopoulos (2007). The Limits of Expertise: Rethinking Pilot Error and the Causes of Airline Accidents (Ashgate Publishing)  Loukopoulos, Dismukes, & Barshi (in press). The Myth of Multitasking: Managing Complexity in Real-World Operations (Ashgate Publishing)  Cardosi (2001). Runway Safety: It’s Everybody’s Business. DOT/FAA/AR-01/66  Cardosi & Yost (2001). Controller and Pilot Error in Airport Operations. DOT/FAA/AR-00/51  DiFiore & Cardosi (2006). Human Factors in Airport Surface Incidents. DOT/FAA/AR-06/5 Human Systems Integration Division

Click to edit Master title style Click to edit Master text styles Second level Third level Fourth level Fifth level 25