BMJ Intended for Health Professionals
Home Help Search/Archive Feedback Table of Contents

PDF of this article
Email this article to a friend
Respond to this article
Read responses to this article
Related letters in BMJ
Other related articles in BMJ
PubMed citation
Related articles in PubMed
Download to Citation Manager
This article has been cited by other articles
Search Medline for articles by:
Helmreich, R. L
Alert me when:
New articles cite this article
Collections under which this article appears:
Quality improvement (including CQI and TQM)
BMJ 2000;320:781-785 ( 18 March )

Education and debate

On error management: lessons from aviation

Robert L Helmreich, professor of psychology

Department of Psychology, University of Texas at Austin, Austin, TX 78712, USA

helmreich@psy.utexas.edu

Pilots and doctors operate in complex environments where teams interact with technology. In both domains, risk varies from low to high with threats coming from a variety of sources in the environment. Safety is paramount for both professions, but cost issues can influence the commitment of resources for safety efforts. Aircraft accidents are infrequent, highly visible, and often involve massive loss of life, resulting in exhaustive investigation into causal factors, public reports, and remedial action. Research by the National Aeronautics and Space Administration into aviation accidents has found that 70% involve human error.1

In contrast, medical adverse events happen to individual patients and seldom receive national publicity. More importantly, there is no standardised method of investigation, documentation, and dissemination. The US Institute of Medicine estimates that each year between 44 000 and 98 000 people die as a result of medical errors. When error is suspected, litigation and new regulations are threats in both medicine and aviation.


Summary points


In aviation, accidents are usually highly visible, and as a result aviation has developed standardised methods of investigating, documenting, and disseminating errors and their lessons

Although operating theatres are not cockpits, medicine could learn from aviation

Observation of flights in operation has identified failures of compliance, communication, procedures, proficiency, and decision making in contributing to errors

Surveys in operating theatres have confirmed that pilots and doctors have common interpersonal problem areas and similarities in professional culture

Accepting the inevitability of error and the importance of reliable data on error and its management will allow systematic efforts to reduce the frequency and severity of adverse events

Error results from physiological and psychological limitations of humans.2 Causes of error include fatigue, workload, and fear as well as cognitive overload, poor interpersonal communications, imperfect information processing, and flawed decision making.3 In both aviation and medicine, teamwork is required, and team error can be defined as action or inaction leading to deviation from team or organisational intentions. Aviation increasingly uses error management strategies to improve safety. Error management is based on understanding the nature and extent of error, changing the conditions that induce error, determining behaviours that prevent or mitigate error, and training personnel in their use.4 Though recognising that operating theatres are not cockpits, I describe approaches that may help improve patient safety.

    Data requirements for error management

Multiple sources of data are essential in assessing aviation safety. Confidential surveys of pilots and other crew members provide insights into perceptions of organisational commitment to safety, appropriate teamwork and leadership, and error.3 Examples of survey results can clarify their importance. Attitudes about the appropriateness of juniors speaking up when problems are observed and leaders soliciting and accepting inputs help define the safety climate. Attitudes about the flying job and personal capabilities define pilots' professional culture. Overwhelmingly, pilots like their work and are proud of their profession. However, their professional culture shows a negative component in denying personal vulnerability. Most of the 30 000 pilots surveyed report that their decision making is as good in emergencies as under normal conditions, that they can leave behind personal problems, and that they perform effectively when fatigued.2 Such inaccurate self perceptions can lead to overconfidence in difficult situations.

A second data source consists of non-punitive incident reporting systems. These provide insights about conditions that induce errors and the errors that result. The United States, Britain, and other countries have national aviation incident reporting systems that remove identifying information about organisations and respondents and allow data to be shared. In the United States, aviation safety action programmes permit pilots to report incidents to their own companies without fear of reprisal, allowing immediate corrective action.5 Because incident reports are voluntary, however, they don't provide data on base rates of risk and error.


Sources of threat and types of error observed during line operations safety audit

Sources of threat

Terrain (mountains, buildings)---58% of flights

Adverse weather---28% of flights

Aircraft malfunctions---15% of flights

Unusual air traffic commands---11% of flights

External errors (air traffic control, maintenance, cabin, dispatch, and ground crew)---8% of flights

Operational pressures---8% of flights

Types of error---with examples

Violation (conscious failure to adhere to procedures or regulations)---performing a checklist from memory

Procedural (followed procedures with wrong execution)---wrong entry into flight management computer

Communications (missing or wrong information exchange or misinterpretation)---misunderstood altitude clearance

Proficiency (error due to lack of knowledge or skill)---inability to progam automation

Decision (decision that unnecessarily increases risk)---unnecessary navigation through adverse weather

A third data source has been under development over 15 years by our project (www.psy.utexas.edu/psy/helmreich/nasaut.htm). It is an observational methodology, the line operations safety audit (LOSA), which uses expert observers in the cockpit during normal flights to record threats to safety, errors and their management, and behaviours identified as critical in preventing accidents. Confidential data have been collected on more than 3500 domestic and international airline flights---an approach supported by the Federal Aviation Administration and the International Civil Aviation Organisation.6

The results of the line operations safety audit confirm that threat and error are ubiquitous in the aviation environment, with an average of two threats and two errors observed per flight.7 The box shows the major sources of threat observed and the five categories of error empirically identified; fig 1 shows the relative frequency of each category. This error classification is useful because different interventions are required to mitigate different types of error.



View larger version (17K):
[in this window]
[in a new window]
 
Fig 1.    Percentage of each type of error and proportion classified as consequential (resulting in undesired aircraft states)

Proficiency errors suggest the need for technical training, whereas communications and decision errors call for team training. Procedural errors may result from human limitations or from inadequate procedures that need to be changed. Violations can stem from a culture of non-compliance, perceptions of invulnerability, or poor procedures. That more than half of observed errors were violations was unexpected. This lack of compliance is a source of concern that has triggered internal reviews of procedures and organisational cultures. Figure 1 also shows the percentage of errors that were classified as consequential---that is, those errors resulting in undesired aircraft states such as near misses, navigational deviation, or other error. Although the percentage of proficiency and decision errors is low, they have a higher probability of being consequential.7 Even non-consequential errors increase risk: teams that violate procedures are 1.4 times more likely to commit other types of errors.8

    Managing error in aviation

Given the ubiquity of threat and error, the key to safety is their effective management. One safety effort is training known as crew resource management (CRM).4 This represents a major change in training, which had previously dealt with only the technical aspects of flying. It considers human performance limiters (such as fatigue and stress) and the nature of human error, and it defines behaviours that are countermeasures to error, such as leadership, briefings, monitoring and cross checking, decision making, and review and modification of plans. Crew resource management is now required for flight crews worldwide, and data support its effectiveness in changing attitudes and behaviour and in enhancing safety.9

Simulation also plays an important role in crew resource management training. Sophisticated simulators allow full crews to practice dealing with error inducing situations without jeopardy and to receive feedback on both their individual and team performance. Two important conclusions emerge from evaluations of crew resource management training: firstly, such training needs to be ongoing, because in the absence of recurrent training and reinforcement, attitudes and practices decay; and secondly, it needs to be tailored to conditions and experience within organisations.7

Understanding how threat and error and their management interact to determine outcomes is critical to safety efforts. To this end, a model has been developed that facilitates analyses both of causes of mishaps and of the effectiveness of avoidance and mitigation strategies. A model should capture the treatment context, including the types of errors, and classify the processes of managing threat and error. Application of the model shows that there is seldom a single cause, but instead a concatenation of contributing factors. The greatest value of analyses using the model is in uncovering latent threats that can induce error.10



View larger version (26K):
[in this window]
[in a new window]
 
Fig 2.    Threat and error model, University of Texas human factors research project

By latent threats we mean existing conditions that may interact with ongoing activities to precipitate error. For example, analysis of a Canadian crash caused by a take-off with wing icing uncovered 10 latent factors, including aircraft design, inadequate oversight by the government, and organisational characteristics including management disregard for de-icing and inadequate maintenance and training.3 Until this post-accident analysis, these risks and threats were mostly hidden. Since accidents occur so infrequently, an examination of threat and error under routine conditions can yield rich data for improving safety margins.

    Applications to medical error

Discussion of applications to medical error will centre on the operating theatre, in which I have some experience as an observer and in which our project has collected observational data. This is a milieu more complex than the cockpit, with differing specialties interacting to treat a patient whose condition and response may have unknown characteristics.11 Aircraft tend to be more predictable than patients.

Though there are legal and cultural barriers to the disclosure of error, aviation's methodologies can be used to gain essential data and to develop comparable interventions. The project team has used both survey and observational methods with operating theatre staff. In observing operations, we noted instances of suboptimal teamwork and communications paralleling those found in the cockpit. Behaviours seen in a European hospital are shown in the box, with examples of negative impact on patients. These are behaviours addressed in crew resource management training.


Behaviours that increase risk to patients in operating theatres

Communication:

  • Failure to inform team of patient's problem---for example, surgeon fails to inform anaesthetist of use of drug before blood pressure is seriously affected
  • Failure to discuss alternative procedures

Leadership:

  • Failure to establish leadership for operating room team

Interpersonal relations, conflict:

  • Overt hostility and frustration---for example, patient deteriorates while surgeon and anaesthetist are in conflict over whether to terminate surgery after pneumothorax

Preparation, planning, vigilance:

  • Failure to plan for contingencies in treatment plan
  • Failure to monitor situation and other team's activities---for example, distracted anaesthetist fails to note drop in blood pressure after monitor's power fails

In addition to these observations, surveys confirm that pilots and doctors have common interpersonal problem areas and similarities in professional culture. 2 12 In response to an open ended query about what is most needed to improve safety and efficiency in the operating theatre, two thirds of doctors and nurses in one hospital cited better communications.11 Most doctors deny the deleterious effects of stressors and proclaim that their decision making is as good in emergencies as in normal situations. In data just collected in a US teaching hospital, 30% of doctors and nurses working in intensive care units denied committing errors.13

Further exploring the relevance of aviation experience, we have started to adapt the threat and error model to the medical environment. A model of threat and error management fits within a general "input-process-outcomes" concept of team performance, in which input factors include individual, team, organisational, environmental, and patient characteristics. Professional and organisational cultures are critical components of such a model.

Threats are defined as factors that increase the likelihood of errors and include environmental conditions such as lighting; staff related conditions such as fatigue and norms of communication and authority; and patient related issues such as difficult airways or undiagnosed conditions. Latent threats are aspects of the system predisposing threat or error, such as staff scheduling policies. The model is shown in fig 2 and is explained more fully, together with a case study (see box for summary), on the BMJ 's website.


Case study: synopsis

An 8 year old boy was admitted for elective surgery on the eardrum. He was anaesthetised and an endotracheal tube inserted, along with internal stethoscope and temperature probe. The anaesthetist did not listen to the chest after inserting the tube. The temperature probe connector was not compatible with the monitor (the hospital had changed brands the previous day). The anaesthetist asked for another but did not connect it; he also did not connect the stethoscope.

Surgery began at 08 20 and carbon dioxide concentrations began to rise after about 30 minutes. The anaesthetist stopped entering CO2 and pulse on the patient's chart. Nurses observed the anaesthetist nodding in his chair, head bobbing; they did not speak to him because they "were afraid of a confrontation."

At 10 15 the surgeon heard a gurgling sound and realised that the airway tube was disconnected. The problem was called out to the anaesthetist, who reconnected the tube. The anaesthetist did not check breathing sounds with the stethoscope.

At 10 30 the patient was breathing so rapidly the surgeon could not operate; he notified the anaesthetist that the rate was 60/min. The anaesthetist did nothing after being alerted.

At 10 45 the monitor showed irregular heartbeats. Just before 11 00 the anaesthetist noted extreme heartbeat irregularity and asked the surgeon to stop operating. The patient was given a dose of lignocaine, but his condition worsened.

At 11 02 the patient's heart stopped beating. The anaesthetist called for code, summoning the emergency team. The endotracheal tube was removed and found to be 50% obstructed by a mucous plug. A new tube was inserted and the patient was ventilated. The emergency team anaesthetist noticed that the airway heater had caused the breathing circuit's plastic tubing to melt and turned the heater off. The patient's temperature was 108°F. The patient died despite the efforts of the code team.

At first glance, the case seems to be a simple instance of gross negligence during surgery by an anaesthetist who contributed to the death of a healthy 8 year old boy by failing to connect sensors and monitor his condition. When the model was applied, however, nine sequential errors were identified, including those of nurses who failed to speak up when they observed the anaesthetist nodding in a chair and the surgeon who continued operating even after the anaesthetist failed to respond to the boy's deteriorating condition. More importantly, latent organisational and professional threats were revealed, including failure to act on reports about the anaesthetist's previous behaviour, lack of policy for monitoring patients, pressure to perform when fatigued, and professional tolerance of peer misbehaviour.

    Establishing error management programmes

Available data, including analyses of adverse events, suggest that aviation's strategies for enhancing teamwork and safety can be applied to medicine. I am not suggesting the mindless import of existing programmes; rather, aviation experience should be used as a template for developing data driven actions reflecting the unique situation of each organisation.

This can be summarised in a six step approach. As in the treatment of disease, action should begin with

The history must include detailed knowledge of the organisation, its norms, and its staff. Diagnosis should include data from confidential incident reporting systems and surveys, systematic observations of team performance, and details of adverse events and near misses.

Further steps are:

Some might conclude that such programmes may add bureaucratic layers and burden to an already overtaxed system. But in aviation, one of the strongest proponents and practitioners of these measures is an airline that eschews anything bureaucratic, learns from everyday mistakes, and enjoys an enviable safety record.

Funding for research into medical error, latent factors in the system, incident reporting systems, and development of training is essential for implementation of such programmes. Research in medicine is historically specific to diseases, but error cuts across all illnesses and medical specialties.

I believe that if organisational and professional cultures accept the inevitability of error and the importance of reliable data on error and its management, systematic efforts to improve safety will reduce the frequency and severity of adverse events.

    Acknowledgments

Thanks to David Musson, Bryan Sexton, William Taggart, and John Wilhelm for their counsel.

    Footnotes

Funding: Partial support was provided by the Gottlieb Daimler und Carl Benz Stiftung.

Competing interests: RH has received grants for research in aviation from the federal government, has been a consultant for airlines, and has received honorariums for speaking to medical groups.

Website extra: View a full explanation of the threat and error management model, with a case study, here

    References

1. Helmreich RL, Foushee HC. Why crew resource management? Empirical and theoretical bases of human factors training in aviation. In: Wiener E, Kanki B, Helmreich R, eds. Cockpit resource management. San Diego: Academic Press, 1993:3-45.
2. Amalberti R. La conduite de systèmes à risques. Paris: Presses Universitaires de France, 1996.
3. Helmreich RL, Merritt AC. Culture at work: national, organisational and professional influences. Aldershot: Ashgate, 1998.
4. Helmreich RL, Merritt AC, Wilhelm JA. The evolution of crew resource management in commercial aviation. Int J Aviation Psychol 1999; 9: 19-32.
5. Federal Aviation Administration. Aviation safety action programs. Washington, DC: FAA, 1999. (Advisory circular 120-66A.)
6. Helmreich RL, Klinect JR, Wilhelm JA. Models of threat, error, and CRM in flight operations. In: Proceedings of the tenth international symposium on aviation psychology. Columbus: Ohio State University, 1999:677-682.
7. Klinect JR, Wilhelm JA, Helmreich RL. Threat and error management: data from line operations safety audits. In: Proceedings of the tenth international symposium on aviation psychology. Columbus: Ohio State University, 1999:683-688.
8. Helmreich RL. Culture and error. In: Safety in aviation: the management commitment: proceedings of a conference. London: Royal Aeronautical Society (in press).
9. Helmreich RL, Wilhelm JA. Outcomes of crew resource management training. Int J Aviation Psychol 1991; 1: 287-300.
10. Reason J. Managing the risks of organisational accidents. Aldershot: Ashgate, 1997.
11. Helmreich RL, Schaefer H-G. Team performance in the operating room. In: Bogner MS, ed. Human error in medicine. Hillside, NJ: Erlbaum, 1994:225-253.
12. Helmreich RL, Davies JM. Human factors in the operating room: Interpersonal determinants of safety, efficiency and morale. In: Aitkenhead AA, ed. Baillière's clinical anaesthesiology: safety and risk management in anaesthesia. London: Ballière Tindall, 1996:277-296.
13. Sexton JB, Thomas EJ, Helmreich RL. Error, stress, and teamwork in medicine and aviation: cross sectional surveys. BMJ 2000; 320: 745-749[Abstract/Free Full Text].


© BMJ 2000

This article has been cited by other articles:


Home page
J Thorac Cardiovasc SurgHome page
M. R. de Leval
Beyond Flatland
J. Thorac. Cardiovasc. Surg., January 1, 2003; 125(1): 12 - 19.
[Full Text] [PDF]


Home page
Acad Emerg MedHome page
K. S. Cosby and P. Croskerry
Patient Safety: A Curriculum for Teaching Patient Safety in Emergency Medicine
Acad. Emerg. Med., January 1, 2003; 10(1): 69 - 78.
[Abstract] [Full Text] [PDF]


Home page
JAMAHome page
L. L. Leape, D. M. Berwick, and D. W. Bates
What Practices Will Most Improve Safety?: Evidence-Based Medicine Meets Patient Safety
JAMA, July 24, 2002; 288(4): 501 - 507.
[Full Text] [PDF]


Home page
J Clin OncolHome page
R. B. Womer, E. Tracy, W. Soo-Hoo, B. Bickert, S. DiTaranto, and J. H. Barnsteiner
Multidisciplinary Systems Approach to Chemotherapy Safety: Rebuilding Processes and Holding the Gains
J. Clin. Oncol., December 15, 2002; 20(24): 4705 - 4712.
[Abstract] [Full Text] [PDF]


Home page
JAMAHome page
R. M. Epstein and E. M. Hundert
Defining and Assessing Professional Competence
JAMA, January 9, 2002; 287(2): 226 - 235.
[Abstract] [Full Text] [PDF]


Home page
Arch Intern MedHome page
T. S. Field, J. H. Gurwitz, J. Avorn, D. McCormick, S. Jain, M. Eckler, M. Benser, and D. W. Bates
Risk Factors for Adverse Drug Events Among Nursing Home Residents
Archives of Internal Medicine, July 9, 2001; 161(13): 1629 - 1634.
[Abstract] [Full Text] [PDF]


Home page
Qual Saf Health CareHome page
D C Aron and L A Headrick
Educating physicians prepared to improve care and safety is no accident: it requires a systematic approach
Qual. Saf. Health Care, January 6, 2002; 11(2): 168 - 173.
[Abstract] [Full Text] [PDF]


Home page
Can. J. AnaesthHome page
S. Fasting and S. E. Gisvold
Serious intraoperative problems - a five-year review of 83,844 anesthetics: [Problemes peroperatoires graves - une revue de 83 844 anesthesies sur cinq ans]
Can J Anesth, June 1, 2002; 49(6): 545 - 553.
[Abstract] [Full Text] [PDF]


Home page
Can Med Assoc JHome page
J. M. Davies
Painful inquiries: lessons from Winnipeg
Can. Med. Assoc. J., November 1, 2001; 165(11): 1503 - 1504.
[Full Text] [PDF]


Home page
Br J AnaesthHome page
G. C. L. Fletcher, P. McGeorge, R. H. Flin, R. J. Glavin, and N. J. Maran
The role of non-technical skills in anaesthesia: a review of current literature
Br. J. Anaesth., March 1, 2002; 88(3): 418 - 429.
[Abstract] [Full Text] [PDF]


Home page
Am J Respir Crit Care MedHome page
M. C. Hartlieb and W. J. Sibbald
Simple Interventions-Significant Impact
Am. J. Respir. Crit. Care Med., February 15, 2002; 165(4): 434 - 435.
[Full Text] [PDF]


Home page
J R Soc MedHome page
P. Young, R. Hamilton, S. Hodgett, M. Moss, C. Rigby, P. Jones, and R. Johanson
Reducing risk by improving standards of intrapartum fetal care
J. R. Soc. Med., January 5, 2001; 94(5): 226 - 231.
[Abstract] [Full Text] [PDF]


Home page
Acad MedHome page
J. B. Battles and C. E. Shea
A System of Analyzing Medical Errors to Improve GME Curricula and Programs
Acad. Med., February 1, 2001; 76(2): 125 - 133.
[Abstract] [Full Text] [PDF]


Home page
Acad Emerg MedHome page
J. A. Handler, M. Gillam, A. B. Sanders, and R. Klasco
Defining, Identifying, and Measuring Error in Emergency Medicine
Acad. Emerg. Med., November 1, 2000; 7(11): 1183 - 1188.
[Abstract] [Full Text]


Home page
BMJHome page
V. Barley, G. Neale, C. Burns-Cox, P. Savage, S. Machin, A. El-Sobky, A. Savage, A. Sheikh, B. Hurwitz, C K Connolly, P. M. Jonsson, G. Tomson, L. Räf, K. McCune, R. Law, R. M Goss, Anonymous, C. A Estrada, J. Carter, C. Brooks, A. C Jobe, J. Aldridge, P. Freeland, B. Ashcroft, M. Eistein, N. Boreham, P. J Shirley, L. Strunin, M. Morgan, P. Cartwright, and R. P Coffey
Reducing error, improving safety
BMJ, August 19, 2000; 321(7259): 505 - 505.
[Full Text]


Home page
BMJHome page
L. L Leape and D. M Berwick
Safe health care: are we up to it?
BMJ, March 18, 2000; 320(7237): 725 - 726.
[Full Text]


Home page
BMJHome page
J B. Sexton, E. J Thomas, and R. L Helmreich
Error, stress, and teamwork in medicine and aviation: cross sectional surveys
BMJ, March 18, 2000; 320(7237): 745 - 749.
[Abstract] [Full Text]

Rapid responses:

Read all Rapid responses

Should crew resource management training be mandatory in anaesthesia?
Peter J Shirley
bmj.com, 22 Mar 2000 [Full text]
Lessons from aviation
Ken McCune
bmj.com, 18 Apr 2000 [Full text]
Reducing Errors by Preventing Legal Sanctions for Error Reporting
Tess M Pape
bmj.com, 26 Apr 2001 [Full text]

Related letters in BMJ:

Reducing error, improving safety
Victor Barley, Graham Neale, Christopher Burns-Cox, Paul Savage, Sam Machin, Adel El-Sobky, Anne Savage, Aziz Sheikh, Brian Hurwitz, C K Connolly, Pia Maria Jonsson, Göran Tomson, Lars Räf, Ken McCune, Ron Law, Roger M Goss, Anonymous, Carlos A Estrada, James Carter, Clyde Brooks, Ann C Jobe, Jonathan Aldridge, Peter Freeland, Brenda Ashcroft, Max Eistein, Nicholas Boreham, Peter J Shirley, Leo Strunin, Maldwyn Morgan, Paul Cartwright, and Robert P Coffey
BMJ 2000 321: 505. [Letter]

Other related articles in BMJ:

EDUCATION AND DEBATE
Anaesthesiology as a model for patient safety in health care.
David M Gaba
BMJ 2000 320: 785-788. [Full text]  



PDF of this article
Email this article to a friend
Respond to this article
Read responses to this article
Related letters in BMJ
Other related articles in BMJ
PubMed citation
Related articles in PubMed
Download to Citation Manager
Search Medline for articles by:
Helmreich, R. L
Alert me when:
New articles cite this article
Collections under which this article appears:
Quality improvement (including CQI and TQM)


Home Help Search/Archive Feedback Table of Contents
BMJ Intended for health professionals
© 2003 BMJ Publishing Group Ltd