Swiss cheese model

For the Swiss Cheese model in physical cosmology, see large-scale structure of the cosmos, galaxy filament, and supercluster.
The Swiss cheese model of accident causation illustrates that, although many layers of defense lie between hazards and accidents, there are flaws in each layer that, if aligned, can allow the accident to occur.

The Swiss Cheese model of accident causation is a model used in risk analysis and risk management, including aviation, engineering, healthcare, and as the principle behind layered security, as used in computer security and defense in depth. It likens human systems to multiple slices of swiss cheese, stacked side by side, in which the risk of a threat becoming a reality is mitigated by the differing layers and types of defenses which are "layered" behind each other. Therefore in theory, lapses and weaknesses in one defense do not allow a risk to materialize, since other defenses also exist, to prevent a single point of weakness. The model was originally formally propounded by Dante Orlandella and James T. Reason of the University of Manchester,[1] and has since gained widespread acceptance. It is sometimes called the cumulative act effect.

Although the Swiss cheese model is respected and considered to be a useful method of relating concepts, it has been subject to criticism that it is used over broadly, and without enough other models or support.[2]

Failure domains

Reason hypothesized that most accidents can be traced to one or more of four failure domains: organizational influences, supervision, preconditions and specific acts. Preconditions for unsafe acts include fatigued air crew or improper communications practices. Unsafe supervision encompasses for example, pairing inexperienced pilots on a night flight into known adverse weather. Organizational influences encompass such things as reduction in expenditure on pilot training in times of financial austerity.[3]

Holes and slices

In the Swiss Cheese model, an organisation's defenses against failure are modeled as a series of barriers, represented as slices of cheese. The holes in the slices represent weaknesses in individual parts of the system and are continually varying in size and position across the slices. The system produces failures when a hole in each slice momentarily aligns, permitting (in Reason's words) "a trajectory of accident opportunity", so that a hazard passes through holes in all of the slices, leading to a failure.[4][5][6]

Frosch[7] described Reason's model in mathematical terms as a model in percolation theory, which he analyses as a Bethe lattice.

Active and latent failures

The Swiss Cheese model includes both active and latent failures. Active failures encompass the unsafe acts that can be directly linked to an accident, such as (in the case of aircraft accidents) pilot error. Latent failures include contributory factors that may lie dormant for days, weeks, or months until they contribute to the accident. Latent failures span the first three domains of failure in Reason's model.[3]

Applications

The same framework applies in healthcare. For example, a latent failure could be the similar packaging of two drugs that are then stored close to each other in a pharmacy. Such a failure would be a contributory factor in the administration of the wrong drug to a patient. Such research led to the realization that medical error can be the result of "system flaws, not character flaws", and that greed, ignorance, malice or laziness are not the only causes of error.[8]

Lubnau, Lubnau, and Okray[9] apply the model to the engineering of firefighting systems, aiming to reduce human errors by "inserting additional layers of cheese into the system", namely the techniques of Crew Resource Management.

This is one of the many models listed, with references, in.[10]

See also

References

  1. Reason 1990.
  2. "Revisiting the Swiss cheese model of accidents" (PDF). Eurocontrol. October 2006.
  3. 1 2 Douglas A. Wiegmann and Scott A. Shappell (2003). A Human Error Approach to Aviation Accident Analysis: The Human Factors Analysis and Classification System. Ashgate Publishing, Ltd. pp. 4849. ISBN 0754618730.
  4. Daryl Raymond Smith, David Frazier, L W Reithmaier, and James C Miller (2001). Controlling Pilot Error. McGraw-Hill Professional. p. 10. ISBN 0071373187.
  5. Jo. H. Wilson, Andrew Symon, Josephine Williams, and John Tingle (2002). Clinical Risk Management in Midwifery: the right to a perfect baby?. Elsevier Health Sciences. pp. 46. ISBN 0750628510.
  6. Tim Amos and Peter Snowden (2005). "Risk management". In Adrian J. B. James, Tim Kendall, and Adrian Worrall. Clinical Governance in Mental Health and Learning Disability Services: A Practical Guide. Gaskell. p. 176. ISBN 1904671128.
  7. Robert A. Frosch (2006). "Notes toward a theory of the management of vulnerability". In Philip E Auerswald, Lewis M Branscomb, Todd M La Porte, and Erwann Michel-Kerjan. Seeds of Disaster, Roots of Response: How Private Action Can Reduce Public Vulnerability. Cambridge University Press. p. 88. ISBN 0521857961.
  8. Patricia Hinton-Walker, Gaya Carlton, Lela Holden, and Patricia W. Stone (2006-06-30). "The intersection of patient safety and nursing research". In Joyce J. Fitzpatrick and Patricia Hinton-Walker. Annual Review of Nursing Research Volume 24: Focus on Patient Safety. Springer Publishing. pp. 89. ISBN 0826141366.
  9. Thomas Lubnau II, Randy Okray, and Thomas Lubnau (2004). Crew Resource Management for the Fire Service. PennWell Books. pp. 2021. ISBN 1593700067.
  10. Taylor, G.A., Easter, K.M., Hegney, R.P. (2004). Enhancing Occupational Safety and Health. Elsevier. pp. 241245, see also pages 140–141 and pages 147–153, also on Kindle. ISBN 0750661976.

Further reading

This article is issued from Wikipedia - version of the Tuesday, February 16, 2016. The text is available under the Creative Commons Attribution/Share Alike but additional terms may apply for the media files.