Movatterモバイル変換


[0]ホーム

URL:


Jump to content
WikipediaThe Free Encyclopedia
Search

Human error

From Wikipedia, the free encyclopedia
Action with unintended consequences
For other uses, seeHuman Error (disambiguation).

Human error is an action that has been done but that was "not intended by the actor; not desired by a set of rules or an external observer; or that led the task or system outside its acceptable limits".[1] Human error has been cited as a primary cause and contributing factor in disasters and accidents in industries as diverse asnuclear power (e.g., theThree Mile Island accident),aviation (e.g.,United Airlines Flight 173),space exploration (e.g., theSpace Shuttle Challenger disaster andSpace Shuttle Columbia disaster), andmedicine. Prevention of human error is generally seen as a major contributor toreliability andsafety of (complex) systems. Human error is one of the many contributing causes ofrisk events.

Definition

[edit]
A sign with aspelling mistake; the word "road" has been spelled incorrectly with a P instead of an R.

Human error refers to something having been done that was "not intended by the actor; not desired by a set of rules or an external observer; or that led the task or system outside its acceptable limits".[1] In short, it is a deviation from intention, expectation or desirability.[1] Logically, human actions can fail to achieve their goal in two different ways: the actions can go as planned, but the plan can be inadequate (leading to mistakes); or, the plan can be satisfactory, but the performance can be deficient (leading toslips andlapses).[2][3] However, a mere failure is not an error if there had been no plan to accomplish something in particular.[1]

Performance

[edit]
The Custom House in Dublin, which was built the wrong way around (the side facing theLiffey was intended to be the side facingGardiner Street)[citation needed].

Human error and performance are two sides of the same coin: "human error" mechanisms are the same as "human performance" mechanisms; performance later categorized as 'error' is done so in hindsight:[3][4] therefore actions later termed "human error" are actually part of the ordinary spectrum of human behaviour. The study ofabsent-mindedness in everyday life provides ample documentation and categorization of such aspects of behavior. Having a sense of awareness is needed to understand when dealing with a potential danger, thus being able to correct it.[5] While human error is firmly entrenched in the classical approaches to accident investigation and risk assessment, it has no role in newer approaches such asresilience engineering.[6]

Categories

[edit]

There are many ways to categorize human error:[7][8]

  • exogenous versusendogenous error (i.e., originating outside versus inside the individual)[9]
  • situation assessment versus response planning[10] and related distinctions in
  • by level of analysis; for example, perceptual (e.g.,optical illusions) versus cognitive versuscommunication versusorganizational
  • physical manipulation error[13]
    • 'slips' occurring when the physical action fails to achieve the immediate objective
    • 'lapses' involve a failure of one's memory or recall
  • active error - observable, physical action that changes equipment, system, or facility state, resulting in immediate undesired consequences
  • latent human error resulting in hidden organization-related weaknesses or equipment flaws that lie dormant; such errors can go unnoticed at the time they occur, having no immediate apparent outcome
  • equipment dependency error – lack of vigilance due to the assumption that hardware controls or physical safety devices will always work
  • team error – lack of vigilance created by the social (interpersonal) interaction between two or more people working together
  • personal dependencies error – unsafe attitudes and traps of human nature leading to complacency and overconfidence

Sources

[edit]

Thecognitive study of human error is a very active research field, including work related to limits ofmemory andattention and also todecision making strategies such as theavailability heuristic and othercognitive biases. Such heuristics and biases are strategies that are useful and often correct, but can lead to systematic patterns of error.

Misunderstandings as a topic in human communication have been studied inconversation analysis, such as the examination of violations of thecooperative principle and Gricean maxims.

Organizational studies of error or dysfunction have included studies ofsafety culture. One technique for analyzing complex systems failure that incorporates organizational analysis ismanagement oversight risk tree analysis.[14][15][16]

Controversies

[edit]
The "Hartlepool monkey" statue inHartlepool, England, commemorating a monkey who was mistaken by locals to be a French soldier and killed.

Some researchers have argued that the dichotomy of human actions as "correct" or "incorrect" is a harmfuloversimplification of acomplex phenomenon.[17][18] A focus on the variability of human performance and how human operators (and organizations) can manage that variability, may be a more fruitful approach. Newer approaches, such as resilience engineering mentioned above, highlight the positive roles that humans can play in complex systems. In resilience engineering, successes (things that go right) and failures (things that go wrong) are seen as having the same basis, namely human performance variability. A specific account of that is theefficiency–thoroughness trade-off principle,[19] which can be found on all levels of human activity, in individuals as well as in groups.

See also

[edit]

References

[edit]
  1. ^abcdSenders, J.W. and Moray, N.P. (1991)Human Error: Cause, Prediction, and Reduction[permanent dead link].Lawrence Erlbaum Associates, p.25.ISBN 0-89859-598-3.
  2. ^Hollnagel, E. (1993)Human Reliability Analysis Context and Control. Academic Press Limited.ISBN 0-12-352658-2.
  3. ^abcReason, James (1990)Human Error.Cambridge University Press.ISBN 0-521-31419-4.
  4. ^Woods, 1990
  5. ^Mohammadfam, Iraj; Mahdinia, Mohsen; Soltanzadeh, Ahmad; Mirzaei Aliabadi, Mostafa; Soltanian, Ali Reza (2021-07-01)."A path analysis model of individual variables predicting safety behavior and human error: The mediating effect of situation awareness".International Journal of Industrial Ergonomics.84 103144.doi:10.1016/j.ergon.2021.103144.ISSN 0169-8141.
  6. ^Hollnagel, E., Woods, D. D. & Leveson, N. G. (2006). Resilience engineering: Concepts and precepts. Aldershot, UK:Ashgate.
  7. ^Jones, 1999
  8. ^Wallace and Ross, 2006
  9. ^Senders and Moray, 1991
  10. ^Roth et al., 1994
  11. ^Sage, 1992
  12. ^Norman, 1988
  13. ^"Human Performance Improvement Handbook".U.S. Department of Energy. June 2009. DOE-HDBK-1028-2009.
  14. ^Rasmussen, Jens; Pejtersen, Annelise M.; Goodstein, L.P. (1994).Cognitive Systems Engineering.John Wiley & Sons.ISBN 0-471-01198-3.
  15. ^"The Management Oversight and Risk Tree (MORT)". International Crisis Management Association. Archived fromthe original on 27 September 2014. Retrieved1 October 2014.
  16. ^Entry for MORT on theFAA Human Factors Workbench
  17. ^Hollnagel, E. (1983)."Human error. (Position Paper for NATO Conference on Human Error, August 1983, Bellagio, Italy".
  18. ^Hollnagel, E. and Amalberti, R. (2001). The Emperor's New Clothes, or whatever happened to "human error"? Invited keynote presentation at 4th International Workshop on Human Error, Safety and System Development.. Linköping, June 11–12, 2001.
  19. ^Hollnagel, Erik (2009).The ETTO principle : efficiency-thoroughness trade-off: why things that go right sometimes go wrong. Farnham, England Burlington, VT: Ashgate.ISBN 978-0-7546-7678-2.OCLC 432428967.

External links

[edit]
Retrieved from "https://en.wikipedia.org/w/index.php?title=Human_error&oldid=1338445739"
Categories:
Hidden categories:

[8]ページ先頭

©2009-2026 Movatter.jp