Interesting People mailing list archives

Proper understanding of "The Human Factor"Risks Digest 23.07


From: Dave Farber <dave () farber net>
Date: Sat, 20 Dec 2003 12:50:39 -0500


Date: Thu, 11 Dec 2003 12:15:00 -0600
From: "Don Norman" <don () jnd org>
Subject: Proper understanding of "The Human Factor"

  [Warning: This is not a posting of some news item. It is an essay -- well,
  a lecture -- triggered by two recent RISKS postings, particularly because
  the second posting completely misunderstood the purpose of the first and
  didn't bother to read the book which was being recommended. And exhibited
  an attitude on the part of designers that is the biggest risk of all risks
  -- because it is the kind of attitude that causes the very problems the
  RISKS group is designed to eliminate.  DN]

If we assume that the people who use technology are stupid ("Bubbas") then
we will continue to design poorly conceived equipment, procedures, and
software, thus leading to more and more accidents, all of which can be
blamed upon the hapless users rather than the root cause -- ill-conceived
software, ill-conceived procedural requirements, ill-conceived business
practices, and ill-conceived design in general. This appears to be a lesson
that must be repeated frequently, even to the supposedly sophisticated
reader/contributor to RISKS.

It is far too easy to blame people when systems fail. The result is that
over 75% of all accidents are blamed on human error.  Wake up people! When
the percentage is that high, it is a signal that something else is at fault
-- namely, the systems are poorly designed from a human point of view. As I
have said many times before (even within these RISKS mailings), if a valve
failed 75% of the time, would you get angry with the valve and simply
continual to replace it? No, you might reconsider the design specs. Yo would
try to figure out why the valve failed and solve the root cause of the
problem. Maybe it is underspecified, maybe there shouldn't be a valve there,
maybe some change needs to be made in the systems that feed into the valve.
Whatever the cause, you would find it and fix it. The same philosophy must
apply to people.

Item. I predict that the municipal water and wastewater treatment industry
is in for a series of serious accidents. Why? Because of postings like that
of Dave Brunberg (RISKS-23.06). He was triggered by Mike Smith's
recommendation for the book "The Human Factor" (RISKS-23.04), but without
bothering to read the book. So he tells us of the "Bubba factor" in his
industry, namely, the belief that operators (named "Bubba") are
characterized by stupidity, laziness, and general ineptness. Brunberg
complains that he must make his software work despite the incompetence of
his operators: "you walk a very fine line between making the plant so
inflexible that operators cannot respond to unforeseen problems and giving
Bubba a little too much latitude."

No wonder we continue to have problems. It is this attitude of developers
that cause the very problems they complain about. The book, the Human
Factor, is in fact an excellent argument against Brunberg's point of view.
In it, the author (Kim Vicente) points out that procedural demands, business
practices that reward productivity and punish safety, and the inability of
system designers to understand the real requirements on the plat operators
are what leads to failure. Poor Bubba is yelled at by his bosses for slowing
up production, penalized if he raises questions about safety. If he follows
procedures, he can't meet production requirements. If he violates them --
which is what everyone is forced to do -- he is punished if an accident
occurs. No matter that lots of other Bubbas have warned about that
likelihood.

Let me also recommend the excellent "Field Guide to Human Error
Investigations." Here, the author (Sidney Dekker) points out that the old
view of human error is that it is the cause of accidents whereas the new
view is that it is a symptom of trouble deeper inside a system. Alas, the
"old" view is in actuality the current view, whereas the "new" view is still
seldom understood. (The "new" view has only been around for 50 years, so I
suppose we need to give it more time.). The Field Guide is about aviation,
but it is very applicable to the waste industry as well -- and to hospitals,
and emergency crews, and manufacturing plants, and any situation where
accidents are being blamed on people.

The most serious RISK in all this is that people take the easy way out,
blame the operator for incompetence, and then smile smugly from their
air-conditioned office, far away from the plant. As long as this attitude
persists, we will have bigger and bigger accidents.

DISCLAIMER (MILD). My strong recommendation for "The Human Factor" appears
on the back jacket of that book and on my website.  My equally strong
recommendation for the "Field Guide" will appear on my website Real Soon
Now.

Dekker, S. (2002). The field guide to human error investigations. Burlington
VT: Ashgate.

Vicente, K. J. (2003). The human factor: revolutionizing the way people live
with technology. Toronto: A. A. Knopf Canada.

Don Norman, Nielsen Norman Group and Northwestern University
norman () nngroup com http://www.jnd.org
-------------------------------------
You are subscribed as interesting-people () lists elistx com
To manage your subscription, go to
 http://v2.listbox.com/member/?listname=ip

Archives at: http://www.interesting-people.org/archives/interesting-people/


Current thread: