Title: Nancy Leveson: Safeware CHAPTER 1: Risk in Modern Society
1Nancy Leveson Safeware
CHAPTER 1 Risk in
Modern Society
- Three Mile Island 1979
- Bhopal, 1984
- Space Shuttle, 1986
- Chernobyl, 1986
- Therac-25 1985 - 1987
2IS NEW TECHNOLOGY MAKING OUR WORLD RISKIER?
- In the United States technological hazards
account for 15 to 25 of human mortality and
have significantly surpassed natural hazards in
impact, cost and general importance. - Ref 87 - Ned Franklin. The accident at
Chernobyl. - The Chemical Engineer, pages 17-22, November,
1986. p.4
3Flood damage in the United States, for example,
has increased as expenditures on flood control
have increased. 154 - Trevor Kletz. Myths of
the Chemical industry. The Institution of
Chemical Engineers, Rugby, Warwickshire, United
Kingdom, 1984.p.4
4In fact, all hazards are affected by complex
interactions among technological, ecological,
sociopolitical, and cultural systems 30, 174,
3339.
- Attempts to control risk by treating it simply
as a technical problem or only as a social issue
are doomed to fail or to be less effective than
possible.
51.1 Changing Attitudes toward Risk
- Not Safe, only Safer.
- Societies are recognizing human and workers
rights - Workers are at the mercy of their employers in
terms of safety - Complete abdication of personal responsibility,
however, is not always wise
6In some instances -- such as the Bhopal accident
-- the public has completely trusted others to
plan for and respond effectively to an emergency,
with tragic results.
- Many aspects of Bhopal guaranteed that an
accident would occur. - emergency and evacuation planning, training, and
equipment were inadequate. - Public not told of simple measures (such as
closing eyes with wet cloth over faces) which
could have saved their lives.
7In writing about the Bhopal tragedy, Bogard
expresses this new attitude
- We are not safe from the risks posed by hazardous
technologies, and any choice of technology
carries with it possible worst case scenarios
that must take into account in any implementation
decision. The public has the right to know
precisely what these worst case scenarios are and
participate in all decisions that directly or
indirectly affect their future health and
well-being.
8In many cases, we must accept the fact that the
result of employing such criteria may be a
decision to forego the implementation of a
hazardous technology altogether 30, p. 109. p.6
9Is risk increasing in our modern society as a
result of new technological achievements, or are
we simply experiencing a new and unjustified form
of Luddhism? p.6
- The Luddite disturbances occurred in Yorkshire
between 1811 and 1816, when workers in the
English woolen industry tried, through violence,
to stem the increasing mechanization of the
mills. Luddism has become a generic term
describing opposition to technological innovation
101.2 Is Increased Concern Justified?
- Is technological risk increasing? Depends on
data used and its interpretation p.6 - Harris and colleagues argue that technological
hazards, in terms or human mortality, were
greater in the earlier, less fully managed stages
of industrial development 112 - Data cited from National Safety Council shows
that occupational death and injury rates have
declined steadily since the early part of this
century 112.
11NSC Concludes that technological hazard mortality
is not currently rising.
- However warning that
- The positive effects of technology have for some
time reached their maximum effect on human
mortality, while the hazards of technology
continue partially unchecked, affecting
particularly the chronic causes of death
currently account for 85 percent of mortality in
the U.S.A. 112 - On other hand, examination of technological
accident rate, rather than the occupational death
and injury rate, suggests that the technological
risk is increasing - 60 of all major industrial disasters from 1921
to 1989 occurred after 1975 30.
12Bogard (1989) argued that 12 of the 19 major
industrial accidents in the twentieth involving
100 or more deaths occurred after 1950.
- If small-scale accidents also included
(transportation, dams, structural collapses)
evidence is more compelling. - Example Military Aviation accident rate has
slowed probably due to emphasis on system
safety. - Past experience does not allow us to predict the
future when the risk factors in the present and
future differ from those in the past. Examining
these changes will help us understand the
problems we face.
131.3 Unique Risk Factors in Industrialized Society
- RISK Likelihood of an accident plus severity
of the potential consequences - Factors include new hazards, increasing
complexity, exposure, energy, automation,
centralization, scale, and pace of technological
change in systems
141.31 The Appearance of New Hazards
- Misuse and overuse of antibiotics have created
resistant microbes - Children no longer work in mines or as chimney
sweeps, but are exposed to man-made chemicals and
pesticides in their food or increased
environmental pollution 57 - Atomic energy has increased potential for death
and injury from radiation exposure
15Redundancy (duplication of components to protect
against individual failures)
- Not effective means of controlling risks
- not effective against hazards that arise from
interactions among components in todays
increasingly complex and interactive systems. - May in fact increase complexity to the point
where the redundancy itself contributes to
accidents
161.3.2 Increasing Complexity
- Perrow distinguishes between accidents caused by
component failures and those which he calls
system accidents, that are caused by interactive
complexity in the presence of tight coupling
259, Normal Accidents - High technology systems are often made up of
networks of closely related subsystems.
Conditions leading to hazards emerge in the
interfaces between subsystems, and disturbances
progress from one component to another.
17 Increasing Complexity/Cont.An example of this
increasingly common type of complexity, modern
petrochemical plants often combine several
separate chemical processes into one continuous
production, without the intermediate storage that
would de-couple the subsystems 274.
- analysis of major industrial accidents
invariably reveal highly complex sequences of
events leading up to accidents, rather than
single component failures. p.8
18Increasing Complexity Cont./
In the past component failure was cited as the
major factor in accidents, today more accidents
result from dangerous design characteristics and
interactions among components 108 p8 top.
19Not only does functional complexity make the
designers task more difficult, but the
complexity and scope of the projects require
numerous people and teams to work together. The
anonymity of team projects dilutes individual
responsibility 172
- Paradox that people are willing to spend money on
complexity but not on simplicity 158, Kletz
WHY IS THIS THE CASE?
20CASE in Point A British Chemical Plant (p.9)
- pump, various pipelines, had several uses
including - transferring methanol from a road tanker to
storage - charging it to the plant,
- moving methanol back from a road tanker to
storage, - charging it to the plant
- moving recovered methanol back from the plant
21On this particular occasion, a tank truck was
being emptied
- The pump had been started from the control panel
but had been stopped by means of a local button.
The next job was to transfer some methanol from
storage to the plant. - the computer set the valves, but as the pump had
been stopped manually it had to be started
manually.
22When the transfer was complete the computer told
the pump to stop, but because it had been stated
manually it did not stop and a spillage occurred
157, p.225. P 9
- a simpler design -- independent pipelines for
different functions, actually installed after the
spill, made errors much less likely and was not
more expensive over the lifetime of the equipment.
23Computers
- may encourage the introduction of unnecessary and
dangerous complexity - enable more interactive, tightly coupled, and
error-prone designs to be built - Kletz has noted Programmable electronic systems
have not introduced new forms of error, but by
increasing the complexity of the the processes
that can be controlled, they have increased the
scope of the introduction of conventional errors
158.
24Accepting Perrows argument that interactive
complexity and coupling are a cause of serious
accidents, then the introduction of computers to
control dangerous systems, may increase risk
unless great care is taken to minimize complexity
and coupling..
Conclusion
251.3.3 More People Exposed to Hazards
- Larger flight capacities
- Dangerous plant facilities closer to population
centers - More of workforce in cities or within commuting
distance. - Interdependencies and complexity cause ripple
effects of hazards magnifying potential
consequences. - 1.3.4 High Energy Sources increase risks
261.3.5 Increasing Automation of Manual Operations
- Automation does not remove humans but tends to
redefine their roles - operators become concerned with maintenance,
repair and higher level supervisory control and
decision making 270
27Operators relegated to central control roomsCase
in Point 1977 NYC Blackout
- Indirect Information
- Operator followed prescribed procedures
- But electrical system was brought to a complete
halt - Operator could not know there were two relay
failures - one leading to a high flow over line normally
carrying little or no current (operator would
have been alerted) - other blocked flow over line making it appear
normal
28- Operator had no way of knowing that zero reading
would appear normal. - Operators become the scapegoat of an automated
system.
29Operators and Embedded Systems
- Embedded systems can mask the occurrence and
subsequent development of a problem - When malfunction is discovered it may be more
difficult to control - Systems may be hidden or distorted
- Such design further limits operator options and
hinders broad comprehension. p.11
30Case China Airlines, 1985
- 747 suffered slow loss of power in right outer
engine - autopilot compensated preventing yaw to the right
- when limit reached, crew had no time to determine
cause - plane rolled and went into vertical dive of
31,500 ft. - Aircraft severely damaged
31Multiple Goals may lead to conflicts
- 1970s attempt to combine energy savings efforts
with process plants lead to complications - Safety and Economy conflicting
- component interactions make system functions less
transparent to designers and operators - Place where trouble first recognized may not be
where it started.
321.3.7 Increasing Scale and Centralization
- Bupp and Derian 1968 observed that
manufacturers were taking orders for nuclear
power plants six times the size of those in
operation. - Previously 1-2 times extrapolations were
considered at the outer boundary of acceptable
risk. - Browns Ferry Accident, 1975 was 10 times size of
plant construction in 1966
33 Supertankers built without sound design and
redundancy
- Mostert writes
- The gigantic scale of vessels creates an abstract
environment in which crews are far removed from
direct experience of the seas unforgiving
qualities and potentially hostile environment.
Heavy automation undermines much of the
old-fashioned vigilance and induces engineers to
lose their occupational instincts -- qualities
that in earlier days of shipping were an
invaluable safety factor.
34Increasing Pace of Technological Change
- Average time of conversion of technical discovery
into commercial product was 30 years 1880 -1919
- now 5 years!
- number of new products increasing exponentially
- dangerous new substances with economic pressures
preventing adequate testing.
35Pace of Technological Change/ cont.
- Learning by trial and error not possible in
modern times - Design and testing procedures must be right from
the start. - Christopher Hinton (1957) writing about nuclear
power pointed out that in other domains, learning
from failures was possible. - Progress continues at torrid pace new standards
and regulatory procedures are necessary.
361.4 How Safe Is Safe Enough?
- The goal is to understand and manage risk in
order to eliminate accidents or to reduce their
consequences. - Frola and Miller (88) claim that system safety
investment has reduced losses where it has been
applied rigorously in military and aerospace
programs.
37How Safe is Safe/ cont.
- Conflicting goals between
- safety, performance, and goals.
- e.g. Industrial arm which cannot be stopped
easily. Slower arm determines that it must hit
something. - human interface(s) designed for ease of use are
often more troublesome
38Risk - Benefit Analysis
- Often viewed as only way to make technology and
risk analysis. - Must be able to
- measure risk
- choose appropriate level for risk
- Systems must be designed and built while
knowledge of their risk is incomplete or
nonexistent.
39Risk Assessment / cont.
- Impossible to measure before system is built.
- Failure rate of 4 (over 10000 years) and use
body is missing - Hard to perform and hard to determine acceptable
risk. - Case of Ford Pinto gas tank
- Optimal Risk involves a tradeoff that minimizes
the sum of all undesirable consequences.
40Risky Systems and Perrow
- Perrow Divides High-Risk systems into 3
categories