KNOWING WHEN AND HOW TO STOP
© 1998 John E. Perkins, III
5201 22nd Ave. NE Ste. 201
Seattle WA 98105
206 524.4496
Send email to the address on our home page:
http://www.drizzle.com/~newroots
Introduction
Excitation or inhibition, the fundamental states of our nervous system's
cells (Pavlov, 1966: 354), also describes in a gross binary way decisions
we face singly, in pairs, in groups, as a nation, and as a species. Poetically,
Hamlet ponders his fate in his "to be or not to be" solloquy;
more prosaically we do the same: does the team stay in town, or leave? Do
I stay in this relationship for the indefinite future, or do I get out now?
Scholars who study decision making like to present the process as one of
stages (Poole & Hirokawa, 1986; Janis and Mann, 1977). Each stage typically
has some tasks which need to be completed before the decision making group
moves to the next stage. After leaving a stage, people will not return to
it as a matter of routine, but will make incremental adjustments to their
original decision as experience accrues (Janis and Mann, 1977: 75). Experience
and research confirm that decisions must be made in an imperfect world;
that is, the decider(s) will not have all of the information, nor will he
or she be able to perfectly predict the behavior of others; sometimes the
decider(s) cannot even predict their own behavior (Staw, 1989; Of Auctions,
1989).
Knowing when, and how to, stop has been the shadow side of decision making.
I personally began to understand this during my time as an anti-war activist
in the 1980's. As I watched government policy as described publicly in the
media, I understood that the governments which had nuclear weapons had no
serious criteria-and thus no intentions-of stopping.
I remember a fellow activist and I discussing this point. She commented
that when she attended public discussions on nuclear power plants, she always
would ask what the utility's plans were if they had to close the plant.
They had started something, without thinking through what they would do
if they had to end it. They never had a plan, and seem confused by the question.
The third source of my reflections on stopping come from when I worked with
elementary students in the South Bronx on issues of decision making and
self-esteem. For one of our exercises, I would ask them to describe "something
you used to say 'yes' to that you now say 'no' to, and something you used
to say 'no' to that you now say 'yes' to." We would role-play the switch
and have fun with people resisting being persuaded.
I turned to the serious problem of their upcoming adolescence. By a show
of hands, all of them felt confident they would not take illegal drugs.
But the reality was that many of them would. A one time experiment might,
or might not, be anything to get too upset about. "But," I asked,
"do you have what it takes to say 'yes' once, then 'no' for the rest
of your life? It's not the first time that makes you an addict-it's the
fifth, the thirteenth, the sixty-fifth time. And remember, when you try
it, it might give you an experience of sights, sounds, colors, and moods,
you've never had before, and could never have without the use of illegal
drugs. Can you then say 'no' to that drug?" I reminded them that the
next six to eight years of their lives would probably be the hardest, and
the decisions they made as fifth and sixth graders about what they would
do, or not do, would affect those years, and the rest of their lives. I
said once they reached junior high school, life and choices would come at
them fast and furiously. I ended by telling them we live in a culture which
is great in encouraging them to start things, but nearly silent about how
to stop or prevent things.
With these personal experiences alerting me, I included this study as a
unit in my program, focusing at the organizational level. Knowing how and
when to stop has basically two discontinuous moments of truth: deciding
to stop before going public and deciding to stop once the decision is public.
Closing what Meyer and Zucker (1989) call permanently failing organizations-deciding
to end projects or institutions which once had experienced success-become
cases of public decision making.
In the case of decisions which some the deciders themselves will later view
as patent folly, the best policy appears to stop early and privately. But
deciders make decisions with incomplete information and only guesses to
how others might behave. Section One, Folly on the March, will take up this
problem in greater detail.
Publicity about a decision, or publicly implementing one, widens the number
of participants and observers, and thus the difficulty of stopping increases
substantially. The decision becomes rigid and resistant to changes (Schulman,
1980: 43). This is because stopping "in the middle" or "before
the end" has added difficulties of how to manage one's own "face"
or public reputation (Bohannon, 1993; Ting-Toomey, 1994); how to persuade
involved others that stopping what you once supported makes more sense now;
and how to deal with incurred psychological, symbolic, material, and social
costs. This complex area of decision making will be considered in Section
Two: In Too Far to Quit?
In the last section of the paper, Can Folly be Prevented?, I will propose
some tentative recommendations and reflect on my experience of researching
this paper.
Section One: Folly on the March
When Barbara Tuchman looked at the March of Folly (1984) through
history, she established some criteria for what she would consider folly.
Her three criteria can help distinguish between foolishness and an honest
mistake. To fit her definition of folly, a course of action:
(1) must have been perceived as counter-productive in its own
time, not by hindsight;
(2) a feasible alternative course of action must have been available; and
(3) the policy in question must be that of a group, not of individual ruler,
and should persist beyond any one political lifetime (p. 5).
Taking up Tuchman's third criteria, in 1972, Irving Janis described many
of the difficulties posed for the "group" in his ground-breaking
book, Groupthink. The groupthink hypothesis asserts that "the existence
of certain antecedent conditions within groups of decision makers result
in defective decision-making processes, which in turn are linked to poor
policy outcomes" (Schafer and Crichlow, 1996: 415). Schafer and Crichlow
in their quantitative review of this theory, simplified Janis's groupthink
model to three stages as shown in this figure.
From an exhaustive review of Janis's works, Schafer and Crichlow developed
a refined list of operational definitions for ten antecedent conditions:
1. Group insulation: Decision makers isolate themselves
from others not in the immediate decision making circle.
2. Lack of tradition of impartial leadership: The president has not
had a history of conducting impartial decision-making processes, which thereby
limits open discussion of a wide range of alternatives.
3. Lack of tradition of methodical procedures: The president has
not established a tradition of using methodical procedures in the decision-making
process in terms of information search, routine and systematic decision-making
meetings, and analysis of pros and cons.
4. Group homogeneity: A lack of disparity exists in the social background
and ideology of the members of the decision-making group.
5. Perceived short time constraint: The group suffers under perceived
temporal limits that affect its ability to consider policy options fully.
6. Low self-esteem caused by recent failure: Recent political or
military defeat weights on the minds of members of the decision-making group
and affects current decisions.
7. High personal stress: The crisis being dealt with causes great
anxiety because of either the stakes involved and the perceived chances
of success or unpleasant policy options.
8. Overestimation of the group: The group operates under an illusion
of invulnerability or a belief in its inherent morality.
9. Closed-mindedness: The group relies on collective rationalizations,
stereotypes of the out-group, or guiding metaphors or analogies.
10. Pressure towards uniformity: One or more of the following exists:
self-censorship, an illusion of unanimity, direct pressure on dissenters,
self-appointed mind guards (pp. 418-419).
Schafer and Crichlow investigated the same 19 Cold War crises that Janis
had used to test the groupthink hypothesis in 1987 (Hekek, et al). They
used bivariate analysis to correlate each of the ten conditions with the
level of information processing errors. Four of the ten antecedent conditions
proved statistically significant and in the predicted direction: lack of
tradition of impartial leadership, lack of tradition of methodical procedures,
overestimation of the group, and closed-mindedness. A fifth condition, pressures
towards uniformity, came very close to significance and accounted for fifteen
percent of the variance on its own and was in the expected direction as
well (p. 423). When lumped together, Schafer and Crichlow found that these
five variables showed almost a one to one correlation between faulty structural
environments and the number of information-processing errors (p. 425).
None of the three situational variables-short time constraint, high personal
stress, and recent failure-proved significant. In fact, short time constraint
and high personal stress seemed to reduce information processing errors.
In other words, some contextual conditions result in more vigilant decision
making.
An Example of Groupwisdom
Interestingly, two of the four significant antecedents to groupthink contain
the phrase "lack of tradition." The American Heritage dictionary
defines tradition as
1. The passing down of elements of a culture from generation
to generation, especially by oral communication.
2.a. A mode of thought or behavior followed by a people continuously
from generation to generation; a custom or usage. b. A set of such
customs and usages viewed as a coherent body of precedents influencing the
present.
Which teases out this idea: not only must the group learn to remain humble
(to overcome overestimation) and learn to tolerate ambiguity (to overcome
close-mindedness), it must continuously examine and deepen its own culture
of decision making, authority, membership, argument, etc. In other words,
in groupthink mode the group lacks a tradition of inquiry into its own modes
of behaving in normal and stressful situations.
Campbell and Nash in A Sense of Mission (1992) describe how some improper
payments by one of its overseas operations prompted Johnson & Johnson's
president James Burke to launch a series of Credo Challenge Meetings. Collins
and Porras (1994) also cite Johnson & Johnson as one of their "visionary
companies" because of the impact these meetings had on the decision
making processes in the company.
In 1975, Johnson & Johnson had annual sales of about $12 billion and
had developed a management culture resistant to centralization (Campbell
and Nash: 138). General Robert Wood Johnson published The Johnson &
Johnson Credo in 1945 for the enlightenment of other business people. The
entire document fits easily on a single page. It lists four valued relationships,
each given a full paragraph which defines the relationship and establishes
Johnson & Johnson's expected obligations to that relationship. The table
on the next page summarizes the 1975 revised Credo.
After much debate, including reservations from the chairman and CEO who
that felt the Credo should be above "challenge," President Burke
decided to hold company-wide discussions about the value of the Credo in
1975. The Credo was then thirty years old. With the Credo Challenge Meetings
Burke began a "long-range program to challenge every manager to make
an informed commitment to the Credo's way of doing business" (p. 143).
In other words, he began what might turn into a new tradition. Indeed, Credo
Challenge Meetings are still held for new managers (p. 146).
Since 1975, over 1200 managers have participated in Credo Challenge Meetings,
gathering for two days in groups of 25. These meetings made a point of always
discussing how the Credo should be implemented in the management of the
Johnson & Johnson company. Interestingly, during these discussions,
Burke's boss, the chairman/CEO, sometimes had to excuse himself because
he "breathed too heavily on the process" (p. 145). In the language
of antecedent conditions to groupthink, he could not maintain leadership
impartiality.
Burke called the results, "A turn on. A genuine happening" (p.
145). He learned that managers felt that balancing all of their responsibilities
required discussion because of the difficulty of the task, and most managers
felt intense commitment to preserving the Credo. Another result of these
Credo Challenge Meetings is that managers facing a decision pause to reflect
upon what the Credo would say to do, and discussions based on the Credo
have become habitual (p. 148). In essence, they added a step which precedes
Janis's antecedents of decision making: continually discuss organizational
values and seek personal commitments to them. Schafer and Crichlow's diagram
can be modified to show how Groupwisdom might operate:
Table 1. Elements in Johnson
& Johnson's 1975 Revised Credo |
---|
Customers: doctors, nurses, patients,
etc. | We must strive to reduce our costs in order to maintain reasonable
prices. Customers' orders must be serviced promptly and accurately. |
---|
Suppliers
| Must have an opportunity to make a fair profit. |
---|
Employees |
We must respect their dignity and recognize their merit. They must have
a sense of security in their jobs. Compensation must be fair and adequate,
and working conditions clean, orderly, and safe. Employees must feel free
to make suggestions and complaints. We must provide competent management
and their actions must be just and ethical. |
---|
Communities where people
live and work | We must be good citizens-support good works and charities
and bear our fair share of taxes. We must maintain in good order the property
we are privileged to use, protecting the environment and natural resources.
|
---|
Stockholders | Business must make a sound profit. We must experiment
with new ideas. Research must be carried on, innovative programs developed
and mistakes paid forWhen we operate according to these principles, the
stockholders should realize a fair return.
(adapted from Campbell and Nash, p. 141). |
---|
Why discuss this in a paper about failures of group decision making? Because
only studying failures may not point to where success lies. The Credo discussions
prepared Johnson & Johnson to succeed dramatically in making value-based
decisions under great time pressure and uninvited public scrutiny.
In 1982, Johnson & Johnson's Tylenol brand pain-killer held 37 percent
of the domestic market. It led the market. Ironically, it had been launched
the same year that the Credo Challenge Meetings began. On September 30,
1982, seven people died in Chicago from taking cyanide-tainted Tylenol.
Shortly after the news reached Johnson & Johnson, they recalled 93,000
bottles and sent almost half a million messages to doctors, hospitals, and
retailers warning them about the tainted capsules. Within a week they would
recall or replace 22 million capsules. On the day after the poisonings,
Johnson & Johnson canceled all advertising for the brand, except for
one ad asking for the public's continued trust in Johnson & Johnson
(pp. 151-152).
They set up a toll-free number-staffed by employees volunteering their time-for
customers to get a refund without submitting proof of purchase. One weekend
the toll-free number handled over a million calls, but many of those were
good wishes from the public.
Everyone, the media, government, and employees, watched how Johnson &
Johnson would respond. At its conclusion, one employee said:
Tylenol was the tangible proof of what they said at the Credo
Challenge Meetings. You came away saying, "My God! You're right: we
really do believe this. It's for real. And we did the right thing."
Actually, the whole organization had spent seven years preparing itself
to act principally. It prepared itself to respond quickly and congruently
to bad news.2 With their masterful handling of the crisis, Johnson &
Johnson returned Tylenol to the shelves with tamper-resistant packaging.
Backed by a massive marketing effort, the brand regained 95 percent of it
previous market share and once again led the field (p. 154).
Responding in a crisis situation defines just one type of decision-making
a group is likely to face. Far more difficult to gauge and respond successfully
to are situations where the problem takes a long time to become salient.
I turn to those types of situations next.
Section Two: In too far to Quit?
Great is the art of beginning,
but greater is the art of ending.
-Henry Wadsworth Longfellow
Anyone attempting to stop something once it becomes public faces a frustrating
task. By public I mean the tangible proofs of a decision, whether begun
or just proposed. This includes: major projects such as the closing of factories,
missions to land a person on the moon, even setting a date for a wedding.
Decision making under public scrutiny has been described by different authors
as escalating situations, permanently failing situations, and large-scale
policy making. In escalating situations efforts to recoup losses simply
increase the losses (Staw and Ross, 1989). In permanently failing organizations
(Meyer and Zucker, 1989) what one group perceives as failure other groups
interpret as acceptable, or even as successful. According to Schulman (1980),
large-scale situations, whether successful or failing, create institutional
and personal commitments which resist being changed.
Escalating Situations
Escalating situations often turn on the very first commitment to a course
of action. In other words, once "go" gets the nod, decision makers
don't look back. Note the care General George C. Marshall, as Secretary
of State, took in 1948 to point out to Congress how difficult it would be
to extricate the country from a military commitment to back the Chinese
Nationalists: It would involve this government in a continuing commitment
from which it would practically be impossible to withdraw" (Neustadt
and May, 1986: 248-249).
Unfortunately, "who lost China" became the shibboleth of the fifties,
which may have prevented the bureaucracy from clearly pointing out the hazards
of escalating in Vietnam. Still, George Ball did attempt to warn President
Johnson of the dangers of escalating in Vietnam:
The decision you face now is crucial. Once large numbers of
U.S. troops are committed to direct combat, they will begin to take heavy
casualties in a war they are ill-equipped to fight in a noncooperative if
not downright hostile countryside. Once we suffer large casualties, we will
have started a well-nigh irreversible process. Our involvement will be so
great that we cannot-without national humiliation-stop short of achieving
our complete objectives. Of the two possibilities I think humiliation will
be more likely than the achievement of our objectives-even after we have
paid terrible costs (quoted by Staw and Ross: 1989: 216, quoting from The
Pentagon Papers).
Staw and Ross's (1989) article "Understanding Behavior in Escalating
Situations" begins with this quote from Ball. They describe "escalating
situations" as "situations in which losses have resulted from
an original course of action, but where there is the possibility of turning
the situation around by investing further time, money, and or effort."
In our daily living we face making decisions in escalating contexts, too:
what to do with a stalled career, falling investments, or a faltering relationship
(Staw and Ross, 1989: 216).
Some insights into escalating situations might come from looking at gambling.
Though a gambler might play dozens of spins of the wheel at the roulette
table in an evening, after the initial decision to play at all, subsequent
bets are not genuine decisions, merely a continuation of that first commitment.
Dostoevsky described this moment before the plunge in The Gambler:
I confess that my heart was pounding in my breast and that I
didn't feel at all cool and detached; probably I had felt for a long time
already that I would leave Roulettenberg a different man and that something
was about to happen which would radically and irrevocably change my life.
I felt that it was bound to happen (in Leonard, 1989: 39).
One can see in another Dostoevsky passage what Janis called overestimation:
I believe I had something like four thousand gulden in my hands
within five minutes. That's when I should have quit. But a funny feeling
came over me, some sort of desire to challenge Fate, an uncontrollable urge
to stick my tongue out at it, to give it a flip on the nose (p. 38).
Note the "funny feeling" Dostoevsky's gambler gets. Here literature
points to a facet of escalation some research misses: there exists something
in the emotion-action-results process of persistent gambling which feeds
a hunger beyond material success or failure. The gambler, confronted with
losing at the wheel, dice, horses, cards, etc. confronts reality's mismatch
to their intense feelings of confidence. He or she attempts to regain both
the confident feeling of precognition and external success by increasing
his or her betting. The gambler continues until all is lost.
From a review of research in the laboratory and the "real world,"
Staw and Ross have uncovered four classes of determinants-project, psychological,
social and organizational-for this persistence in losing situations (1989:
216). Each of these determinants possesses, naturally, both rational and
meta-rational dimensions. Though the research into these determinants has
been uneven, the existence of multiple determinants of commitment and persistence
leads Staw and Ross to view escalation as multidetermined. Persistent pursuit
of a failing course may most likely happen in situations characterized by
a series of seemingly insufficient small-impact variables, as say, the loss
of a few dollars on a particular bet out of a original stake of several
hundred dollars. They add, "A slow and irregular decline may not only
make a line of behavior difficult to extinguish (in the reinforcement theory
sense), but may allow the forces for persistence to grow over time"
(p. 219).
"Slow and irregular decline" can also be named permanent failure.
In their book on Permanently Failing Organizations, Meyer and Zucker (1989)
researched the literature on low performing organizations and presented
four case studies:
The decline of the Los Angeles Herald Examiner, the only paper actually
founded by William Randolph Hearst, and once the flagship of the Hearst
publishing empire-endured a nine year strike from 1967 to 1976 and the family
remains divided on whether to sell the paper or to attempt a comeback;
A worker-owned meat packing company where workers picketed their own company-it
declined for 15 years under various organizational schemes before going
bankrupt;
The attempt of the Los Angeles Archdiocese to sell off an under-enrolled
Catholic boys' high school without anticipating the reactions of the staff
or the Latino alumni, ninety percent of whom had gone on to college and
successful careers-the Archdiocese reversed its decision after eighteen
months of controversy and in the face of a well coordinated "Save Cathedral
High" campaign; and
The divergent motivations of steel mill owners when contrasted with those
of their employees and the communities in which the mills operated-workers
and communities established new quasi-governmental institutions which could
take over mills by exercising powers of eminent domain with the stated purpose
of protecting the jobs, and communities, dependent on the mills (pp. 31-40).
These cases share some characteristics identified by Meyer and Zucker: "anticipated
future benefits, multiple goals, and decline" (p.80). Anticipated future
benefits mean that some people involved believe that continuing the organization,
even in the face of disappointing performance or long-term decline, may
still bring benefits in the future. Competing groups contest how to frame
the benefits (and decline) and to whom the benefits accrue, which leads
to multiple goals. In the case of the closing steel mills, for example,
stockholders would view closing uncompetitive mills as a sound business
decision while communities viewed the same news as "bombshells"
(p. 40). Naturally, given the divergent ways in which parties measure success,
whether or not an organization is even "failing" and in a state
of decline becomes a matter of opinion. For example, in the case of the
high school, measured financially it had ceased to pay its way, but measured
educationally it continued to be an outstanding success story. Who gets
to name the measures of performance?
To the three common qualities identified by Meyer and Zucker, I humbly submit
three more. First, none had an explicit and public process for closing down
operations. In the incorporation papers for organizations the initial founders
must describe what they will do in the case the organization has to dissolve.
When an organization is one week old, this can be answered with a phrase;
when it is several decades old, more is needed.
Second, three of the four cases had a wide range of people who advocated
for a right to participate in the decision, even though they were not "owners."
These people often recruited other people of wider social or political influence
to the cause of keeping the organization operating. Meyer and Zucker called
these secondary interests "dependent actors" (p. 24) and offered
this description of this role:
Workers, who receive solely wage or salary compensation; the
community, which receives only the side benefits of firm operation in the
form of employment, purchasing power, and access to goods and services provided
by the firm; and the organizations using the firm's products as inputs,
requiring the firm's services, or selling their products to the firm, are
in a position of dependency and therefore have the motivation to support
the maintenance of the firm apart from its efficiency and resulting performance.
This motivation is relatively constant, hence it weakens the relation of
performance to persistence (p. 93-94).
Third, the people initially identifying the organization as being in decline
failed to understand the emotional-symbolic connections others had with
the organization. Organizations take on significant symbolic meaning for
people whose lives they have previously touched, as well as for current
dependent actors, as can be seen in the Herald Examiner and Catholic high
school cases. Failing to understand this connection leaves those seeking
to suspend operations baffled by the emotion and energy people put into
keeping the organization "alive." They also misstep significantly
right at the beginning of their effort to close by not informing dependent
actors well enough in advance and by not making sure people can find a way
to ritualistically say good-bye to-while somehow remaining emotionally in
touch with-their "symbol" (Bridges, 1991: 31). They unwittingly
step in to the quagmire of changing behaviors and expectations about large-scale
operations, which I will turn to next.
Large-Scale Decisions
The political scientist Paul Schulman makes a case in Large-Scale Policy
Making (1980) that large-scale projects need special handling, almost a
new way of being conceptualized. The way forward, or back, may not be smooth
and continuous. He notes that
large-scale policy pursuits are beset by organizational thresholds
or "critical mass" points closely associated with both their initiation
and subsequent developmentThe large-scale policy objective typically confronts
psychological, technological, organizational, and administrative barriers
over which it must "leap" discontinuously if it is to establish
and sustain itself (p.28).
And I will add that these discontinuous leaps reappear when a large-scale
project needs to reverse itself and shut down operations. Essentially, large-scale
projects generate formal and informal agreements among lots of people, who
then participate in doing the work of the project. The Nobel Laureate economist
Kenneth Arrow has neatly summarized the dilemma:
The problem is that agreements are typically harder to change
than individual decisions. When you have committed not only yourself but
many others to an enterprise, the difficulty of changing becomes considerable.
If it is done on a conscious level, we have all the formalities involved
in persuading others to change their minds. What may be hardest of all to
change are unconscious agreements, agreements whose very purpose is lost
in our minds (1974: 28).
The discontinuities exist externally as well as within individual decision
makers-and within involved decision making groups. A change, even a small
one, begins to feel to the decision maker, and well may be interpreted by
others, as a personal error and failure, though he or she or they decided
and acted in good faith for the benefit of the organization. This results
in overprotection and rigidity in decision-making processes by those with
power because even small changes come to mean admissions of error (Schulman,
p. 50). Paradoxically, persisting in the error, sometimes at enormous costs,
becomes preferred by people identified with the decision to admitting to
a mistake in judgment. Persisting in folly, as Barbara Tuchman points out,
displays a certain lack of courage:
Persistence in error is the problem. there is always the freedom
of choice to change or desist from a counter-productive course if the policy-maker
has the moral courage to exercise it. He [or she] is not a fated creature
blown by the whims of Homeric gods. Yet to recognize error, to cut losses,
to alter course, is the most repugnant option in government (p. 383).
Section Three: Can Folly be Prevented?
Can folly be prevented? Maybe not, but its incidence might be reduced. Barbara
Tuchman thinks that individual leaders need better and longer training,
and that the public needs to peer into their character before allowing them
to assume the reins (puns with reigns) of office (p. 384). What can we do
while waiting for this enlightened educational policy, and the enlightened
populace which will nurture it? Plenty, I think. I will call my recommendations
anticipatory conditions.
The leading anticipatory condition must be the continuous practice of wide-ranging
discussions of an organization's values and moral beliefs. This keeps decision
makers conscious of how their actions might support or undermine reputed
values. Clearly, the example provided by Johnson & Johnson shows the
value of talking about morals. It is noteworthy that moral issues raised
by Senator Fulbright and Arthur Schlesinger did not get discussed by Kennedy's
advisors before the Bay of Pig invasion in 1962 (Janis, 1972: 157).
Anticipatory conditions two through six are the active, positive rephrasings
of the significant antecedents Schafer and Crichlow identified. These will
be listed later.
One of the fascinating features of groupthink is how partial the decision
makers thinking becomes, perhaps as a way to cope with their anxiety by
limiting the scope of what they would think about. Neustadt and May in Thinking
in Time (1986) have identified this tendency to "ignore whatever seems
not to fit and to define the problem as one calling for solutions they [politicians]
have ready" (p. 235). Tuchman called this trait wooden-headedness,
and commented that for Philip II of Spain "no experience of the failure
of his policy could shake his belief in its essential excellence" (Tuchman:
7).
Neustadt and May provided several remedies for this malaise. My anticipatory
condition number seven is Neustadt and May's suggestion that decision makers
make time lines of their own organization, of their competitors' organizations,
and of the current situation. They recommend leaving nothing of note out,
and weaving into it the details of personal and organizational history (Neustadt
and May: 238).
To balance this literal mapping of the history of the people and issues
before the group, they suggest making deliberate use of analogies: that
is, making a worksheet which lists possible historical analogies to what
is perceived as the current challenge, and then note where the analogy has
likenesses-and also note the differences-to the current problem (p. 273).
A spontaneous example of this from Robert Kennedy avoided a pre-emptive
strike during Cuban Missile Crisis: he called a surprise attack on Cuba
a "Pearl Harbor in reverse" (emphasis added, Janis, 1972: 157).
A ninth anticipatory condition will also be taken from Neustadt and May:
decision makers need to think through their assumptions about cause and
effect. For example, when deliberating on whether to approve the Bay of
Pigs invasion, Kennedy or his aides, needed only to have written on a yellow
the completion to this sentence: "For the objective of bringing Castro
down, a landing at the Bay of Pigs is the best option because" Even
if they had sidestepped basic beliefs ("truths"), they would have
encountered key "ifthen" presumptions. Those could have been tested-or
discarded-on sight (p. 238). Janis points out, with Arthur Schlesinger,
that some of their assumptions about the military operations might have
been challenged by simply looking at a map of Cuba after the plan was changed
from a landing at Trinidad to a landing at the Bay of Pigs:
Schlesinger acknowledges that he and the others attending the
White House meetings simply overlooked the geography of Cuba: "I don't
think we fully realized that the Escambray Mountains lay 80 miles from the
Bay of Pigs, across a hopeless tangle of swamps and jungle." This oversight
might have been corrected if someone in the advisory group had taken the
trouble to look a map of Cuba (1972: 29).
My tenth, and last, anticipatory condition is deciding before acting how
and when one will want to voluntary stop the activity. This is to be done
for every option under discussion. In escalating situations, where tempting
possibilities and disregarded dangers lurk at every turn, a pre-decision
as to when to stop can prevent great loss (Of Auctions, 1989: Staw, 1989).
Of course, decision makers must always seriously consider doing nothing.
But action not required now may be required later. This means setting up
a monitoring system for detecting the needed information. (see Neustadt
and May, p. 238). To summarize my ten anticipatory conditions recommendations:
1. Host on-going discussions about the moral and ethical values
of the organization.
2. Create a tradition of impartial leadership.
3. Create a tradition of methodical procedures.
4. Keep the estimation of the group limited and underestimated. Stay humble.
5. Practice tolerance for ambiguity as the group struggles to make sense
of all of the available information. Available information includes the
feelings, hunches, and metaphorical reasoning of members of the group.
6. Continually strive to see the situation in new ways in order to avoid
becoming closed-minded.
7. Make a time line of the history of the group, of any opposing groups,
and of the situation. Weave in personal details.
8. Make deliberate use of analogies by listing them and how they are similar-and
different-to the current situation.
9. Probe thinking for cultural, personal, and institutional assumptions
about cause and effect.
10. Discuss for every suggested decision option how the group would want
to voluntarily stop.
Of course, these are just suggestions. There exists no royal road. As Neustadt
and May put it after discussing their time-line:
This bring us to the absolute frontier of our considerations
and experiments up to now. We have no more to offer than a general attitude,
a cast of mind, an outlook, not a method(p. 237).
Final Words
Despite the positive tone of my recommendations, I could add many more.
This means to me that the question of knowing how and when to stop remains
an open one. If I can assert certainty about anything I have learned on
this topic, it is that all proposals must be put forward as provisional
and subject to later, or better, periodic review and revisions. Public discourse
about decisions and large-scale efforts may need to always include reminders
of the provisionalness of our planning, and specifically point out the types
of new data or changes in circumstances which would trip a major reconsideration.
I also have learned that large-scale changes involves the changing of a
lot of minds, so continuous and open dialogue and evaluation of both the
criteria and data of performance will pay dividends later.
Knowing how and when to stop will be different for every specific instance.
One map cannot be drawn for how to transverse this territory, for each instance
has different players, obstacles, histories, and anticipated futures. Still,
as Kenneth Arrow points out:
There are moments in history when we simply must act, fully
knowing our ignorance of possible consequences, but to retain our full rationality
we must sustain the burden of action without certitude, we must always keep
open the possibility of recognizing past errors and changing course (p.
29).
Though I began my researches for this paper with characteristic enthusiasm,
believing that, at last, I would be able to learn from the mistakes of others
to distill for myself a few "rules" for knowing when and how to
stop, I leave this project with many more questions and plausible lines
of research.
Bibliography
Texts
Arrow, Kenneth. (1974). The Limits of Organization. New York: Norton.
Bennis, Warren. (1989). Why Leaders Can't Lead. San Francisco: Jossey-Bass.
Bennis, Warren. (1993). An Invented Life: Reflections on Leadership and
Change. Reading, MA: Addison-Wesley.
Bergler, Edmund. (1967). The Psychology of Gambling. In Herman, Robert,
D,, editor, Gambling. New York: Harper and Row.
Bridges, William. (1991). Managing Transitions: Making the Most of Changes.
Reading, MA: Addison-Wesley.
Bruteau, Beatrice. (1979). The Psychic Grid: How We Create the World We
Know. Wheaton, IL: The Theosophical Publishing House.
Campbell, Andrew, and Nash, Laura L. (1992). A Sense of Mission: Defining
Direction for the Large Corporation. Reading, MA: Addison-Wesley.
Dauten, Dale A. (1980). Quitting: Knowing When to Leave. New York: Walker
and Company.
Farnham, Alan. (1994). It's a Bird! It's a Plane! It's a Flop! Fortune,
May 2, 1994, pp. 108-110.
Ferarnside, W. Ward, and Holther, William B. (1959). Fallacy: The Counterfeit
of Argument. Englewood Cliffs, NJ: Prentice-Hall.
Herek, G. M., Janis, L, and Huth, P. (1987). Decision making during international
crisis: Is Quality of Process Related to Outcome? In Journal of Conflict
Resolution, 31:203-226. Cited in Schafer and Crichlow (1996).
Janis, Irving L. (1972). Victims of Group Think. Boston: Houghton Mifflin.
Janis, Irving L., and Mann, Leon. (1977). Decision Making. New York: Free
Press.
Keynes, John Maynard. (1977, originally 1932). The World's Economic Outlook.
In Desaulniers, Louise, editor. Highlights from 125 Years of the Atlantic.
No City: Atlantic Subscriber Edition, pp. 334-338.
Kolb, Deborah M., and Bartunek, Jean M., editors. (1992). Hidden Conflict
in Organizations: Uncovering Behind-the-Scenes Disputes. Newbury Park, CA:
Sage Publications.
la Brecque, Mort. (1980). On Making Sounder Judgments: Strategies and Snares.
Psychology Today, June 1980, pp. 33-42.
Leonard, Linda S. (1990). Witness to the Fire: Creativity and the Veil of
Addiction. Boston: Shambhala.
Lord, John, with Wold, Jeffrey. (1992(, Song of the Phoenix: The Hidden
Rewards of Failure. Stockbridge, MA: Berkshire House.
Meyer, Marshall, and Zucker, Lynne G. (1990). Permanently Failing Organizations.
Newbury Park, CA: Sage Publications.
Neustadt, Richard E., and May, Ernest R. (1986). Thinking in Time: The Uses
of History for Decision-Makers. New York: The Free Press.
Of Auctions, Dilemmas, and Bloodletting: Model of Escalation Behaviour.
(1989). The Lancet December 23/30, 1989:1487-1488. Failing
Pavlov, Ivan P. (1966, originally 1932). Essay on the Physiological Concept
of The Symptomatology of Hysteria. In Marks, Robert W., editor. (1966).
Great Ideas in Psychology. New York: Bantam, pp. 353-375.
Poole, Marshall, and Hirokawa, Randy Y. (1986). Communication and Group
Decision Making. Newbury Park, CA: Sage Publications.
Schafer, Mark, and Crichlow, Scott. (1996). Antecedents of Groupthink: A
Quantitative Study. In Journal of Conflict Resolution, 40(3):415-435.
Schulman, Paul R. (1980). Large-Scale Policy Making. New York: Elsevier.
Slovic, Paul, Fischhoff, Baruch, and Lichtenstein, Sarah. (1980). Risky
Assumptions. Psychology Today, June 1980, pp. 44-48.
Staw, Barry M. and Ross, Jerry. (1989). Understanding Behavior in Escalation
Situations. Science 246(10/13/89):216-220. PDE/Failing
Teger, Allan I. (1980). Too Much Invested to Quit. New York: Pergamon.
Thompson, James C., Jr. (1977 originally 1968). How Could Vietnam Happen:
An Autopsy. In Desaulniers, Louise, editor. Highlights from 125 Years of
the Atlantic. No City: Atlantic Subscriber Edition, pp. 558-567.
Ting-Toomey, Stella, editor. (1994). The Challenge of Facework: Cross Cultural
and Interpersonal Issues. Albany, NY: State University of New York.
Tuchman, Barbara. (1984). The March of Folly: From Troy to Vietnam. New
York: Knopf.
Valentin, Erhard K. (1994). Anatomy of a Fatal Business Strategy. Journal
of Management Studies, 31(3):359-382.
Tapes
Bohannan, Paul, et al. (1993). Face Saving Devices as a Mode of Conflict
Settlement. Portland, OR: National Conference on Peacemaking and Conflict
Resolution.
Internet
Forsyth, Donelson R. (1996). Lecture Notes: Decision Making in Groups. Available
12/23/96 at http://www.vcu.edu/hasweb/psy/psy341/decide.html.
Francis, Donald P. (1996). A Voice for the Public's Health. Available 12/23/96
at http://www.gene.com/AE/TSN/SS/bkgnd_paper.html.
Healing Mission: William Foege, MPH Harvard 1995, Reflects on a Career Devoted
to Reducing Human Suffering. (1996). Available on 12/23/96 at http://biosun1.harvard.edu/hphr/94fall/healing.html.
Johnson, Jo-Ann. (1996). Workers Avert Plant Closings, Save Jobs. Available
12/28/96 at http://www.americannews.com/story/ans-14.html.
NOTES
1 Comparing this version with the 1945 original, shows that a fifth
category has been deleted: "Our third responsbility is to our management.
Our executives must be persons of talent, education, experience, and ability.
They must be persons of common sense and full understanding" (Collins
and Porras, 1994: 58). I have not seen any references to why it was deleted.
2 Note that without this preparation and deep organizational understanding
of its ideology, Johnson & Johnson might have reacted as Bristol-Myers
did to a problem of Excedrin being tampered with in the Denver area: Bristol-Myers
recalled tablets only in Colorado and did not alert the public. (Collins
and Porras: 81)