The years between the first hydrogen bomb
tests and the Limited Test Ban Treaty in 1963 saw more than just
increased anxiety about the effects of nuclear testing on weather. They
also saw increased interest in large-scale, purposeful environmental
modification. Most climate modification enthusiasts spoke of increasing
global temperatures, in the hopes that this would increase the quantity
of cultivated land and make for fairer weather. Some suggested
blackening deserts or snowy areas, to increase absorption of radiation.
Covering large areas with carbon dust, so the theory went, would raise
temperatures. Alternatively, if several hydrogen bombs were exploded
underwater, they might evaporate seawater and create an ice cloud that
would block the escape of radiation. Meteorologist Harry Wexler had
little patience for those who wanted to add weather and climate
modification to the set of tools in man’s possession. But by 1958 even he
acknowledged that serious proposals for massive changes, using nuclear
weapons as tools, were inevitable. Like most professional
meteorologists, in the past he had dismissed the idea that hydrogen
bombs had affected the weather. But with the prospect of determined
experiments designed to bring about such changes, he warned of “the
unhappy situation of the cure being worse than the ailment.”
Whatever
one might have thought about the wisdom of tinkering with the weather
in peacetime, the manipulation of nature on a vast scale for military
purposes seemed to be a perfectly legitimate application of scientific
knowledge. While planning a total war against the Soviet Union, every
avenue begged for exploration. Let’s explore how the scientific advisors
of America’s key allies in NATO saw the alliance fighting in the future.
Numerous ideas for creating catastrophic events through natural
processes were presented, especially using hydrogen bombs as triggers.
In these discussions, held as early as 1960, top scientists debated the
fundamental environmental question — can humans have a long-term effect
on the global environment?
The
desire for novel military technology seemed especially urgent by the
early 1960s. Although officially part of the International Geophysical
Year, the Soviet Union’s launch of Sputnik in October 1957 had clear
military ramifications. Not only did it begin the space race but it also
took the arms race to a new stage that included communications
satellites and intercontinental ballistic missiles. The launch of
Sputnik made the world seem smaller and made the most far-fetched
visions of the future seem possible. The gee-whiz, Buck Rogers feel of
the immediate postwar years returned. But this wave of technological
enthusiasm was darker, because instead of coming on the tide of a war
victory, it came as a foreboding new competition. For years the
Americans had been preparing for the missile age, gathering data on the
atmosphere and on the earth’s gravity over the poles. The Soviets
clearly had kept the pace. Sputnik served as a justification for a vast
array of projects to use scientific knowledge to tamper with nature on a
large scale.
Reinforcing the sense of urgency, President
Eisenhower’s special committee on weather modification submitted its final
report in January 1958, just months after Sputnik’s launch. The
committee’s chairman, retired Navy Captain Howard T. Orville, said at a
press conference that he suspected that the Soviets already had begun a
large, secret program on weather control. Despite routine dismissals of
the idea throughout the decade by meteorologists, the high-level
committee ranked weather control ahead of hydrogen bombs and satellites
in military significance. Orville urged the government to support
research on controlling large-scale weather systems, not just
rainmaking. He further suggested that finding ways to manipulate the heat
balance between the sun and earth might be the key to weather and
climate control. The earth already had been heated up by man’s efforts,
by introducing carbon dioxide into the atmosphere through the burning of
fossil fuels. This carbon dioxide helped to trap the heat and create,
as the New York Times put it, a “greenhouse effect.” It might be possible
to harness this greenhouse effect. “If such steps are feasible,”
journalist John Finney reported, “then New York City might be put under a
few hundred feet of ice or a few hundred feet of water depending on
whether the temperature was raised or lowered.”
Rumors spread
quickly about scientists in the United States and Soviet Union
experimenting with unprecedented tools for controlling nature. Were the
Soviets planning to dam the Bering Strait? Were the Americans able to
steer storms? Naysayers pointed out that meteorologists could not even
predict naturally occurring weather, so how could anyone control it? One
author opined in the New York Times, “For would it not be foolish for
anyone to talk of controlling an intricate piece of apparatus until he
knew precisely how it worked?” After the report of Eisenhower’s special
committee was made public, scientists in allied countries received
strange, sheepish letters from their defense establishments, asking if
the latest rumors about American research could be true. For example, a
British Air Ministry scientific advisor, E. V. Truefitt, presented his
countryman, oceanographer George Deacon, with “one or two questions
which have come up in odd conversations.” He called them “wild cat”
ideas that he did not really take seriously, yet they appeared to be in
discussion in the United States. Despite his instinct that they could
not possibly be real, he felt obligated to run them by a competent man
of science.
One of the ideas was to melt the polar ice cap by
exploding nuclear weapons on it, thus raising the global sea level. The
Soviets might be considering it, so the rumor went, to drown cities in
the United States and Western Europe. Another idea was to change ocean
currents or temperatures to interfere with an enemy’s climate and food
production. Truefitt had no idea how assess an ocean-initiated climate
change, but he had made a rough calculation to determine what was needed
to melt the polar ice cap. He believed that it would take about a
million tons of fissile material to melt enough to raise sea level by 30
feet. “This is a large amount of fissile material whichever way you look
at it,” he wrote to Deacon, “and consequently my guess is that it is not
the kind of project that even the Americans would embark on under the
influence of Sputniks.”
Desperate to find “weapon of the future”
The
truth was that the immediate post-Sputnik years had a peculiar air,
both of desperation and of opportunity. Doors were wide open to a range
of technological possibilities. Nearly anything that was technically
feasible made it to the highest levels of discussion. For starters, that
meant revisiting the questions surrounding biological, chemical, and
radiological weapons. But it also sparked discussion of the ambitious,
the horrendous, and the quirky. Like wildcatters exploring for oil,
American scientists grasped desperately around them, striving to find the
next weapon of the future.
There were several post-Sputnik efforts
to push the limits of the “possible,” to explore exotic ideas that
might prove decisive 5, 10, or 20 years into the future. Some actions to
direct this scientific work were high profile and public. President
Eisenhower created a science advisory committee to guide the course of
American technology and ensure that the Americans did not fall behind
the Soviet Union. This President’s Science Advisory Committee (PSAC)
also existed to rein in some of the wilder ideas, to avoid wasteful
spending. Other brain trusts, often dominated by physicists with
expertise in nuclear affairs, sprang up behind closed doors to advise
military establishments. One of these was “JASON,” an elite group of
scientists who got together during the summer months to assess major
scientific and technological problems of military significance. Paid by
government contract through several different bodies throughout its
existence, “the Jasons,” as they called themselves, were drawn from the
cream of civilian academic science. Despite their outsider status, the
Jasons gained the respect and trust of officials in the Defense Department
and the armed services, and their advice often revolutionized military
thinking during the nuclear era.
Sputnik did not just spark new
scientific projects, however. It also revolutionized military strategy,
making it grimmer than ever. The American and Soviet air forces realized
they were going to have to rethink the basic notion of national
vulnerability. No longer could the Air Force’s Strategic Air Command
count on scrambling bombers and flying them over the North Pole. Most of
the war’s damage would have been done before bombers left the Western
Hemisphere.
More than that, as a secret National Academy of
Sciences group advised the Air Force in 1958, the range of possible
wars soon would expand exponentially. Conflicts were going to become both
more total and more limited at the same time. On the one hand, the
United States was losing its ability to incapacitate enemy forces. In
practice that meant that the most attractive targets all over the world
would be centers of population — cities — rather than armies or
airfields. Making an effective attack against enemy military forces seemed
a dwindling prospect in an era when missiles could be put into hardened
silos, mobile rocket units, or submarines patrolling the oceans.
Cities, by contrast, would be ripe for plucking. “Weapon yields,
delivery accuracies, and force level requirements for city destruction
are modest,” these scientists concluded, while attacking heavily
fortified bunkers would require large and accurate payloads. That meant
that finding ways of maximizing civilian death would assume an even
greater importance than it already had.
On the other hand, nuclear
parity would make full-blown conflict less likely, meaning that all of
the armed services would have to reorient themselves back to
conventional warfare. As RAND Corporation game theorists had long
feared, the atomic bomb was a wasting asset—and the window of
opportunity to “win” decisively in a war against the Soviet Union had
passed. By the late 1950s, the new orthodoxy in strategic thinking
accepted that the Soviet Union was committed to avoiding a nuclear
holocaust and that it intended to encourage “brushfire” wars instead.
Small wars like those in Malaya and Korea would become more common. As
the 1960s dawned, military strategists wondered about the fate of
Vietnam, which the French had failed to hold. By treaty the country had
split between North and South. Should the communist North Vietnamese
invade, would the Americans consider using nuclear weapons?
Some may
have argued that nuclear bombs were America’s answer to the human
population imbalance against, say, people in China or Southeast Asia.
But new studies at RAND had dismissed this possibility, showing that
nuclear weapons would be ineffective against guerrilla forces in
Southeast Asia and would visit enormous collateral damage upon friendly
population centers. So the military would need to let go of President
Eisenhower’s preferred strategy of massive retaliation as America’s
basic posture.
The Air Force would have to stop relying on
aircraft designed purely to deliver nuclear weapons. Instead, it would
need to find ways of fighting men, tanks, rockets, and airplanes—all
without nuclear weapons. A decade earlier the Navy had bitterly opposed
the Air Force’s claims that the era of aircraft carriers and battleships
had ended. Now it seemed that the Navy had been right. The new
conventional wisdom, which President Kennedy (a former Navy man) soon
would establish as the doctrine of “flexible response,” was that the
nations with the greatest range, flexibility, and cleverness in weapons
systems would stand strongest. This meant conducting research on weapons
at various levels of destruction up to and including nuclear bombs and
being creative about their uses.
It also meant combining modes of
warfare across scientific disciplines. Geophysical and biological
knowledge might be united, for example, in developing dispersal
mechanisms for pathogens. In trying to achieve large area coverage, one
might fall back on cloud-seeding techniques—with the important difference
that the “seeds” would not be silver iodide to cause rain but pathogens
to spread disease far and wide. For example, certain phenomena in air
masses, such as “Polar Outbreaks” (thrusts of cold air from the poles
toward the equator), seemed to have great potential for such seeding,
especially given the Soviet Union’s meteorological vulnerability from
the north.
Military research embraces science-fiction
The
post-Sputnik national pall of gloom encouraged American scientists to
explore unorthodox weapons, and they left no stone unturned. The U.S.
military forged ahead with research on weapons using radiation, particle
beams, nuclear energy, and kinetic energy. The Army Chemical Corps even
investigated the use of lysergic acid diethylamide (LSD) and cannabis
as non-lethal, incapacitating agents. The National Academy of Sciences
noted this approvingly in 1958 and suggested that the Air Force begin
administering LSD to airmen as soon as possible, to judge whether to add
it to the arsenal of chemical weapons.
With so many wide-ranging
ideas being vetted, NATO allies worried that the Americans were moving
in too many directions at once. It was fine to support science in the
United States and to speak grandly about possibly controlling forces of
nature—but which ideas could be incorporated into actual NATO war plans?
In 1960 NATO members agreed to convene a special group of scientists
and military leaders to assess the long-term prospects of war. They
wanted to know what would really be feasible by the 1970s, and what was
just science fiction.
This kind of science forecasting was not just
a matter of intelligent people guessing the future. By 1960 it had a
distinguished history of shaping policy, particularly in some parts of
the American military establishment. The Air Force, for example,
understood in the 1950s that much of its strength relied on continuous
research and development (R&D). Toward the end of World War II,
General Henry “Hap” Arnold, commander of the then-Army Air Force,
famously said that “for twenty years the Air Force was built around
pilots, pilots, and more pilots. . . . The next twenty years is going to
be built around scientists.” Throughout the Cold War, such brain
trusts—in think tanks like RAND, secret groups like JASON, and many
others—exercised a remarkable influence on policies.
When NATO
tried, in 1960, to estimate the next 10 to 15 years of weapons, it
enlisted the leadership of Theodore von Kármán, the grand old man of
science forecasting. By then he was 79 years old. Born in Hungary, von
Kármán had been one of the world’s foremost experts in aerodynamics. He
even had helped the Austrian military design aircraft during the First
World War. In 1929 he came to the United States to head up an
aeronautical laboratory at Caltech, helping to kick-start the aviation
industry in southern California. Acting as scientific advisor to United
States air forces during World War II, von Kármán had initiated a
long-term study of air power that amassed some of the best brains in
physics and aeronautics. The resultant report, Where We Stand, became a
road map for postwar air power research. In subsequent years, von Kármán
repeated this process with other studies, and in fact he chaired the
1958 secret committee advising the Air Force, under the auspices of the
National Academy of Sciences. In 1960 he embarked on a study that would
be the capstone of his long career: NATO’s attempt to grasp the future
face of battle over the entire earth.
Battle: earth
Known
simply as the Von Kármán Committee, the new group included the chief
scientific advisor of each national defense organization in the United
States, Britain, Canada, France, and West Germany. With several working
groups of scientists under them, they ran the gamut of new weapons in an
era of “total war.” They included the typical range of military
subjects, including aircraft, weaponry, and ships. But they also delved
deeply into the implications of the global physical environment,
particularly in light of the extraordinary size of thermonuclear
weapons, the global reach of ballistic missiles, and the extent of
global monitoring begun during the International Geophysical Year.
The
buzzword of the IGY had been “synoptic.” Taken literally, it meant
observing more than one place at the same time—viewing together. The
IGY’s concept was to take a huge number of observations, spread out over
a variety of geophysical disciplines and geographic areas, all within
an 18-month period. Doing so would provide a portrait of the earth that
was
more true and comprehensive than anything ever attempted.
The
Von Kármán Committee adopted the word “synoptic” too, but applied it to
weapons. Weapons of a “synoptic scale” meant control and domination of
whole physical systems. In military shorthand, the word synoptic called
to mind vastness, encompassing large portions of the earth—or perhaps
all of it. The IGY had brought this idea into military planners’ field of
vision. But while the IGY was concerned with synoptic-scale
measurement, NATO was concerned with synoptic-scale manipulation.
Once
they began to meet, the members of the Von Kármán Committee realized
that they all agreed on at least one thing: the global observations
initiated in the IGY would have to continue indefinitely. The geophysical
factors of modern war involved knowledge of an operational
environment—in other words, how would the sea, land, or air affect troops
and ships? NATO forces needed to be able to operate in any kind of
environment. If it was on planet Earth, NATO should be prepared to fight
there and win.
In fact the U.S. armed services already were
developing environment-specific training centers to give American forces
mastery of three classes of extreme conditions: polar, desert, and
jungle. Given that the northern polar region was “the only large
uncommitted area lying between the world’s strongest antagonists,” polar
operations weighed heavily on defense planners’ minds. Already polar
and Arctic training centers existed at locations in Greenland, Canada,
and in the state of Alaska. The United States also operated a desert
warfare center in Yuma, Arizona. Still needed were centers approximating
Mediterranean conditions and tropical ones.
To take advantage of
the apparent shrinkage of the earth due to ballistic missiles, NATO
advisors also pointed out the need to revolutionize the field of
geodesy—earth measurement. Mapmakers relied on data taken from a variety
of oceanic or terrestrial expeditions, sometimes decades or more old.
No one had seen the earth from space, much less taken accurate
measurements based on satellites. Intercontinental ballistic missiles
would require precision. But NATO literally did not know where the
Soviet Union was. “On a world wide scale, we are not sure of the
position of North America in relation to the Eurasian continent.”
Knowledge of anything in the Southern Hemisphere was even less accurate.
The only decent data came from the Americas, Western Europe, Japan, and
some of the former and current European colonial territories. The
Soviets could target the West with accuracy, but the West could not do
the same. Any kind of exact targeting of the Soviet Union would prove
impossible before satellites could take comprehensive measurements. In
the meantime, constant earth measurement from the air would prove
essential. Fortunately, international scientific projects were providing
that data.
The IGY had convinced scientists and military planners
of the usefulness of synoptic data collection. If done in real time, or
close to it, data collection could be automated and collected over a
large territory, perhaps even globally. Individual scientists might
never analyze the vast amounts of data, but the data could be fed into
computers in order to monitor and
predict environmental conditions.
Already the Americans were working on an anti-submarine warfare
“environmental prediction system.” It collected oceanographic
information—to estimate sonar performance—and combined it with
meteorological information to predict future oceanographic conditions.
Had
the members of the Von Kármán Committee been military historians, there
is little doubt about what they would have cast as the “decisive
moment” in the history of global strategy. Time and again they called to
mind the changes brought about by the advent of earth-orbiting
satellites. It would prove to be, they believed, a dividing line between
military eras. It promised total monitoring of the global environment, a
vision of the future that was pervasive across the range of sciences
and military operations. By 1970, these NATO advisors predicted,
scientists would be able to identify and track thunderstorms as they
occurred all over the entire earth and to keep the earth’s radiation
under constant surveillance. Old charts would be discarded, in favor of a
constantly refreshing set of data beamed down from the heavens.
Automated data systems would be necessary to achieve accuracy of
measurement and improved forecasting. As the committee put it: “The
concept of inaccessible geographical areas is no longer
valid—observations over enemy-held, oceanic and uninhabited areas are as
easily made as elsewhere.” Reliance on existing charts and data,
collected laboriously by error-prone humans, rarely uniform from country
to country, seemed archaic. New methods of continuous, uniform data
collection of the oceans, land, and space would provide the kind of
mastery of the global environment that the Von Kármán committee
envisioned.
Climate change as warfare
Aside
from this unprecedented ability to forecast conditions and improve
global accuracy, the NATO science advisors also predicted ambitious,
large-scale manipulation of the environment. The brass ring of military
geophysics was weather control. Scientists already had achieved modest
results in increasing rainfall or dissipating fogs. But these successes
required optimal conditions and certainly could not be projected over a
large area or from a long distance. But what about climate control?
In
a 1956 Fortune article, mathematician John von Neumann had suggested
that militaries would be able to make large-scale changes to climate. He
pointed out various ways to alter oceans and seas. One was to blanket
ice sheets with blackening agents, to absorb more light and melt them.
If it could be done to Greenland, its ice sheet alone would raise sea
levels by about 10 feet “and cause great discomfort to most world
ports.” Another scheme was to divert the Gulf Stream, which would
severely change the climate of Northern Europe. Still another idea was
to dam the Bering Strait. Such alterations would have clear, long-term
effects on world climate. And these changes seemed possible. Reflecting on
von Neumann’s predictions, the NATO group believed that an
extraordinary tool lay in the hands of military planners: the hydrogen
bomb. “It is perhaps true,” the committee concluded, “that means
presently within man’s reach could be employed so as to alter global
climate for long periods.”
Given the later controversy about the
role of carbon dioxide in inducing global climate change, the focus on
the hydrogen bomb might seem surprising. But the reason for this was
simple. Advised by physicists, the defense establishments of NATO’s
strongest members believed that in order for “synoptic scale” weapons to
be feasible, man had to achieve physical power that was comparable to
nature’s power. The only tool that seemed likely to provide that was the
hydrogen bomb. Although professional meteorologists had insisted that
hydrogen bomb tests had not created the extreme winters of 1954, 1958,
and 1962, these military advisors were less adamant. They knew that the
energies of nature were vast, but felt they might be shaped by man. It
seemed that the Soviets were working hard on the problem. Canadian
scientists repeated the oft-heard rumor that the Soviets were planning
large-scale manipulation of the oceans, along with drastic modification
of climate, by damming up the Bering Strait. The Canadians reasoned:
surely the Russians had in mind the use of nuclear bombs?
NATO scientists found the prospects of such power over nature intriguing. They called it
environmental warfare.
“This kind of warfare has the peculiarity that it could look like our
image of nuclear war, or could be so subtle that the ‘weapons’ and
‘battles’ are hard to identify.” The enemy might undertake a vast
engineering project to change the climate of a whole region, “leading
gradually to economic ruin and loss of strength.” This could be done
even without declaring war.
Once again ecological vulnerability
emerged as a crucial area in need of study for military purposes. The
NATO science advisors did not yet understand their true vulnerability to
what they called “living weapons.” But new data were coming in. Since
the late 1950s, American engineers had planned to use thermonuclear
explosions to excavate a harbor in Alaska—a project dubbed “Plowshare.”
Beforehand they put together what today might be called an environmental
impact statement and discovered that the effect on the Eskimos’ diet
might not be as negligible as originally assumed. For this and other
reasons, the project was scrapped.
But that knowledge had been
useful for military thinking. Scientists had traced the pathway of
radioactivity through the food chain. NATO scientists now used the
example of the Eskimos’ ecosystem to argue for more advanced knowledge
of ecological warfare. Within that ecosystem, Eskimos lived
interdependently with seals, otter, fish, caribou, and plankton. If the
plankton were all killed, an Eskimo’s ecological community would be
utterly destroyed. “At best he would have to move,” the group pointed
out. “At worst he would die.” This kind of thinking could be tailored to
particular regions: “The people of Asia depend on rice and a very few
other crops. Something like a lethal rice-rust or blight could make life
in Asia much more difficult and perhaps untenable.”
As a weapon
system, ecological links went further than killing—they also promised
biological coercion. Destruction of the enemy need not be the goal.
Getting rid of plankton, for example, would make the Eskimos’ entire
food system collapse and force them to be entirely dependent on food
supplied from outside the region. To achieve this, toxic agents “may be
developed to attack essential links in various ecological chains.” The
aim would be to shape an existing interdependent web along new lines,
“to force the ecology to accept dependence on some crop or animal which
cannot live at all in the homeland.” Doing this would put the victim in
an extremely disadvantageous position, “leading to a gradual loss of
power and position and inevitable vassalage.”
Von Kármán died
shortly after the first of his committee reports was completed.
As
colleagues remembered his contributions to aeronautics and to scientific
advising, his death lent the committee’s findings an extraordinary amount
of authority within NATO. The reports had the air of a final act of
service; the chairman’s passing only augmented the committee’s
importance. With Von Kármán gone, the reports themselves were a
foreboding, Cassandra-like vision of the future that military planners
could ignore only at their peril. This was especially true of subjects
that the committee felt it did not yet understand fully.
Environmental warfare becomes real
Environmental
warfare had captured the imagination of the committee but the results
had been unsatisfying. It seemed in keeping with the direction of
science—toward global, synoptic-scale activities. Yet it was unclear how
it might shape weaponry. The experience of the Von Kármán Committee
established “environmental warfare” as a distinct concept, and it was
not long before NATO reconvened the members to look into the subject
more fully. They realized that there were commonalities between the work
on geophysics and the ongoing work on radiological, biological, and
chemical weapons. Both involved alterations to the natural world with
potentially devastating human consequences. Military technology seemed
on the verge of an unprecedented ability to tap the forces of nature on a
massive scale.
Thus in late 1962, NATO summoned scientists and
military planners to Paris to hammer out what might legitimately come
out of “environmental warfare” and what the long-term consequences might
be. The man who tried to fill Von Kármán’s shoes was another Hungarian,
nuclear physicist Edward Teller, who joined the group as a “special
advisor.” Known widely as the father of the hydrogen bomb, Teller
already was deeply committed to using nuclear explosions for massive
earthmoving projects, such as the construction of harbors. He also saw
great potential in developing novel uses of nuclear weapons in wartime.
Along with Teller, committee members were drawn from national defense
establishments and from the U.S. Advanced Research Projects Agency
(ARPA).
The central question almost always remained the same: were
natural forces susceptible to human influence on a large, even global,
scale? In methodical fashion, these military planners broke down
environmental warfare into distinct spheres of possibility,
corresponding with the layers of the earth and its atmosphere as it
extended into space: lithosphere and hydrosphere (land and oceans),
troposphere (lower atmosphere), stratosphere and ionosphere (upper
atmosphere), and exosphere (outer space). Some of the earlier “wildcat”
ideas were quickly dispensed with as impractical, such as using hydrogen
bombs to melt the polar ice caps. But other wildcat ideas were
feasible, particularly using nuclear weapons as triggers for tsunamis in
the oceans, or for altering the weather.
One only had to open a
newspaper to see what natural catastrophes could accomplish. In 1958, in
Alaska’s Lituya Bay, there was a landslide so powerful that it carried
the energy equivalent to a one-kiloton explosion. In May 1960, a wall of
water smashed the Chilean coast over a stretch of several hundred
miles, with wave heights of 5.5 to 13.5 meters. The Chilean earthquake
sent storm waves across a large area of the Pacific at speeds in excess
of 400 miles per hour. Even as far away as Hawaii, low-lying areas were
flooded. Thousands of Chileans were killed, and millions were left
homeless. Reporters described the relentless devastation:
The
quakes went on for all of the week, demolishing or damaging thousands
of homes and other buildings, and burying some small communities under
landslides. Whole villages were swept away by tsunamis as high as
twenty-four feet. The quakes were so violent that mountains disappeared,
new lakes were formed and the earth’s surface dropped as much as 1,000
feet in twenty-five miles. The worst quake, last Sunday, released energy
of 240 megatons, equal to that of 1,200 atomic bombs of the type dropped
on Hiroshima and far more than the 174 megatons released by all the
nuclear explosions to date.
Noting deaths all over
the Pacific Rim, the New York Times reported that the Chilean earthquake
“gave tragic testimony that in this age of the conquest of the atom and
of triumphs in outer space man is still helpless against the vast and
still largely unpredictable forces that frequently go berserk in his
immediate environment—hurricanes, volcanoes and earthquakes.”
NATO
saw it differently. Environmental cataclysms could become part of the
alliance’s arsenal, with the help of a well-placed nuclear explosion.
The cascading effects of energy release from the existing instabilities
of nature could be, quite literally, earth shattering. The power over
nature was tempting: “The large engineering capability which is provided
by multi-megaton nuclear weapons might open up the possibility of
changing the course of ocean streams which are known to affect climate
and continents.” Narrow straits could indeed be dammed up, as some
feared the Soviets planned for the Bering Straits. Peninsulas could be
turned into islands, changing the patterns of water flow and mixing. With
enough nuclear bombs, the sea floor in some areas might be reconfigured
entirely.
Even weather control seemed poised to make a quantum
leap forward with the nuclear bomb as a tool. “Real weather control,”
NATO scientists argued, “would mean control of synoptic scale
disturbances—the centers of high pressure and low pressure found on the
daily weather maps.” Such large-scale systems seemed inherently
susceptible to influence, despite the huge energies required to do it.
The sun imparted energy into the air masses constantly, but only some of
it became kinetic energy. Most of the energy was stored, ready to be
released. The results could be quite violent, as in the case of
cyclones. A relatively small release of energy—say, a nuclear bomb—could
trigger a much larger release of natural energy.
One reason that
such widespread and even long-term changes in the earth’s systems seemed
feasible—at least in theory—was the growing realization of how serious
an effect humans already were having upon the upper atmosphere. High in
the sky, major effects seemed well within NATO’s grasp. Nuclear
explosions could create electron clouds some 70–90 kilometers up,
disrupting high-frequency communication. One of the leading researchers
on electron cloud disruption, Jerome Pressman, had been advising the
U.S. Army Signal Corps, the Air Force, and ARPA on this subject for
years. He told the rest of the environmental warfare committee that even
a single nuclear burst could disrupt long-distance communication over a
stretch of a thousand kilometers. If nuclear weapons were exploded in
the atmosphere as a defense against incoming missiles, the range of this
electron cloud would be vast indeed. High-frequency communication
equipment and long-distance radar systems might be rendered useless.
Out
in space—the exosphere—NATO saw great promise in the radiation belts
that American and Soviet satellites had measured during the
International Geophysical Year. The Van Allen belts were actually giant
regions of charged particles trapped by the earth’s magnetic field. They
were sources of intense, persistent radiation that endangered any
equipment or living thing in space. Although the Van Allen belts were
natural phenomena, similar belts could be created artificially by
exploding a nuclear weapon at an altitude of at least 400 kilometers.
Large bombs at even higher altitudes would create an extraordinarily
powerful radiation environment in space. The belts would cloak the
earth, challenging any exit or entrance by missile, satellite, or
spacecraft. Because the belts would be= trapped by the earth’s magnetic
field, there would be holes in the radiation cloak at the north and south
geomagnetic poles.
Whoever controlled these entry points would have
comparatively easy access to space. That would make the poles even more
important as strategic regions.
In fact, manipulation of the Van
Allen belts already had begun. In 1958 the United States discovered that
its high-altitude tests of “small kilotonnage” had created electron
shells around the earth, about 60 miles thick. Because the operation in
which these tests occurred had been dubbed “ARGUS,” the creation of the
shell became the “ARGUS effect.” Just a few months prior to these NATO
meetings, the United States detonated an even larger explosion at high
altitude—the “Starfish” experiment. As Edward Teller reported, “this is
the first time that the Argus effect was demonstrated on a really big
scale.” An immense number of electrons were caught in the earth’s
magnetic field and “are forming now a new Van Allen belt greater in
electron density than any of the known Van Allen belts.” He confided that
the electrons had damaged the solar cells in
American satellites.
“Why not just drop a bomb?”
Despite
their fascination with these weapons, the committee members struggled
to overcome the possibilities that defied the logic of nuclear warfare.
The military significance of triggering natural catastrophes was not
readily obvious. “If the weapon can be exploded a few miles offshore, it
can probably be delivered on, or close to, the target itself, and a far
larger proportion of the energy available would be expended on the
target and not on long lengths of unimportant coast line.” The same
argument could be made against any effort to influence the flow of ocean
currents and thus modify the world’s climate. Why not just drop a bomb
on a city? It seemed more logical.
On the other hand, there might
be great value in environmental devastation in a total war. NATO
advisors had already moved beyond “cities” as targets and had begun to
imagine much larger swathes of territory. Aside from the blast and
radioactive contamination, thermonuclear bombs could have wide-ranging
horrific consequences. Disruptions of dams and levees would lead to
widespread flooding. Drowning and starvation would result, posing a
serious threat to those who managed to survive the bombs.
The most
ghastly environmental threat was the prospect of large-scale fire. In
Whole World on Fire (2004), Lynn Eden has written that military planners
routinely ignored the consequences of huge firestorms caused by a
nuclear explosion’s thermal radiation. She suggests that this led
nuclear strategists to underestimate the catastrophic effects of nuclear
explosions throughout the Cold War. While war plans typically focused on
blast effects, not everyone ignored the totality of death and
destruction from fires. Some military planners considered it part of
environmental warfare. In the early 1960s, scientists and military
planners at the highest levels of NATO faced a stomach-churning analysis
that cast them as a way of arming the countryside against the enemy
even when his cities were destroyed.
These fires would
instantaneously ignite a huge area due to the explosion’s initial
thermal radiation, regardless of blast effects. Rather than just use
bombs directly against cities, one could explode a large bomb of about
100 megatons high in the atmosphere, at about 80 kilometers. Doing so
would maximize the amount of thermal radiation that would reach the
earth. Such radiation would ignite flammable material instantly, over an
area of nearly a million square kilometers. As a point of comparison,
the largest recorded forest fire in the United States occurred in 1871 in
Wisconsin and Michigan, which claimed 1,683 lives and spread over
15,000 square kilometers. Setting fire to forests, in an area of a
million square kilometers, would pose intractable problems to an enemy.
Outside the bombed-out cities, the countryside would provide no shelter,
no food, and no hope of survival.
A fire from thermal radiation
would differ from a typical forest fire because it would not need to
spread—instead, the whole area would go up in flames at the same time.
Oxygen would rapidly deplete, leaving any survivors suffocating to death.
It would be impossible to run from it. Rushes of air would create
firestorms with “strong winds of up to hurricane force,” far more intense
than the deadly firestorms created in German and Japanese cities during
World War II. Edward Teller guessed that the energy released in a fire
would exceed that of the nuclear explosion, roughly the equivalent of a
thousand megatons. “This is the most violent and wide-spread
environmental change which can be expected from a nuclear attack,” he
said. If total war were the goal, fires from thermal radiation could
achieve it on a continental scale.
These discussions, recorded for
posterity in NATO meeting minutes, have a surreal feel to them.
Scientists argued about whether hydrogen bombs were more effective as
triggers of vast environmental events, or if they should just be dropped
directly on their targets. Scientists quibbled over the extent of
damage from a fire-raising weapon. Some doubted, for example, that
hurricane-force winds would ensue. It was difficult to argue with the
conclusion, however: “The immediate result would be beyond all
experience.” But some insisted that it would only “likely” be beyond all
experience.
Such intellectualized detachment from human
experience reached new heights when the long-term ecological
consequences of nuclear weapons were imagined. The NATO group recognized
that using nuclear weapons in this way might have severe consequences
for the earth in the long run. But while acknowledging that the effect on
weather and climate might be significant, scientists had little data
with which to generate specific predictions. As for the devastation of
the land, NATO was confident that a succession of vegetation “would
sooner or later re-establish itself, and over a few decades there would
be some ecological recovery.”
The only thing not in doubt in these
discussions was that maximizing human death was the principal goal.
Which was better, Teller and his colleagues asked—drowning villages
along the coast, igniting the countryside with thermal radiation, or
simply laying waste a city? Should humans be contaminated through the
food chain, or beat into submission through ecological dependence? While
praising the ingenuity of these wildcat ideas, Teller’s own preference
was to bomb cities. If death and devastation were the goals, he
reasoned, why not keep it simple? Mammals, including humans, were more
sensitive to radioactivity than insects, seed plants, or bacteria. It
made little sense to attempt to contaminate man through these less
susceptible organisms when the bomb would do the trick. “Thus the most
economic way to attack populations with nuclear radiation,” the
committee concluded, “is to do so directly rather than through some
element of their surroundings.”
For many in NATO, looking at the
world as a zero-sum game between the nuclear-armed United States and the
Soviet Union, environmental warfare seemed like an inefficient sideshow.
As interesting as ocean manipulation and weather control might be,
nuclear explosions would be required to produce them. In that case,
presumably a real war would have begun, and the enemy could be bombed
directly without resorting to exotic methods such as these. Even in the
case of biological, radiological, and chemical weapons, changing the
environment would be a more circuitous route than attacking directly.
In
trying to imagine uses of environmental weapons, military analysts
working with NATO confronted the same question that has stood at the
center of environmental issues ever since: can human actions have
long-lasting, detrimental consequences upon the earth? As an advocate of
peacetime nuclear testing, Teller had reason to minimize the long-term
impacts of human action, particularly nuclear fallout. He spoke at
length to the committee about how some scientists had exaggerated these
effects, and his point of view prevailed.
The NATO committee concluded
that the danger of sickness and disease from contamination “are no worse
than the other hazards which would have to be faced by the survivors of
a nuclear war.” As for the long-term genetic effects upon future
generations, the committee toed the protesting line that the ultimate
effects on future generations could not be predicted with certainty.
Nevertheless,
some on the committee were convinced that humans were capable of making
large alterations to the environment. Throughout the Von Kármán reports
were repeated references to unpredictable consequences of human action
on the atmosphere. Increasing or decreasing the ozone concentration in
the atmosphere was certainly possible, altering the amount of
ultraviolet light reaching the earth. Deliberate creation of an ozone
hole might confuse surveillance systems; deteriorate aircraft materials
such as rubber, plastic, glass; and harm humans and crops. Less
purposeful might be the introduction of chemicals from rocket fuel or
other sources, resulting in “large inadvertent changes” in atmospheric
properties.
NATO concluded its assessment of environmental warfare
with a warning that major changes might already be under way. “Much of
the military planning of today assumes that the earth’s atmosphere will
remain substantially as it is,” it wrote.
Reprinted from
“Arming Mother Nature: The Birth of Catastrophic Environmentalism” with permission from Oxford University Press USA. Copyright © Oxford University Press 2013