Nuclear Risk
Authors
Philip Taubman
News Type
Commentary
Date
Paragraphs

The Obama administration seems ready to resuscitate relations with Russia, including by renewing nuclear-arms-reduction talks. Even before the inaugural parade wound down, the White House Web site offered up a list of ambitious nuclear policy goals, with everything from making bomb-making materials more secure to the eventual abolition of nuclear weapons.

That's welcome news, but for such goals to be realized, the White House will need to be prepared to reimagine and reshape the nuclear era and, against strong opposition, break free from cold war thinking and better address the threats America faces today.

George W. Bush actually started down this road. He reached an agreement with the Kremlin in 2002 to cut the number of operational strategic warheads on each side to between 1,700 and 2,200 by the year 2012, a two-thirds reduction. Washington is likely to reach that goal ahead of schedule. President Bush's efforts were propelled by the Nuclear Posture Review - a periodic reassessment of nuclear forces and policies - in December 2001. While still grounded in the belief that nuclear weapons are the silver bullets of American defense, the review let a little daylight into the nuclear bunker by acknowledging that nuclear-weapons policy had to be readjusted to deal with rapidly changing threats. Soon, however, the president's initiatives were overshadowed by the wars in Afghanistan and Iraq, his administration's absorption with the threat of terrorism and the gradual breakdown in relations with Russia.

President Bush's agreement with Moscow, which was built upon weapons reductions made by Presidents Bill Clinton and George H. W. Bush, is President Obama's starting point. But rather than settle for the next level - 1,000 active weapons seems to be the likely goal - the White House should reconsider the entire superstructure of nuclear-weapons strategy. This won't be easy. The mandarins of the nuclear establishment remain enthralled by elaborate deterrence theories premised on the notion that the ultimate defense against a variety of military threats is a bristling nuclear arsenal.

It's true that America's nuclear weapons still offer the hope of deterring attacks from countries like North Korea and, if it soon goes nuclear, Iran. But it is hard to imagine how they would dissuade a band of elusive, stateless terrorists from making a nuclear bomb and detonating it in New York, Washington or Los Angeles.

One provocative road map for moving away from nuclear deterrence comes from a quartet of cold war leaders - Henry Kissinger and George Shultz, former secretaries of state; William Perry, a former secretary of defense; and Sam Nunn, a former chairman of the Senate Armed Services Committee. Two years ago, they bridged their ideological differences to call, improbably, for the abolition of nuclear weapons, and they proposed a series of interim steps to reduce nuclear dangers, stop the spread of bomb-making materials and lay the groundwork for a nuclear-free world.

Even the quartet recognizes that "getting to zero" will be exceedingly difficult. But the issue today isn't whether the elimination of nuclear weapons is feasible. That's a distant goal.

An achievable immediate goal should be to cut the United States' and Russia's nuclear stockpiles down to the bare minimum of operational warheads needed to backstop conventional forces. As long as these two countries have far and away the most nuclear weapons, Washington looks hypocritical when it lectures other nations about the size of their arsenals or their efforts to develop nuclear weapons.

There's reasonable disagreement among experts about the minimum number of nuclear weapons the United States and Russia should maintain. The more emphasis you put on nuclear deterrence, the more potent you think the arsenal should be. And the more you want to engage the world in arms reduction and prevent proliferation, the more you consider radical cuts. To bring the number down below 1,000 would require determined presidential leadership.

The president's determination will be measured by how effectively he makes the case for Senate ratification of the 1996 Comprehensive Nuclear Test Ban Treaty. Leading scientists say that technological advances over the past decade have erased doubts about whether an international monitoring system can detect and locate underground tests outlawed by the treaty. The scientists also say that the United States has the technical expertise and tools to maintain the effectiveness of its nuclear weapons without underground testing, as has been successfully demonstrated since the United States stopped testing in 1992.

Ratification of the test-ban treaty would help build momentum for a 2010 review of the Nuclear Nonproliferation Treaty, the increasingly frail 1968 accord aimed at limiting the spread of nuclear weapons and eventually eliminating them. American leadership is essential to reinvigorating the treaty and buttressing nonproliferation efforts. The best way to avoid nuclear terrorism is to prevent terrorists from acquiring the highly enriched uranium needed to make the simplest nuclear bomb.

Listening to the discussion at a recent nuclear-weapons conference in Washington, I felt as though I had slipped back in time to the cold war and its arcane, often surreal debates about waging nuclear war and the doctrine of mutual assured destruction. It's heartening to see President Obama and his national-security team promising to elevate nuclear-weapons policy and free it from the shibboleths of cold war nuclear theology. Now they must put their words into action.

All News button
1
-

Abstract: The expected increasing world energy demand makes it necessary for us to seriously and urgently study the questions of global warming due to greenhouse gas effect emissions and the depletion of fossil resources. This clearly means producing more energy, while emitting a minimum amount of CO2, and keeping the costs under control and acceptable for the user.

A growing number of prospective studies thus envision that nuclear energy, because it is carbon-free, will play an important and essential role in the world energy mix of the 21st century.

However, the increased use of nuclear power to generate electricity brings with it, threats to regional and global security - specifically, increased risks of nuclear weapon proliferation and nuclear terrorism: nuclear power reactors inevitably produce plutonium as a by-product, plutonium that could be used by countries or terrorist groups to fabricate nuclear weapons. Several states still have not signed the NPT, while others have not clarified their real intentions.

Even though this aspect should by no means be neglected, the issue of nuclear energy expansion should be examined globally, accounting for the context, the current needs, as well as all kinds of concerns.

The context is the one described above, characterized by growing energy demand and climate change: nuclear energy is unanimously recognized as a solution well adapted to such a context. Its overall assets are numerous, it is a clean and competitive source of energy, which has very good safety records, with more improvements to come, it contributes to security of energy supply. All these assets should not be swept away for reasons solely linked to proliferation concerns. As a matter of fact, intensive works are being carried out, to improve even more nuclear energy's track record, by ensuring its sustainability: waste minimisation, increased safety, competitiveness, economy of uranium resources, resistance to nuclear proliferation, and application to fields wider than shear electricity production.

Jacques Bouchard is Special Adviser to the Chairman of the French Atomic Energy Commission (CEA). In 2006, he was appointed Chairman of the Generation IV International Forum (GIF) for 3 years.

Born in 1939, Jacques Bouchard holds an engineering degree from the "Ecole Centrale de Paris", and specialized in reactor physics.

Mr. Bouchard joined the CEA in 1964 and became Head of the Experimental Physics unit in 1973, then head of the Nuclear Engineering Department in 1975. In that capacity, the work he conducted was mainly in support of pressurized water reactor technology, and he also led studies in physics for fuel cycle applications.

In 1982, he became head of the Fast Neutron Reactor Department in Cadarache. In 1990, he was appointed head of the CEA's Nuclear Reactor Division, then, from 1994 to 2000, he became the Director of CEA's military application division.

From 2000 to 2004, he was in charge of the entire nuclear energy sector in CEA.

Since 2005, he is Special Adviser to the Chairman of the CEA.

Jacques Bouchard was also the President of the French Nuclear Energy Society from 2001 to 2003 and professor at the reknown "Ecole des Mines de Paris". He has serve on the board of directors of several companies working in the nuclear field, and he is member of many advisory committees to national and international nuclear organizations.

If you would like to be added to the email announcement list, please visit https://mailman.stanford.edu/mailman/listinfo/stsseminar 

Reuben W. Hills Conference Room

Jacques Bouchard Special Adviser to the Chairman of the French Atomic Energy Commission Speaker
Seminars
Authors
Varun Rai
News Type
News
Date
Paragraphs

Executive Summary

Carbon capture and storage (CCS) is a promising technology that might allow for significant reductions in CO2 emissions. But at present CCS is very expensive and its performance is highly uncertain at the scale of commercial power plants. Such challenges to deployment, though, are not new to students of technological change. Several successful technologies, including energy technologies, have faced similar challenges as CCS faces now. In this paper we draw lessons for the CCS industry from the history of other energy technologies that, as with CCS today, were risky and expensive early in their commercial development. Specifically, we analyze the development of the US nuclear-power industry, the US SO2-scrubber industry, and the global LNG industry.

We focus on three major questions in the development of these analogous industries. First, we consider the creation of the initial market to prove the technology: how and by whom was the initial niche market for these industries created? Second, we look at how risk-reduction strategies for path-breaking projects allowed the technology to evolve into a form so that it could capture a wider market and diffuse broadly into service. Third, we explore the "learning curves" that describe the cost reduction as these technologies started to capture significant market share.

Our findings suggest that directly applying to CCS the conventional wisdom that is prevalent regarding the deployment and diffusion of technologies can be very misleading. The conventional wisdom may be summarized as: "Technologies are best deployed if left in the hands of private players"; "Don't pick technology winners" or "Technology forcing is wrong"; and "Technology costs reduce as its cumulative installed capacity increases". We find that none of these readily applies when thinking about deployment of CCS.

Hero Image
ccs logo
All News button
1
Paragraphs

Carbon capture and storage (CCS) is a promising technology that might allow for significant reductions in CO2 emissions. But at present CCS is very expensive and its performance is highly uncertain at the scale of commercial power plants. Such challenges to deployment, though, are not new to students of technological change. Several successful technologies, including energy technologies, have faced similar challenges as CCS faces now. In this paper we draw lessons for the CCS industry from the history of other energy technologies that, as with CCS today, were risky and expensive early in their commercial development. Specifically, we analyze the development of the US nuclear-power industry, the US SO2-scrubber industry, and the global LNG industry.

We focus on three major questions in the development of these analogous industries. First, we consider the creation of the initial market to prove the technology: how and by whom was the initial niche market for these industries created? Second, we look at how risk-reduction strategies for path-breaking projects allowed the technology to evolve into a form so that it could capture a wider market and diffuse broadly into service. Third, we explore the "learning curves" that describe the cost reduction as these technologies started to capture significant market share.

Our findings suggest that directly applying to CCS the conventional wisdom that is prevalent regarding the deployment and diffusion of technologies can be very misleading. The conventional wisdom may be summarized as: "Technologies are best deployed if left in the hands of private players"; "Don't pick technology winners" or "Technology forcing is wrong"; and "Technology costs reduce as its cumulative installed capacity increases". We find that none of these readily applies when thinking about deployment of CCS.

Through analyzing the development the analogous industries, we arrive at three principal observations:  

  • First, government played a decisive role in the development of all of these analogous technologies. Much of the early government role was to provide direct backing for R&D work and demonstration projects that validated the technological concepts. For example, the US government directly supported for over two decades most of the basic science and engineering research in both SO2 scrubbers and nuclear power. Most of the demonstration projects were significantly underwritten by government as well; the Japanese government was the principal backer of LNG technology through its promises to buy most of the world's LNG output over many years. Direct government support created the niche opportunities for these technologies.
  • Second, diffusion of these technologies beyond the early demonstration and niche projects hinged on the credibility of incentives for industry to invest in commercial-scale projects. In each of the historical cases, government made a shift in its support strategy as the technology diffused more widely. In the early phase (when commercial uncertainties were so high that businesses found it extremely risky to participate in more than small, isolated projects) success in achieving technology diffusion required a direct role for government. But as uncertainties about the technology's performance reduced and operational experience accumulated, direct financial support became less important, and indirect instruments to lower commercial risk rose in prominence. Those instruments included tax breaks, portfolio/performance standards, purchase guarantees, and low-interest-rate loans linked to specific commercial-scale investments. It is conceivable that such incentives could have been supplied by non-governmental institutions, such as large firms or industry associations, but the three analogs point strongly to a governmental role-perhaps because only government action was viewed as credible. (In the United States, many of the key decisions to support new technologies were crafted at the state level, such as through rate base decisions to allow utilities to purchase nuclear plants.)
  • Third, the conventional wisdom that experience with technologies inevitably reduces costs does not necessarily hold. Risky and capital-intensive technologies may be particularly vulnerable to diffusion without accompanying reductions in cost. In fact, we find the opposite of the conventional wisdom to be true for nuclear power in the US (1960-1980) and global LNG (1960-1995). Costs increased as cumulative installed capacity increased. A very rapid expansion of nuclear power plants in the US around 1970 led to spiraling costs, as the industry had no chance to pass lessons from one generation of investment to the next-a fact evident, for example, in the failure to standardize design and regulation that would allow firms to exploit economies of scale. For natural gas liquefaction plants, costs stayed high for decades due to a market structure marked by little competition among technology suppliers and the presence of a single dominant customer (Japanese firms organized by the Japanese government) willing to pay a premium for safety and security of supply. The same attributes that allowed LNG to expand rapidly-namely, promises of assured demand made credible by the singular backing of the Japanese state-were also a special liability as the technology struggled to compete in other markets. The experience with SO2 scrubbers was more encouraging-costs declined fairly promptly once industrial-scale investment was under way. But that happened only after sufficient clarity on technological performance and capability of FGD systems had been established. What followed was a strict performance standard-in the form of a government mandate, imposed by environmental regulators-that effectively picked FGD as a technology winner. The guaranteed market for FGD led to serious investment, innovations, and learning-by-doing cost reductions. We do not argue that this technology-forcing approach was economically efficient but merely underscore that rates of diffusion of FGD technology akin to what is imagined for CCS technology today were possible only under this technology-forcing regulatory regime.

As CCS commercialization proceeds, policymakers must remain mindful that cost reduction is not automatic-it can be derailed especially by non-competitive markets, unanticipated shifts in regulation, and unexpected technological challenges. At the same time, there may be some inevitable tradeoffs, at least for a period, between providing credible mechanisms to reduce commercial risk, such as promises of assured demand for early technology providers, and stimulating market competition that can lead to lower costs. History suggests that government-backed assurances are essential to creating the market for capital-intensive technologies; yet those very assurances can also create the context that makes it difficult for investors to feel the pressure of competition that, over successive generations of technology, leads to learning and lower costs.

We are also mindful that our history here-drawn on the experience of three technologies that have been successful in obtaining a substantial market share-is a biased one. By looking at successes we are perhaps overly prone to derive lessons for success when, in fact, most visions for substantial technological change actually fail to get traction.

All Publications button
1
Publication Type
Working Papers
Publication Date
Journal Publisher
Program on Energy and Sustainable Development, Working Paper #81
Authors
Varun Rai
David G. Victor
Mark C. Thurber
-

North Korea is widely regarded internationally as a long-term threat to regional peace and stability and to the international nuclear nonproliferation regime. Conservatives support the use of sanctions and other pressures to counter these threats, while liberals hope that reasonable efforts at accommodation will persuade the Pyongyang regime to change course. Professor Lankov maintains that neither approach will work. He urges a new approach, based on engagement but with the long-term goal of inducing regime change from within North Korea.

Andrei Lankov, a historian of Korea and one of the world's top experts on North Korea, is an associate professor at Kookmin University in Seoul. He received undergraduate and graduate degrees at Leningrad State University and attended Kim Il Sung University in Pyongyang, North Korea. He has also taught at Leningrad State University and Australian National University. Lankov is the author of many books in English, Korean, and Russian, including From Stalin to Kim Il Sung: The Formation of North Korea, 1945-1960; Crisis in North Korea: The Failure of De-Stalinization, 1956; North of the DMZ: Essays on Daily Life in North Korea; and The Dawn of Modern Korea. Among his most recent articles is "Staying Alive: Why North Korea Will Not Change," which appeared in the March/April 2008 edition of Foreign Affairs. He is also a columnist for the Chosun Ilbo and Korea Times in South Korea.

This special seminar is supported by a generous grant from Koret Foundation.

Philippines Conference Room

Andrei Lankov Associate Professor, Kookmin University, Seoul, Korea Speaker
Conferences
Paragraphs

In the aftermath of a terrorist attack political stakes are high: legislators fear being seen as lenient or indifferent and often grant the executive broader authorities without thorough debate. The judiciary's role, too, is restricted: constitutional structure and cultural norms narrow the courts' ability to check the executive at all but the margins. The dominant "Security or Freedom" framework for evaluating counterterrorist law thus fails to capture an important characteristic: increased executive power that shifts the balance between branches of government. This book re-calculates the cost of counterterrorist law to the United Kingdom and the United States, arguing that the damage caused is significantly greater than first appears. Donohue warns that the proliferation of biological and nuclear materials, together with willingness on the part of extremists to sacrifice themselves, may drive each country to take increasingly drastic measures with a resultant shift in the basic structure of both states.

“Laura Donohue’s sophisticated and complex analysis of counterterrorism law in Britain and the United States warns of the risks to fundamental individual rights when democracies establish counterterrorist regimes. Although governments frame their initiatives in terms of a choice between security and freedom, Donohue challenges this logic. Loss of liberty is not necessarily balanced by gain in safety. Compromises intended to be temporary turn out to be permanent. Leaders and citizens of democracies would be well advised to heed this pointed and timely warning.”

- Martha Crenshaw, Senior Fellow, Center for International Security and Cooperation (CISAC), Stanford University

An ambitious argument against the "Security or Freedom" framework, which is the dominant paradigm for thinking about counterterrorist law. The first book to compare the history of both British and American counterterrorist law. Argues that counterterrorist law is a danger to the rights central to liberal democracy: life, liberty, property, privacy and free speech.

All Publications button
1
Publication Type
Books
Publication Date
Journal Publisher
Cambridge University Press
Authors
Number
ISBN-13: 9780521605878
Subscribe to Nuclear Risk