Nuclear Weapons
NNSA Needs to Refine and More Effectively Manage Its New Approach for Assessing and Certifying Nuclear Weapons
Gao ID: GAO-06-261 February 3, 2006
In 1992, the United States began a unilateral moratorium on the testing of nuclear weapons. To compensate for the lack of testing, the Department of Energy's National Nuclear Security Administration (NNSA) developed the Stockpile Stewardship Program to assess and certify the safety and reliability of the nation's nuclear stockpile without nuclear testing. In 2001, NNSA's weapons laboratories began developing what is intended to be a common framework for a new methodology for assessing and certifying the safety and reliability of the nuclear stockpile without nuclear testing. GAO was asked to evaluate (1) the new methodology NNSA is developing and (2) NNSA's management of the implementation of this new methodology.
NNSA has endorsed the use of the "quantification of margins and uncertainties" (QMU) methodology as its principal method for assessing and certifying the safety and reliability of the nuclear stockpile. Starting in 2001, Los Alamos National Laboratory (LANL) and Lawrence Livermore National Laboratory (LLNL) officials began developing QMU, which focuses on creating a common "watch list" of factors that are the most critical to the operation and performance of a nuclear weapon. QMU seeks to quantify (1) how close each critical factor is to the point at which it would fail to perform as designed (i.e., the margin to failure) and (2) the uncertainty that exists in calculating the margin, in order to ensure that the margin is sufficiently larger than the uncertainty. According to NNSA and laboratory officials, they intend to use their calculations of margins and uncertainties to more effectively target their resources, as well as to certify any redesigned weapons envisioned by the Reliable Replacement Warhead program. According to NNSA and weapons laboratory officials, they have made progress in applying the principles of QMU to the assessment and certification of nuclear warheads in the stockpile. NNSA has commissioned two technical reviews of the implementation of QMU. While strongly supporting QMU, the reviews found that the development and implementation of QMU was still in its early stages and recommended that NNSA further define the technical details supporting the implementation of QMU and integrate the activities of the three weapons laboratories in implementing QMU. GAO also found important differences in the understanding and application of QMU among the weapons laboratories. For example, while LLNL and LANL both agree on the fundamental tenets of QMU at a high level, they are pursuing different approaches to calculating and combining uncertainties. NNSA uses a planning structure that it calls "campaigns" to organize and fund its scientific research. According to NNSA policies, campaign managers at NNSA headquarters are responsible for developing plans and high-level milestones, overseeing the execution of these plans, and providing input to the evaluation of the performance of the weapons laboratories. However, NNSA's management of these processes is deficient in four key areas. First, NNSA's existing plans do not adequately integrate the scientific research currently conducted across the weapon complex to support the development and implementation of QMU. Second, NNSA has not developed a clear, consistent set of milestones to guide the development and implementation of QMU. Third, NNSA has not established formal requirements for conducting annual, technical reviews of the implementation of QMU at the three laboratories or for certifying the completion of QMU-related milestones. Finally, NNSA has not established adequate performance measures to determine the progress of the three laboratories in developing and implementing QMU.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-06-261, Nuclear Weapons: NNSA Needs to Refine and More Effectively Manage Its New Approach for Assessing and Certifying Nuclear Weapons
This is the accessible text file for GAO report number GAO-06-261
entitled 'Nuclear Weapons: NNSA Needs to Refine and More Effectively
Manage Its New Approach for Assessing and Certifying Nuclear Weapons'
which was released on February 3, 2006.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to the Subcommittee on Strategic Forces, Committee on Armed
Services, House of Representatives:
February 2006:
Nuclear Weapons:
NNSA Needs to Refine and More Effectively Manage Its New Approach for
Assessing and Certifying Nuclear Weapons:
GAO-06-261:
GAO Highlights:
Highlights of GAO-06-261, a report to the Subcommittee on Strategic
Forces, Committee on Armed Services, House of Representatives:
Why GAO Did This Study:
In 1992, the United States began a unilateral moratorium on the testing
of nuclear weapons. To compensate for the lack of testing, the
Department of Energy‘s National Nuclear Security Administration (NNSA)
developed the Stockpile Stewardship Program to assess and certify the
safety and reliability of the nation‘s nuclear stockpile without
nuclear testing. In 2001, NNSA‘s weapons laboratories began developing
what is intended to be a common framework for a new methodology for
assessing and certifying the safety and reliability of the nuclear
stockpile without nuclear testing. GAO was asked to evaluate (1) the
new methodology NNSA is developing and (2) NNSA‘s management of the
implementation of this new methodology.
What GAO Found:
NNSA has endorsed the use of the ’quantification of margins and
uncertainties“ (QMU) methodology as its principal method for assessing
and certifying the safety and reliability of the nuclear stockpile.
Starting in 2001, Los Alamos National Laboratory (LANL) and Lawrence
Livermore National Laboratory (LLNL) officials began developing QMU,
which focuses on creating a common ’watch list“ of factors that are the
most critical to the operation and performance of a nuclear weapon. QMU
seeks to quantify (1) how close each critical factor is to the point at
which it would fail to perform as designed (i.e., the margin to
failure) and (2) the uncertainty that exists in calculating the margin,
in order to ensure that the margin is sufficiently larger than the
uncertainty. According to NNSA and laboratory officials, they intend to
use their calculations of margins and uncertainties to more effectively
target their resources, as well as to certify any redesigned weapons
envisioned by the Reliable Replacement Warhead program.
According to NNSA and weapons laboratory officials, they have made
progress in applying the principles of QMU to the assessment and
certification of nuclear warheads in the stockpile. NNSA has
commissioned two technical reviews of the implementation of QMU. While
strongly supporting QMU, the reviews found that the development and
implementation of QMU was still in its early stages and recommended
that NNSA further define the technical details supporting the
implementation of QMU and integrate the activities of the three weapons
laboratories in implementing QMU. GAO also found important differences
in the understanding and application of QMU among the weapons
laboratories. For example, while LLNL and LANL both agree on the
fundamental tenets of QMU at a high level, they are pursuing different
approaches to calculating and combining uncertainties.
NNSA uses a planning structure that it calls ’campaigns“ to organize
and fund its scientific research. According to NNSA policies, campaign
managers at NNSA headquarters are responsible for developing plans and
high-level milestones, overseeing the execution of these plans, and
providing input to the evaluation of the performance of the weapons
laboratories. However, NNSA‘s management of these processes is
deficient in four key areas. First, NNSA‘s existing plans do not
adequately integrate the scientific research currently conducted across
the weapon complex to support the development and implementation of
QMU. Second, NNSA has not developed a clear, consistent set of
milestones to guide the development and implementation of QMU. Third,
NNSA has not established formal requirements for conducting annual,
technical reviews of the implementation of QMU at the three
laboratories or for certifying the completion of QMU-related
milestones. Finally, NNSA has not established adequate performance
measures to determine the progress of the three laboratories in
developing and implementing QMU.
What GAO Recommends:
GAO is making five recommendations to the Administrator of NNSA to (1)
ensure that the three laboratories have an agreed-upon technical
approach for implementing QMU and (2) improve NNSA‘s management of the
development and implementation of QMU.
While NNSA raised concerns with some of GAO‘s recommendations, it
agreed that it needed to better manage QMU‘s development and
implementation. NNSA also said that GAO had not given it credit for its
success in implementing QMU. GAO clarified its report to address NNSA‘s
concerns.
www.gao.gov/cgi-bin/getrpt?GAO-06-261.
To view the full product, including the scope and methodology, click on
the link above. For more information, contact Gene Aloise at (202) 512-
3841 or aloisee@gao.gov.
[End of section]
Contents:
Letter:
Results in Brief:
Background:
The QMU Methodology Is Highly Promising but Still in the Early Stages
of Development:
NNSA's Management of the Development and Implementation of QMU Is
Deficient in Four Key Areas:
Conclusions:
Recommendations for Executive Action:
Agency Comments and Our Evaluation:
Appendixes:
Appendix I: Comments from the National Nuclear Security Administration:
Appendix II: GAO Contact and Staff Acknowledgments:
Tables:
Table 1: Nuclear Weapons in the Enduring Stockpile:
Table 2: NNSA Funding for the Scientific Campaigns, Fiscal Years 2001-
2005:
Table 3: NNSA Funding Requests and Projections for the Scientific
Campaigns, Fiscal Years 2006-2010:
Table 4: NNSA Level-1 Milestones Related to the Development and
Implementation of QMU:
Table 5: Primary Campaign Level-2 Milestones Related to the Development
and Implementation of QMU:
Abbreviations:
ASC: Advanced Simulation and Computing:
DOE: Department of Energy:
ICF: Inertial Confinement Fusion:
LANL: Los Alamos National Laboratory:
LLNL: Lawrence Livermore National Laboratory:
NNSA: National Nuclear Security Administration:
Primary: Primary Assessment Technologies:
QMU: quantification of margins and uncertainties:
RRW: Reliable Replacement Warhead:
Science Council: NNSA's Office of Defense Programs Science Council:
Secondary: Secondary Assessment Technologies:
SNL: Sandia National Laboratory:
Letter February 3, 2006:
The Honorable Terry Everett:
Chairman:
The Honorable Silvestre Reyes:
Ranking Minority Member:
Subcommittee on Strategic Forces:
Committee on Armed Services House of Representatives:
In 1992, the United States began a unilateral moratorium on the testing
of nuclear weapons. Prior to the moratorium, underground nuclear
testing was a critical component of the evaluation and certification of
the performance of a nuclear weapon. Confidence in the continued
performance of stockpiled weapons relied heavily on the expert judgment
of weapon designers who had significant experience with successful
nuclear tests. In addition, the training of new weapon designers
depended on continued nuclear testing. In 1993, the Department of
Energy (DOE), at the direction of the President and the Congress,
established the Stockpile Stewardship Program to ensure the
preservation of the United States' core intellectual and technical
competencies in nuclear weapons without testing.[Footnote 1] The
National Nuclear Security Administration (NNSA), a separately organized
agency within DOE, is now responsible for carrying out the Stockpile
Stewardship Program, which includes activities associated with the
research, design, development, simulation, modeling, and nonnuclear
testing of nuclear weapons. The three nuclear weapons design
laboratories--Lawrence Livermore National Laboratory (LLNL) in
California, Los Alamos National Laboratory (LANL) in New Mexico, and
Sandia National Laboratories (SNL) in California and New Mexico--use
the results of these activities to annually assess the safety and
reliability of the nation's nuclear weapons stockpile and to certify to
the President that the resumption of underground nuclear weapons
testing is not needed.
When the moratorium began in 1992, DOE (and subsequently NNSA) faced
several challenges in fulfilling its new mission of stockpile
stewardship. For example, since both expected and unexpected changes
occur as the nuclear stockpile ages, NNSA has become more concerned
with gaining a detailed understanding of how such changes might affect
the safety and reliability of stockpiled weapons. However, unlike the
rest of a nuclear weapon, the nuclear explosive package--which contains
the primary and the secondary[Footnote 2]--cannot be tested simply by
evaluating individual components. Specifically, because the operation
of the nuclear explosive package is highly integrated, nonlinear,
occurs during a very short period of time, and reaches extreme
temperatures and pressures, there are portions of the nuclear explosive
package that cannot be tested outside of a nuclear explosion. In
addition, although the United States conducted about 1,000 nuclear
weapons tests prior to the moratorium, only a few tests were designed
to collect data on uncertainties associated with a particular part of
the nuclear explosive package. As a result, much of the scientific
basis for the examination of an exploding nuclear weapon must be
extrapolated from other phenomena. Finally, since nuclear testing is no
longer available to train new weapons designers, NNSA and the weapons
laboratories are faced with the need to develop a rigorous,
transparent, and explainable approach to all aspects of the weapon
design process, including the assessment and certification of the
performance of nuclear weapons.
To address these challenges, in 1999, DOE established 18 programs--
which it referred to as "campaigns"--six of which were intended to
develop the scientific knowledge, tools, and methods required to
provide confidence in the assessment and certification of the safety
and reliability of the nuclear stockpile in the absence of nuclear
testing. These scientific campaigns include the (1) Primary Assessment
Technologies (Primary), (2) Secondary Assessment Technologies
(Secondary), (3) Advanced Simulation and Computing (ASC), (4) Advanced
Radiography, (5) Dynamic Materials Properties, and (6) Inertial
Confinement Fusion and High Yield (ICF) campaigns. In particular, the
Primary and Secondary campaigns are designed to analyze and understand
the different scientific phenomena that occur in the primary and
secondary stages of a nuclear weapon during detonation. As such, the
Primary and Secondary campaigns are intended to set the requirements
for the computer models and experimental data provided by the other
campaigns that are needed to assess and certify the safety and
reliability of nuclear weapons.
While the campaign structure brought increased organization to the
scientific research conducted across the weapons complex, NNSA still
lacked a coherent strategy for relating the scientific research
conducted by the weapons laboratories to the needs of the nuclear
stockpile and the Stockpile Stewardship Program. Consequently, in 2001,
LLNL and LANL began developing what is intended to be a common
framework for a new methodology for assessing and certifying the safety
and reliability of warheads in the nuclear stockpile in the absence of
nuclear testing.
The Stockpile Stewardship Program is now over 10 years old, NNSA's
campaign structure is in its sixth year, and 4 years have passed since
LLNL and LANL began their effort to develop a new assessment and
certification methodology. As the weapons in the nuclear stockpile
continue to age, and as more experienced weapon designers and other
scientists and technicians retire, NNSA is faced with increased urgency
in meeting the goals of the Stockpile Stewardship Program. Furthermore,
NNSA has recently created an effort, known as the Reliable Replacement
Warhead (RRW) program, to study a new approach to maintaining nuclear
warheads over the long term. The RRW program would redesign weapon
components to be easier to manufacture, maintain, dismantle, and
certify without nuclear testing, potentially allowing NNSA to
transition to a smaller, more efficient weapons complex. NNSA's ability
to successfully manage these efforts will have a dramatic impact on the
future of the U.S. nuclear stockpile and, ultimately, will affect the
President's decision of whether a return to nuclear testing is required
to maintain confidence in the safety and reliability of the stockpile.
In this context, you asked us to evaluate (1) the new methodology NNSA
is developing for assessing and certifying the safety and reliability
of the nuclear stockpile in the absence of nuclear testing and (2)
NNSA's management of the implementation of this methodology.
To evaluate the new methodology NNSA is developing for assessing and
certifying the safety and reliability of the nuclear stockpile in the
absence of nuclear testing, we reviewed relevant policy and planning
documents from NNSA and the three weapons laboratories, including
implementation plans and program plans for the six scientific
campaigns. We focused our work principally on the Primary and Secondary
campaigns because the primary and secondary are the key components of
the nuclear explosive package and because the Primary and Secondary
campaigns are intended to set the requirements for the experimental
data and computer models needed to assess and certify the performance
of nuclear weapons. We also reviewed relevant reports, including those
from NNSA's Office of Defense Programs Science Council, the MITRE
Corporation's JASON panel,[Footnote 3] University of California review
committees for LANL and LLNL, and the Strategic Advisory Group
Stockpile Assessment Team for U.S. Strategic Command. In addition, we
interviewed officials from NNSA headquarters and site offices, as well
as contractors who operate NNSA sites. Our primary source of
information was NNSA's Office of Defense Programs. We also met with
officials at LANL, LLNL, and SNL. Finally, we interviewed nuclear
weapons experts, senior scientists, and other relevant officials
outside of NNSA and the laboratories, including members of NNSA's
Office of Defense Programs Science Council, the JASON panel, University
of California review committees for LANL and LLNL, the Strategic
Advisory Group Stockpile Assessment Team for U.S. Strategic Command,
and the Deputy Assistant to the Secretary of Defense (Nuclear Matters)
for the Department of Defense.
To evaluate NNSA's management of the implementation of its new
methodology to assess and certify the safety and reliability of nuclear
weapons in the absence of nuclear testing, we reviewed relevant NNSA
policy, planning, and evaluation documents, including the Office of
Defense Program's Program Management Manual, campaign program and
implementation plans, contractor performance evaluation plans and
reports, and internal reviews of NNSA management. We also reviewed
contractor planning and evaluation documents, including LANL, LLNL, and
SNL performance evaluation plans and reports. Finally, we met with
campaign managers and other officials at NNSA headquarters and site
offices, LANL, LLNL, and SNL. We performed our work between August 2004
and December 2005 in accordance with generally accepted government
auditing standards.
Results in Brief:
NNSA has endorsed the use of the "quantification of margins and
uncertainties" (QMU) methodology as its principal method for assessing
and certifying the safety and reliability of the existing nuclear
stockpile in the absence of nuclear testing. The QMU methodology
focuses on creating a "watch list" of factors that, in the judgment of
nuclear weapon experts, are the most critical to the operation and
performance of a nuclear weapon. Starting in 2001, LANL and LLNL
officials began developing QMU, which they described as a common
methodology for quantifying how close each critical factor is to the
point at which it would fail to perform as designed (i.e., the margin
to failure), as well as quantifying the uncertainty that exists in
calculating the margin, in order to ensure that the margin is
sufficiently greater than the uncertainty. According to NNSA and
laboratory officials, the weapons laboratories intend to use their
calculations of margins and uncertainties to more effectively target
their resources to either increasing the margin in a nuclear weapon or
reducing the uncertainties associated with calculating the margin. In
addition, they said that QMU will be vital to certifying any redesigned
weapons, such as those envisioned by the RRW program.
NNSA and laboratory officials told us that they have made progress in
applying the principles of QMU to the certification and assessment of
nuclear warheads in the stockpile. However, QMU is still in its early
stages of development, and important differences exist among the three
laboratories in their application of QMU. To date, NNSA has
commissioned two technical reviews of the implementation of QMU at the
weapons laboratories. While strongly supporting QMU, the reviews found
that the development and implementation of QMU was still in its early
stages. For example, one review stated that, in the course of its work,
it became evident that there were a variety of differing and sometimes
diverging views of what QMU really was and how it was working in
practice. The reviews recommended that NNSA take steps to further
define the technical details supporting the implementation of QMU and
integrate the activities of the three weapons laboratories in
implementing QMU. However, NNSA and the weapons laboratories have not
fully implemented these recommendations. Beyond the issues raised in
the two reports, we also found differences in the understanding and
application of QMU among the three laboratories. For example, LLNL and
LANL officials told us that the QMU methodology only applies to the
nuclear explosive package and not to the nonnuclear components that
control the use, arming, and firing of the nuclear warhead. However,
SNL officials told us that they have been applying their own version of
QMU to nonnuclear components for a long time. In addition, we found
that while LLNL and LANL both agree on the fundamental tenets of QMU at
a high level, their application of the QMU methodology differs in some
important respects. Specifically, LLNL and LANL are pursuing different
approaches to calculating and combining uncertainties. While there will
be methodological differences among the laboratories in the detailed
application of QMU to specific weapon systems, it is fundamentally
important that these differences be understood and, if need be,
reconciled, to ensure that QMU achieves the goal of the common
methodology NNSA has stated it needs to support the continued
assessment of the existing stockpile or the certification of redesigned
nuclear components under the RRW program.
NNSA relies on its Primary and Secondary campaigns to manage the
development and implementation of QMU. According to NNSA policies,
campaign managers at NNSA headquarters are responsible for developing
campaign plans and high-level milestones, overseeing the execution of
these plans, and providing input to the evaluation of the performance
of the weapons laboratories. However, NNSA's management of these
processes is deficient in four key areas. First, the planning documents
that NNSA has established for the Primary and Secondary campaigns do
not adequately integrate the scientific research currently conducted
that supports the development and implementation of QMU. Specifically,
a significant portion of the scientific research that is relevant to
the Primary and Secondary campaigns, and the implementation of QMU, is
funded and carried out by a variety of campaigns and other programs
within the Stockpile Stewardship Program. Second, NNSA has not
developed a clear, consistent set of milestones to guide the
development and implementation of QMU. For example, while one key
campaign plan envisions a two-stage path to identify and reduce key
uncertainties in nuclear weapon performance using QMU by 2014, the
performance measures in NNSA's fiscal year 2006 budget request call for
the completion of QMU by 2010. Third, NNSA has not established formal
requirements for conducting annual, technical reviews of the
implementation of QMU at the three weapons laboratories or for
certifying the completion of QMU-related milestones. Finally, NNSA has
not established adequate performance measures to determine the progress
of the laboratories in developing and implementing QMU. Specifically,
NNSA officials were not able to show how they are able to measure
progress toward current performance targets related to the development
and implementation of QMU (e.g., NNSA's statement that the development
and implementation of QMU was 10 percent complete at the end of fiscal
year 2004). As a result of these deficiencies, NNSA cannot fully ensure
that it will be able to meet key deadlines for implementing QMU.
GAO is making five recommendations to the Administrator of NNSA to (1)
ensure that the three weapons laboratories have an agreed upon
technical approach for implementing QMU and (2) improve NNSA's
management of the development and implementation of QMU.
We provided NNSA with a draft of this report for their review and
comment. Overall, NNSA generally agreed that there was a need for an
agreed-upon technical approach for implementing QMU and that NNSA
needed to improve the management of QMU through clearer, long-term
milestones and better integration across the program. However, NNSA
stated that QMU had already been effectively implemented and that we
had not given NNSA sufficient credit for its success. In addition, NNSA
raised several issues about our conclusions and recommendations
regarding their management of the QMU effort. We have modified our
report to more fully recognize that QMU is being used by the
laboratories to address stockpile issues and to more completely
characterize its current state of development. NNSA also made technical
clarifications, which we incorporated in this report as appropriate.
Background:
Most modern nuclear warheads contain a nuclear explosive package, which
contains the primary and the secondary, and a set of nonnuclear
components.[Footnote 4] The nuclear detonation of the primary produces
energy that drives the secondary, which produces further nuclear energy
of a militarily significant yield. The nonnuclear components control
the use, arming, and firing of the warhead. All nuclear weapons
developed to date rely on nuclear fission to initiate their explosive
release of energy. Most also rely on nuclear fusion to increase their
total energy yield. Nuclear fission occurs when the nucleus of a heavy,
unstable atom (such as uranium-235) is split into two lighter parts,
which releases neutrons and produces large amounts of energy. Nuclear
fusion occurs when the nuclei of two light atoms (such as deuterium and
tritium) are joined, or fused, to form a heavier atom, with an
accompanying release of neutrons and larger amounts of energy.
The U.S. nuclear stockpile consists of nine weapon types. (See table
1.) The lifetimes of the weapons currently in the stockpile have been
extended well beyond the minimum life for which they were originally
designed--generally about 20 years--increasing the average age of the
stockpile and, for the first time, leaving NNSA with large numbers of
weapons that are close to 30 years old.
Table 1: Nuclear Weapons in the Enduring Stockpile:
Warhead or bomb mark: B61 3/4/10;
Description: Tactical bomb;
Date of entry into stockpile: 1979/1979/1990;
Laboratory: LANL, SNL;
Military service: Air Force.
Warhead or bomb mark: B61 7/11;
Description: Strategic bomb;
Date of entry into stockpile: 1985/1996;
Laboratory: LANL, SNL;
Military service: Air Force.
Warhead or bomb mark: W62;
Description: ICBM warhead[A];
Date of entry into stockpile: 1970;
Laboratory: LLNL, SNL;
Military service: Air Force.
Warhead or bomb mark: W76;
Description: SLBM warhead[B];
Date of entry into stockpile: 1978;
Laboratory: LANL, SNL;
Military service: Navy.
Warhead or bomb mark: W78;
Description: ICBM warhead[A];
Date of entry into stockpile: 1979;
Laboratory: LANL, SNL;
Military service: Air Force.
Warhead or bomb mark: W80 0/1;
Description: Cruise missile warhead;
Date of entry into stockpile: 1984/1982;
Laboratory: LLNL, SNL;
Military service: Air Force/Navy.
Warhead or bomb mark: B83 0/1;
Description: Strategic bomb;
Date of entry into stockpile: 1983/1993;
Laboratory: LLNL, SNL;
Military service: Air Force.
Warhead or bomb mark: W87;
Description: ICBM warhead[A];
Date of entry into stockpile: 1986;
Laboratory: LLNL, SNL;
Military service: Air Force.
Warhead or bomb mark: W88;
Description: SLBM warhead[B];
Date of entry into stockpile: 1989;
Laboratory: LANL, SNL;
Military service: Navy.
Source: NNSA.
Note: The dates of entry into the enduring nuclear stockpile are based
on when the weapon reached phase 6 of the weapons development and
production cycle. As of 2005, responsibility for the W80 0/1 was
transferred from LANL to LLNL.
[A] ICBM = intercontinental ballistic missile.
[B] SLBM = submarine launched ballistic missile.
[End of table]
Established in 1993, the Stockpile Stewardship Program faces two main
technical challenges: provide (1) a better scientific understanding of
the basic phenomena associated with nuclear weapons and (2) an improved
capability to predict the impact of aging and remanufactured components
on the safety and reliability of nuclear weapons. Specifically,
* An exploding nuclear weapon creates the highest pressures, greatest
temperatures, and most extreme densities ever made by man on earth,
within some of the shortest times ever measured. When combined, these
variables exist nowhere else in nature. While the United States
conducted about 1,000 nuclear weapons tests prior to the moratorium,
these tests were conducted mainly to look at broad indicators of weapon
performance (such as the yield of a weapon) and were often not designed
to collect data on specific properties of nuclear weapons physics.
After more than 60 years of developing nuclear weapons, while many of
the physical processes are well understood and accurately modeled, the
United States still does not possess a set of completely known and
expressed laws and equations of nuclear weapons physics that link the
physical event to first principles.
* As nuclear weapons age, a number of physical changes can take place.
The effects of aging are not always gradual, and the potential for
unexpected changes in materials causes significant concerns as to
whether weapons will continue to function properly. Replacing aging
components is, therefore, essential to ensure that the weapon will
function as designed. However, it may be difficult or impossible to
ensure that all specifications for the manufacturing of new components
are precisely met, especially since each weapon was essentially
handmade. In addition, some of the manufacturing process lines used for
the original production have been disassembled.
In 1995, the President established an annual assessment and reporting
requirement designed to help ensure that nuclear weapons remain safe
and reliable without underground testing.[Footnote 5] As part of this
requirement, the three weapons laboratories are required to issue a
series of reports and letters that address the safety, reliability,
performance, and military effectiveness of each weapon type in the
stockpile. The letters, submitted to the Secretary of Energy
individually by the laboratory directors, summarize the results of the
assessment reports and, among other things, express the directors'
conclusions regarding whether an underground nuclear test is needed and
the adequacy of various tools and methods currently in use to evaluate
the stockpile.
To address these challenges, in 1999 DOE developed a new three-part
program structure for the Stockpile Stewardship Program that included a
series of campaigns, which DOE defined as technically challenging,
multiyear, multifunctional efforts to develop and maintain the critical
capabilities needed to continue assessing the safety and reliability of
the nuclear stockpile into the foreseeable future without underground
testing. DOE originally created 18 campaigns that were designed to
focus its efforts in science and computing, applied science and
engineering, and production readiness. Six of these campaigns currently
focus on the development and improvement of the scientific knowledge,
tools, and methods required to provide confidence in the assessment and
certification of the safety and reliability of the nuclear stockpile in
the absence of nuclear testing. These six campaigns are as follows:
* The Primary and Secondary campaigns were established to analyze and
understand the different scientific phenomena that occur in the primary
and secondary stages of a nuclear weapon during detonation. As such,
the Primary and Secondary campaigns are intended to support the
development and implementation of the QMU methodology and to set the
requirements for the computers, computer models, and experimental data
needed to assess and certify the performance of nuclear weapons.
* The ASC campaign provides the leading-edge supercomputers and models
that are used to simulate the detonation and performance of nuclear
weapons.
* Two campaigns--Advanced Radiography and Dynamic Materials Properties-
-provide data from laboratory experiments to support nuclear weapons
theory and computational modeling. For example, the Advanced
Radiography campaign conducts experiments that measure how stockpile
materials behave when exposed to explosively driven shocks. One of the
major facilities being built to support this campaign is the Dual Axis
Radiographic Hydrodynamic Test Facility at LANL.
* The ICF campaign develops experimental capabilities and conducts
experiments to examine phenomena at high temperature and pressure
regimes that approach but do not equal those occurring in a nuclear
weapon. As a result, scientists currently have to extrapolate from the
results of these experiments to understand similar phenomena in a
nuclear weapon. One of the major facilities being built as part of this
campaign is the National Ignition Facility at LLNL.
The other two program activities associated with the Stockpile
Stewardship Program are "Directed Stockpile Work" and "Readiness in
Technical Base and Facilities." Directed Stockpile Work includes the
activities that directly support specific weapons in the stockpile,
such as the Stockpile Life Extension Program, which employs a
standardized approach for planning and carrying out nuclear weapons
refurbishment activities to extend the operational lives of the weapons
in the stockpile well beyond their original design lives. The life
extension for the W87 was completed in 2004, and three other weapon
systems--the B61, W76, and W80--are currently undergoing life
extensions. Each life extension program is specific to that weapon
type, with different parts being replaced or refurbished for each
weapon type. Readiness in Technical Base and Facilities includes the
physical infrastructure and operational readiness required to conduct
campaign and Directed Stockpile Work activities across the nuclear
weapons complex. The complex includes the three nuclear weapons design
laboratories (LANL, LLNL, and SNL), the Nevada Test Site, and four
production plants--the Pantex Plant in Texas, the Y-12 Plant in
Tennessee, a portion of the Savannah River Site in South Carolina, and
the Kansas City Plant in Missouri.
From fiscal year 2001 through fiscal year 2005, NNSA spent over $7
billion on the six scientific campaigns (in inflation-adjusted
dollars). (See table 2.) NNSA has requested almost $7 billion in
funding for these campaigns over the next 5 years. (See table 3.)
Table 2: NNSA Funding for the Scientific Campaigns, Fiscal Years 2001-
2005:
Dollars in millions.
Primary;
FY 2001: $49.8;
FY 2002: $52.4;
FY 2003: $48.7;
FY 2004: $41.2;
FY 2005: $73.4;
Total: $265.5.
Secondary;
FY 2001: 43.7;
FY 2002: 42.1;
FY 2003: 49.2;
FY 2004: 54.6;
FY 2005: 57.2;
Total: 246.8.
ASC;
FY 2001: 770.9;
FY 2002: 692.2;
FY 2003: 799.3;
FY 2004: 738.9;
FY 2005: 685.9;
Total: 3,687.2.
Advanced Radiography;
FY 2001: 85.7;
FY 2002: 100.3;
FY 2003: 74.2;
FY 2004: 53.5;
FY 2005: 52.7;
Total: 366.4.
Dynamic Materials Properties;
FY 2001: 79.4;
FY 2002: 80.7;
FY 2003: 85.2;
FY 2004: 87.8;
FY 2005: 74.2;
Total: 407.3.
ICF;
FY 2001: 515.7;
FY 2002: 593.3;
FY 2003: 518.9;
FY 2004: 480.1;
FY 2005: 492.1;
Total: 2,600.1.
Total;
FY 2001: $1,545.2;
FY 2002: $1,561.0;
FY 2003: $1,575.5;
FY 2004: $1,456.1;
FY 2005: $1,435.5;
Total: $7,573.3.
[End of table]
Source: NNSA.
Note: In constant dollars, base year 2005.
Table 3: NNSA Funding Requests and Projections for the Scientific
Campaigns, Fiscal Years 2006-2010:
Primary;
FY 2006: $45.2;
FY 2007: $47.5;
FY 2008: $48.9;
FY 2009: $48.7;
FY 2010: $45.6;
Total: $235.9.
Secondary;
FY 2006: $61.3;
FY 2007: $63.9;
FY 2008: $65.0;
FY 2009: $65.0;
FY 2010: $65.0;
Total: $320.2.
ASC;
FY 2006: $660.8;
FY 2007: $666.0;
FY 2008: $666.0;
FY 2009: $666.0;
FY 2010: $666.0;
Total: $3,324.8.
Advanced Radiography;
FY 2006: $49.5;
FY 2007: $42.7;
FY 2008: $39.5;
FY 2009: $38.7;
FY 2010: $41.9;
Total: $212.3.
Dynamic Materials Properties;
FY 2006: $80.9;
FY 2007: $85.1;
FY 2008: $86.5;
FY 2009: $87.4;
FY 2010: $87.4;
Total: $427.3.
ICF;
FY 2006: $460.4;
FY 2007: $461.6;
FY 2008: $461.6;
FY 2009: $461.6;
FY 2010: $461.6;
Total: $2,306.8.
Total;
FY 2006: $1,358.1;
FY 2007: $1,366.8;
FY 2008: $1,367.5;
FY 2009: $1,367.4;
FY 2010: $1,367.5;
Total: $6,827.3.
Source: DOE, FY 2006 Congressional Budget Request, February 2005.
[End of table]
Within NNSA, the Office of Defense Programs is responsible for managing
the campaigns and the Stockpile Stewardship Program in general. Within
this office, two organizations share responsibility for overall
management of the scientific campaigns: the Office of the Assistant
Deputy Administrator for Research, Development, and Simulation and the
Office of the Assistant Deputy Administrator for Inertial Confinement
Fusion and the National Ignition Facility Project. The first office
oversees campaign activities associated with the Primary and Secondary
campaigns--as well as the ASC, Advanced Radiography, and Dynamic
Materials Properties campaigns--with a staff of about 13 people. The
second office oversees activities associated with the ICF campaign with
a single staff person. Actual campaign activities are conducted by
scientists and other staff at the three weapons laboratories. LANL and
LLNL conduct activities associated with the nuclear explosive package,
while SNL performs activities associated with the nonnuclear components
that control the use, arming, and firing of the nuclear warhead.
The QMU Methodology Is Highly Promising but Still in the Early Stages
of Development:
NNSA has endorsed the use of a new common methodology, known as the
quantification of margins and uncertainties, or QMU, for assessing and
certifying the safety and reliability of the nuclear stockpile. NNSA
and laboratory officials told us that they have made progress in
applying the principles of QMU to the certification and assessment of
nuclear warheads in the stockpile. However, QMU is still in its early
stages of development, and important differences exist among the three
laboratories in their application of QMU. To date, NNSA has
commissioned two technical reviews of the implementation of QMU at the
weapons laboratories. While strongly supporting QMU, the reviews found
that the development and implementation of QMU was still in its early
stages. The reviews recommended that NNSA take steps to further define
the technical details supporting the implementation of QMU and
integrate the activities of the three weapons laboratories in
implementing QMU. However, NNSA and the weapons laboratories have not
fully implemented these recommendations. Beyond the issues raised in
the two reports, we also found differences in the understanding and
application of QMU among the three laboratories.
NNSA Has Endorsed QMU as a New, Common Methodology for Assessing and
Certifying Stockpile Safety and Reliability:
When the Primary and Secondary campaigns were established in 1999, they
brought some organization and overall goals to the scientific research
conducted across the weapons complex. For example, as we noted in April
2005, the Primary campaign set an initial goal in the 2005 to 2010 time
frame for certifying the performance of the primary of a nuclear weapon
to within a stated yield level.[Footnote 6] However, according to
senior NNSA officials, NNSA still lacked a coherent strategy for
relating the scientific work conducted by the weapons laboratories
under the campaigns to the needs of the nuclear stockpile and the
overall Stockpile Stewardship Program. This view was echoed by a NNSA
advisory committee report, which stated in 2002 that the process used
by the weapons laboratories to certify the safety and reliability of
nuclear weapons was ill defined and unevenly applied, leading to major
delays and inefficiencies in programs.[Footnote 7]
Starting in 2001, LLNL and LANL began developing what is intended to be
a common methodology for assessing and certifying the performance and
safety of nuclear weapons in the absence of nuclear testing. In 2003,
the associate directors for nuclear weapons at LLNL and LANL published
a white paper--entitled "National Certification Methodology for the
Nuclear Weapon Stockpile"--that described this new methodology, which
they referred to as the quantification of margins and uncertainties or
QMU. According to the white paper, QMU is based on an adaptation of
standard engineering practices and lends itself to the development of
"rigorous, quantitative, and explicit criteria for judging the
robustness of weapon system and component performance at a detailed
level." Moreover, the quantitative results of this process would enable
NNSA and the weapons laboratories to set priorities for their
activities and thereby make rational decisions about allocating program
resources to the nuclear stockpile.
The process envisaged in the white paper focuses on creating a "watch
list" of factors that, in the judgment of nuclear weapons experts, are
the most critical to the operation and performance of a nuclear weapon.
These factors include key operating characteristics and components of
the nuclear weapon. For each identified, critical factor leading to a
nuclear explosion, nuclear weapons experts would define performance
metrics. These performance metrics would represent the experts' best
judgment of what constitutes acceptable behavior--i.e., the range of
acceptable values for a critical function to successfully occur or for
a critical component to function properly--as well as what constitutes
unacceptable behavior or failure. To use an analogy, consider the
operation of a gasoline engine. Some of the events critical to the
operation of the engine would include the opening and closing of
valves, the firing of the spark plugs, and the ignition of the fuel in
each cylinder. Relevant performance metrics for the ignition of fuel in
a cylinder would include information on the condition of the spark
plugs (e.g., whether they are corroded) and the fuel/air mixture in the
cylinder.
Once nuclear experts have identified the relevant performance metrics
for each critical factor, according to the 2003 white paper, the goal
of QMU is to quantify these metrics. Specifically, the QMU methodology
seeks to quantify (1) how close each critical factor is to the point at
which it would fail to perform as designed (i.e., the performance
margin or the margin to failure) and (2) the uncertainty in calculating
the margin. According to the white paper, the weapons laboratories
would be able to use their calculated values of margins and
uncertainties as a way to assess their confidence in the performance of
a nuclear weapon. That is, the laboratories would establish a
"confidence ratio" for each critical factor --they would divide their
calculated value for the margin ("M") by their calculations of the
associated uncertainty ("U") and arrive at a single number ("M/U").
According to the white paper, the weapons laboratories would only have
confidence in the performance of a nuclear weapon if the margin
"significantly" exceeds uncertainty for all critical issues. However,
the white paper did not define what the term "significantly" meant.
In a broad range of key planning and management documents that have
followed the issuance of the white paper, NNSA and the weapons
laboratories have endorsed the use of the QMU methodology as the
principal tool for assessing and certifying the safety and reliability
of the nuclear stockpile in the absence of nuclear testing. For
example, in its fiscal year 2006 implementation plan for the Primary
campaign, NNSA stated as a strategic objective that it needs to develop
the capabilities and understanding necessary to apply QMU as the
assessment and certification methodology for the nuclear explosive
package. In addition, in its fiscal year 2006 budget request, NNSA
selected its progress toward the development and implementation of QMU
as one of its major performance indicators. Finally, in the plans that
NNSA uses to evaluate the performance of LANL and LLNL, NNSA has
established an overall objective for LANL and LLNL to assess and
certify the safety and reliability of nuclear weapons using a common
QMU methodology.
Officials at NNSA and the weapons laboratories have also stated that
QMU will be vital to certifying any weapon redesigns, such as are
envisioned by the RRW program. For example, senior NNSA officials told
us that the Stockpile Stewardship Program will not be sustainable if it
only involves the continued refurbishment in perpetuity of existing
weapons in the current nuclear stockpile. They stated that the
accumulation of small changes over the extended lifetime of the current
nuclear stockpile will result in increasing levels of uncertainty about
its performance. If NNSA moves forward with the RRW program, according
to NNSA documents and officials, the future goal of the weapons program
will be to use QMU to replace existing stockpile weapons with an RRW
whose safety and reliability could be assured with the highest
confidence, without nuclear testing, for as long as the United States
requires nuclear forces.
The Development and Implementation of QMU Is at an Early Stage and
Important Differences Exist Among the Weapons Laboratories in their
Application of QMU:
According to NNSA and laboratory officials, the weapons laboratories
have made progress in applying the principles of QMU to the
certification of life extension programs and to the annual stockpile
assessment process. For example, LLNL officials told us that they are
applying QMU to the assessment of the W80, which is currently
undergoing a life extension.[Footnote 8] They said that, in applying
the QMU methodology, they tend to focus their efforts on identifying
credible "failure modes," which are based on observable problems, such
as might be caused by the redesign of components in a nuclear weapon,
changes to the manufacturing process for components, or the performance
of a nuclear weapon under aged conditions. They said that, for the W80
life extension program, they have developed a list of failure modes and
quantified the margins and uncertainties associated with these failure
modes. Based on their calculations, they said that they have increased
their confidence in the performance of the W80.
Similarly, LANL officials told us that they are applying QMU to the
W76, which is also currently undergoing a life extension and is
scheduled to finish its first production unit in 2007. They said that,
in applying the QMU methodology, they tend to focus their efforts on
defining "performance gates," which are based on a number of critical
points during the explosion of a nuclear weapon that separate the
nuclear explosion into natural stages of operation. The performance
gates identify the characteristics that a nuclear weapon must have at a
particular time during its operation to meet its performance
requirements (e.g., to reach its expected yield). LANL officials told
us that they have developed a list of performance gates for the W76
life extension program and are beginning to quantify the margins and
uncertainties associated with these performance gates.
Despite this progress, we found that QMU is still in its early stages
of development and that important differences exist among the weapons
laboratories in their application of QMU. To date, NNSA has
commissioned two technical reviews of the implementation of QMU at the
weapons laboratories. The first review was conducted by NNSA's Office
of Defense Programs Science Council (Science Council)--which advises
NNSA on scientific matters across a range of activities, including
those associated with the scientific campaigns--and resulted in a March
2004 report.[Footnote 9] The second review was conducted by the MITRE
Corporation's JASON panel and resulted in a February 2005
report.[Footnote 10] Both reports endorsed the use of QMU by the
weapons laboratories and listed several potential benefits that QMU
could bring to the nuclear weapons program. For example, according to
the Science Council report, QMU will serve an important role in
training the next generation of nuclear weapon designers and will
quantify and increase NNSA's confidence in the assessment and
certification of the nuclear stockpile. According to the JASON report,
QMU could become a useful management tool for directing investments in
a given weapon system where they would be most effective in increasing
confidence, as required by the life extension programs. In addition,
the JASON report described how LANL and LLNL officials had identified
potential failure modes in several weapon systems and calculated the
associated margins and uncertainties. The report noted that, for most
of these failure modes, the margin for success was large compared with
the uncertainty in the performance.
However, according to both the Science Council and the JASON reports,
the development and implementation of QMU is still in its early stages.
For example, the JASON report described QMU as highly promising but
unfinished, incomplete and evolving, and in the early stages of
development. Moreover, the chair of the JASON panel on QMU told us in
June 2005 that, during the course of his review, members of the JASON
panel found that QMU was not mature enough to assess its reliability or
usefulness. The reports also stated that the weapons laboratories have
not fully developed or agreed upon the technical details supporting the
implementation and application of QMU. For example, the JASON report
stated that, in the course of its review, it became evident that there
were a variety of differing and sometimes diverging reviews of what QMU
really was and how it was working in practice. As an example, the
report stated that some of the scientists, designers, and engineers at
LANL and LLNL saw the role of expert judgment as an integral part of
the QMU process, while others did not. In discussions with the weapons
laboratories about the two reports, LANL officials told us that they
believed that the details of QMU as a formal methodology are still
evolving, while LLNL officials stated that QMU was "embryonic" and not
fully developed.
While supporting QMU, the two reports noted that the weapons
laboratories face challenges in successfully implementing a coherent
and credible analytical method based on the QMU methodology. For
example, in its 2004 report, the Science Council stated that, in its
view, the QMU methodology is based on the following core assumptions:
* Computer simulations can accurately predict the behavior of a complex
nuclear explosive system as a function of time.
* It is sufficient for the assessment of the performance of a nuclear
weapon to examine the simulation of the time evolution of a nuclear
explosive system at a number of discrete time intervals and to
determine whether the behavior of the system at each interval is within
acceptable bounds.
* The laboratories' determinations of acceptable behavior can be made
quantitatively--that is, they will make a quantitative estimate of a
system's margins and uncertainties.
* Given these quantitative measures of the margins and uncertainties,
it is possible to calculate the probability (or confidence level) that
the nuclear explosive system will perform as desired.
However, the Science Council's report noted that extraordinary degrees
of complexity are involved in a rational implementation of QMU that are
only beginning to be understood. For example, in order for the QMU
methodology to have validity, it must sufficiently identify all
critical failure modes, critical events, and associated performance
metrics. However, as described earlier, the operation of an exploding
nuclear weapon is highly integrated and nonlinear, occurs during a very
short period of time, and reaches extreme temperatures and pressures.
In addition, the United States does not possess a set of completely
known and expressed laws and equations of nuclear weapons physics.
Given these complexities, it will be difficult to demonstrate the
successful implementation of QMU, according to the report. In addition,
the Science Council stated that it was not presented with any evidence
that there exists a method--even in principle--for calculating an
overall probability that a nuclear explosive package will perform as
designed from the set of quantitative margins and uncertainties at each
time interval.
To address these and other issues, the two reports recommended that
NNSA take steps to further define the technical details supporting the
implementation of QMU and to integrate the activities of the three
weapons laboratories in implementing QMU. For example, the 2004 Science
Council report recommended that NNSA direct the associate directors for
nuclear weapons at LANL and LLNL to undertake a major effort to define
the details of QMU. In particular, the report recommended that a
trilaboratory team be charged with defining a common language for QMU
and identifying the important performance gates, failure modes, and
other criteria in the QMU approach. The report stated that this agreed-
upon "reference" set could then be used to support all analyses of
stockpile issues. In addition, the report recommended that NNSA
consider establishing annual or semiannual workshops for the three
weapons laboratories to improve the identification, study, and
prioritization of potential failure modes and other factors that are
critical to the operation and performance of nuclear weapons.
Similarly, the 2005 JASON panel report noted that the meaning and
implications of QMU are currently unclear. To rectify this problem, the
report recommended that the associate directors for nuclear weapons at
LANL and LLNL write a new, and authoritative, paper defining QMU and
submit it to NNSA. Furthermore, the report recommended that the
laboratories establish a formal process to (1) identify all failure
modes and performance gates associated with QMU, using the same
methodology for all weapon systems, and (2) establish better
relationships between the concepts of failure modes and performance
gates for all weapon systems in the stockpile.
However, NNSA and laboratory officials have not fully implemented these
recommendations, particularly the recommendations of the Science
Council. For example, while LLNL and LANL officials are drafting a new
"white paper" on QMU that attempts to clarify some fundamental tenets
of the methodology, officials from SNL are not involved in the drafting
of this paper. In addition, NNSA has not required the three weapons
laboratories to hold regular meetings or workshops to improve the
identification, prioritization, and integration of failure modes,
performance gates, and other critical factors.
According to NNSA's Assistant Deputy Administrator for Research,
Development, and Simulation, NNSA has not fully implemented the
recommendations of the Science Council's report partly because the
report was intended more to give NNSA a sense of the status of the
implementation of QMU than it was to provide recommendations. For
example, the 2004 report states that the "friendly review," as the
report is referred to by NNSA, would not have budget implications and
that the report's findings and recommendations would be reported only
to the senior management of the weapons laboratories. As a result, the
Assistant Deputy Administrator told us that he had referred the
recommendations to the directors of the weapons laboratories and told
them to implement the recommendations as they saw fit.
Furthermore, LLNL and LANL officials disagreed with some of the
statements in the Science Council report and stressed that, in using
QMU, they do not attempt to assign an overall probability that the
nuclear explosive package will perform as desired. That is, they do not
attempt to add up calculations of margins and uncertainties for all the
critical factors to arrive at a single estimate of margin and
uncertainty, or a single confidence ratio, for the entire nuclear
explosive package. Instead, they said that they focus on ensuring that
the margin for each identified critical factor in the explosion of a
nuclear weapon is greater than the uncertainty. However, they said
that, for a given critical factor, they do combine various calculations
of individual uncertainties that contribute to the total amount of
uncertainty for that factor.
In addition, in addressing comments in the JASON report, LLNL and LANL
officials stressed that QMU has always relied, and will continue to
rely heavily, on the judgment of nuclear weapons experts. For example,
LLNL officials told us that since there is no single definition of what
constitutes a threshold for failure, they use expert judgment to decide
what to put on their list of failure modes. They also said that the QMU
methodology provides a way to make the entire annual assessment and
certification process more transparent to peer review. Similarly, LANL
officials said that they use expert judgment extensively in
establishing performance metrics and threshold values for their
performance gates. They said that expert judgment will always be a part
of the scientific process and a part of QMU.
Beyond the issues raised in the two reports, we found that there are
differences in the understanding and application of QMU among the three
laboratories. For example, the three laboratories do not agree about
the application of QMU to areas outside of the nuclear explosive
package. Specifically, LLNL officials told us that the QMU methodology,
as currently developed, only applies to the nuclear explosive package
and not to the nonnuclear components that control the use, arming, and
firing of the nuclear warhead. According to LLNL and LANL officials,
SNL scientists can run hundreds of experiments to test their components
and, therefore, can use normal statistical analysis in certifying the
performance of nonnuclear components. As a result, according to LLNL
and LANL officials, SNL does not have to cope with real uncertainty and
does not "do" QMU. Furthermore, according to LLNL officials, SNL has
chosen not to participate in the development of QMU with LLNL and LANL.
However, SNL officials told us that while some of the nonnuclear
components are testable to a degree, SNL is as challenged as the other
two weapons laboratories in certifying the performance of their systems
without actual testing. For example, SNL officials said that they
simply do not have enough money to perform enough tests on all of their
nonnuclear components to be able to rely completely on statistical
analysis to meet their safety performance levels. In addition, SNL
scientists are not able to test their components under the conditions
of a nuclear explosion but are still required to certify the
performance of the components under these conditions. Thus, SNL
officials told us that they had been using their own version of QMU for
a long time.
SNL officials told us that they define QMU as a way to make risk-
informed decisions about the effect of variabilities and uncertainties
on the performance of a nuclear weapon, including the nonnuclear
components that control the use, arming, and firing of the nuclear
warhead. Moreover, they said that this kind of risk-informed approach
is not unique to the nuclear weapons laboratories and is used
extensively in areas such as nuclear reactor safety. However, they told
us that they have been left out in the development of QMU by the two
other weapons laboratories. Specifically, they said that while SNL
scientists have worked with other scientists at LANL and LLNL at a
"grass roots" level, there has only been limited cooperation and
dialogue between upper-level management at the three laboratories
concerning the development and implementation of QMU.
In addition, we found that while LLNL and LANL both agree on the
fundamental tenets of QMU at a high level, their application of the QMU
methodology differs in some important respects. For example, LLNL and
LANL officials told us that, at a detailed level, the two laboratories
are pursuing different approaches to calculating and combining
uncertainties. For the W80 life extension program, LLNL officials
showed us how they combined calculations of individual uncertainties
that contributed to the total uncertainty for a key failure mode of the
primary--the amount of primary yield necessary to drive the secondary.
However, they said that the scientific support for their method for
combining individual calculations of uncertainty was limited, and they
stated that they are pursuing a variety of more sophisticated analyses
to improve their current approach.
Moreover, the two laboratories are taking a different approach to
generating a confidence ratio for each critical factor, as described in
the 2003 white paper on QMU. For example, for the W80 life extension
program, LLNL officials showed us how they calculated a single
confidence ratio for a key failure mode of the primary, based on their
calculations of margin and uncertainty. They said that the weapon
systems for which they are responsible have a lot of margin built into
them, and they feel comfortable generating this number. In contrast, in
discussions with LANL officials about the W76 life extension program,
LANL officials told us that they prefer not to calculate a single
confidence ratio for a performance gate, partly because they are
concerned that their customers (e.g., the Department of Defense) might
think that the QMU methodology is more formal than it is currently.
In commenting on the differences between the two laboratories, NNSA
officials stated that the two laboratories are pursuing complementary
approaches, and that these differences are part of the rationale for a
national policy decision to maintain two nuclear design laboratories.
In addition, they stated that the confidence in the correctness of
scientific research is improved by achieving the same answer through
multiple approaches. LLNL officials also made similar comments, stating
that the nation will benefit from some amount of independence between
the laboratories to assure that the best methodology for assessing the
stockpile in the absence of nuclear testing is achieved.
NNSA's Management of the Development and Implementation of QMU Is
Deficient in Four Key Areas:
NNSA relies on its Primary and Secondary campaigns to manage the
development and implementation of QMU. According to NNSA policies,
campaign managers at NNSA headquarters are responsible for developing
campaign plans and high-level milestones, overseeing the execution of
these plans, and providing input to the evaluation of the performance
of the weapons laboratories. However, NNSA's management of these
processes is deficient in four key areas. First, the planning documents
that NNSA has established for the Primary and Secondary campaigns do
not adequately integrate the scientific research currently conducted
that supports the development and implementation of QMU. Second, NNSA
has not developed a clear, consistent set of milestones to guide the
development and implementation of QMU. Third, NNSA has not established
formal requirements for conducting annual, technical reviews of the
implementation of QMU or for certifying the completion of QMU-related
milestones. Finally, NNSA has not established adequate performance
measures to determine the progress of the laboratories in developing
and implementing QMU.
Campaign Planning Documents Do Not Adequately Integrate the Scientific
Activities Supporting QMU:
As part of its planning structure, NNSA requires the use of program and
implementation plans to set requirements and manage resources for the
campaigns and other programs associated with the Stockpile Stewardship
Program. Program plans are strategic in nature and identify the long-
term goals, high-level milestones, and resources needed to support a
particular program over a 7-year period, while implementation plans
establish performance expectations for the program and each
participating site for the current year of execution. According to NNSA
policies, program and implementation plans should flow from and
interact with each other using a set of cascading goals and
requirements.
NNSA has established a single program plan, which it calls the "Science
campaign program plan," that encompasses the Primary and the Secondary
campaigns, as well as two other campaigns--Advanced Radiography and
Dynamic Materials Properties. NNSA has also established separate
implementation plans for each of these campaigns, including the Primary
and Secondary campaigns. According to NNSA, it relies on these plans--
and in particular the plans related to the Primary and Secondary
campaigns--to manage the development and implementation of QMU, as well
as to determine the requirements for the experimental data and computer
modeling needed to analyze and understand the different scientific
phenomena that occur in a nuclear weapon during detonation.
However, the current Primary and Secondary campaign plans do not
contain a comprehensive, integrated list of the relevant scientific
research being conducted across the weapons complex to support the
development and implementation of QMU. For example, according to the
NNSA campaign manager for the Primary campaign, he had to hold a
workshop in 2005 with officials from the weapons laboratories in order
to catalogue all of the scientific activities that are currently
performed under the heading of "primary assessment" regardless of the
NNSA funding source. According to this official, the existing Primary
campaign implementation plan does not provide the integration across
NNSA programs that is needed to achieve the goals of the Primary
campaign and to develop and implement QMU.
According to NNSA officials, the lack of integration has occurred in
large part because a significant portion of the scientific research
that is relevant to the Primary and Secondary campaigns is funded and
carried out by different campaigns and other programs. Specifically,
different NNSA campaign managers use different campaign planning
documents to plan and oversee research and funding for activities that
are directly relevant to the Primary and Secondary campaigns and the
development and implementation of QMU. For example, the ASC campaign
provides the supercomputing capability that the weapons laboratories
use to simulate and predict the behavior of an exploding nuclear
weapon. Moreover, the weapons laboratories rely on ASC supercomputers
to quantify their uncertainties with respect to the accuracy of these
computer simulations--a key component in the implementation of QMU. As
a result, the ASC campaign plans and funds activities that are critical
to the development and implementation of QMU.
To address this problem, according to NNSA officials, NNSA is taking
steps to establish better relationships among the campaign plans. For
example, NNSA is currently drafting a new plan--which it calls the
Primary Assessment Plan--in an attempt to better coordinate the
activities covered under the separate program and implementation plans.
The draft plan outlines high-level research priorities, time lines, and
proposed milestones necessary to support (1) NNSA's responsibilities
for the current stockpile, (2) primary physics design for the
development of an RRW, and (3) certification of an RRW in the 2012 time
frame and a second RRW in the 2018 time frame. According to NNSA
officials, they expect to finalize this plan by the third quarter of
fiscal year 2006. In addition, they expect to have a similar plan for
the Secondary campaign finalized by December 2006 and are considering
combining both plans into a full-system assessment plan. According to
one NNSA official responsible for the Primary and Secondary campaigns,
NNSA will revise the existing campaign program and implementation plans
to be consistent with the Primary Assessment Plan.
More fundamentally, some nuclear weapons experts have suggested that
NNSA's planning structure should be reorganized to better reflect the
use of QMU as NNSA's main strategy for assessing and certifying the
performance of nuclear weapons. For example, the chair of the LLNL
Defense and Nuclear Technologies Director's Review Committee--which
conducts technical reviews of LLNL's nuclear weapons activities for the
University of California--told us that the current campaign structure
has become a series of "stovepipes" that NNSA uses to manage stockpile
stewardship. He said that in order for NNSA to realize its long-term
goals for implementing QMU, NNSA is going to have to reorganize itself
around something that he called an "uncertainty spreadsheet" for each
element of a weapon's performance (e.g., implosion of the primary,
transfer of energy to the secondary, etc.), leading to the weapon's
yield. He said that the laboratories should develop a spreadsheet for
each weapon in the stockpile that (1) identifies the major sources of
uncertainty at each critical event in their assessment of the weapon's
performance and (2) relates the laboratory's scientific activities and
milestones to these identified sources of uncertainty. He said that the
development and use of these spreadsheets would essentially capture the
intent of the scientific campaigns and make them unnecessary.
NNSA Does Not Have a Clear, Consistent Set of QMU-Related Milestones:
NNSA has established a number of milestones that relate to the
development and implementation of QMU. Within the Science campaign
program plan, NNSA has established a series of high-level milestones,
which it calls "level-1" milestones. According to NNSA policies, level-
1 milestones should be sufficient enough to allow strategic integration
between sites involved in the campaigns and between programs in NNSA.
Within the implementation plans for the Primary and Secondary
campaigns, NNSA has established a number of lower-level milestones,
which it calls "level-2" milestones, which NNSA campaign managers use
to track major activities for the current year of execution. The level-
1 milestones related to QMU are shown in table 4, and the level-2
milestones related to QMU for the Primary campaign are shown in table
5.
Table 4: NNSA Level-1 Milestones Related to the Development and
Implementation of QMU:
Due date: FY2007;
Milestone number: M46;
Milestone description: Publish documented plan to reduce major sources
of uncertainty. (Cycle I).
Due date: FY2010;
Milestone number: M47;
Milestone description: Accounting for simulation and experimental
uncertainties, assess ability to reproduce the full underground test
data sets for a representative group of nuclear tests with a consistent
set of models.
Due date: FY2011;
Milestone number: M48;
Milestone description: Publish documented plan to reduce the major
sources of uncertainty assessed in fiscal year 2010. (Cycle II).
Due date: FY2014;
Milestone number: M20;
Milestone description: Accounting for simulation and experimental
uncertainties, reassess ability to reproduce the full underground test
data sets for a representative group of nuclear tests with a consistent
set of models.
Source: NNSA, FY2006 Science campaign program plan.
[End of table]
Table 5: Primary Campaign Level-2 Milestones Related to the Development
and Implementation of QMU:
Due date: FY2004;
Milestone description: Analyze specific underground test events in the
support of QMU.
Due date: FY2004;
Milestone description: Develop QMU certification logic to support the
W76.
Due date: FY2004;
Milestone description: Develop QMU certification logic to support the
W88.
Due date: FY2005;
Milestone description: Analyze specific underground test events in the
support of QMU.
Due date: FY2005;
Milestone description: Predict primary performance and identify major
sources of uncertainty for the W-76 LEP. Quantify these sources where
possible or develop requirements of a plan to do so.
Due date: FY2005;
Milestone description: Develop probabilistic tools and methods to
combine various sources of uncertainty for primary performance.
Source: NNSA Primary campaign implementation plans, fiscal years 2004
and 2005.
[End of table]
According to NNSA officials, the level-1 milestones in table 4
represent a two-stage path to systematically identify uncertainties and
reduce them through analyzing past underground test results, developing
new experimental capabilities, and performing new experiments to
understand the relevant physical processes. According to these level-1
milestones, NNSA expects to complete the second stage or "cycle" of
this process by fiscal year 2014 (i.e., milestone M20), at which time
NNSA will have sufficiently reduced major sources of uncertainties and
will have confidence in its ability to predict the performance of
nuclear weapons in the absence of nuclear testing.
However, we identified several problems with the NNSA milestones
related to the development and implementation of QMU. Specifically, the
level-1 milestones in the Science campaign program plan have the
following problems:
* The milestones are not well-defined and never explicitly mention QMU.
According to NNSA officials responsible for overseeing the Primary
campaign, these milestones are too qualitative and too far in the
future to enable NNSA to effectively plan for and oversee the
implementation of QMU. They described these milestones as "fuzzy" and
said that they need to be better defined. However, NNSA officials also
stated that these milestones are not just for QMU but for the entire
Science campaign, of which QMU is only a part.
* The milestones conflict with the performance measures shown in other
important NNSA management documents. Specifically, while the Science
campaign program plan envisions a two-stage path to identify and reduce
key uncertainties related to nuclear weapon operations using QMU by
2014, the performance measures in NNSA's fiscal year 2006 budget
request and in Appendix A of the Science campaign program plan call for
the completion of QMU by 2010.
* The milestones have not been integrated with other QMU-related level-
1 milestones in other planning documents. For example, the current ASC
campaign program plan contains a series of level-1 milestones for
completing the certification of several weapon systems--including the
B61, W80, W76, and W88--with quantified margins and uncertainties by
the end of fiscal year 2007. However, these milestones do not appear in
and are not referenced by the Science campaign program plan. Moreover,
the ASC campaign manager told us that, until recently, he was not aware
of the existence of the level-1 milestones for implementing QMU that
are contained in the Science campaign program plan.
In addition, we found that neither the Science campaign program plan
nor the Primary campaign implementation plan describe how the level-2
milestones on QMU in the Primary campaign implementation plan are
related to the level-1 milestones on QMU in the Science campaign
program plan. Consequently, it is unclear how the achievement of
specific level-2 milestones--such as the development of probabilistic
tools and methods to combine various sources of uncertainty for primary
performance--will result in the achievement of level-1 milestones for
the implementation of QMU or how NNSA expects to certify several major
nuclear weapon systems using QMU before the QMU methodology is fully
developed and implemented.
NNSA, as well as laboratory officials, agreed that there are weaknesses
with the current QMU milestones. According to NNSA officials, when NNSA
established the current tiered structure for campaign milestones in
2003, the different tiers of milestones served different purposes and,
therefore, were never well-integrated. For example, NNSA officials said
that the level-1 milestones were originally created to reflect measures
that were deemed to be important to senior NNSA officials, while level-
2 milestones were created to be used by NNSA campaign managers to
perform more technical oversight of the weapons laboratories.
Furthermore, according to NNSA officials, the current level-2
milestones are only representative of campaign activities conducted by
the weapons laboratories. That is, the level-2 milestones were never
designed to cover the entire scope of work being conducted by the
weapons laboratories and are, therefore, not comprehensive in scope.
To address these problems, according to NNSA officials, NNSA is taking
steps to develop better milestones to track the implementation of the
QMU methodology. For example, in the draft Primary Assessment Plan,
NNSA has established 19 "high-level" milestones that cover the time
period from fiscal year 2006 to fiscal year 2018. According to these
draft milestones, by fiscal year 2010, NNSA expects to "complete the
experimental work and methodology development needed to demonstrate the
ability of primary certification tools to support certification of
existing stockpile system and RRW." In addition, NNSA expects to
certify a RRW in fiscal year 2012 and a second RRW in fiscal year 2018.
NNSA Has Not Established Formal Requirements for Conducting Technical
Reviews or Certifying the Completion of QMU-Related Milestones:
According to NNSA policies, campaign managers are required to track the
status of level-1 and level-2 milestones and provide routine, formal
reports on the status of their programs. For example, campaign managers
are required to track, modify, and score the status of level-1 and
level-2 milestones through the use of an Internet-based application
called the Milestone Reporting Tool. On a quarterly basis, campaign
managers assign one of four possible scores for each milestone listed
in the application: (1) "blue" for completed milestones, (2) "green"
for milestones that are on track to be finished by the end of the
fiscal year, (3) "yellow" for milestones that may not be completed by
the end of the fiscal year, and (4) "red" for milestones that will not
be completed by the end of the fiscal year. At quarterly program review
meetings, campaign managers brief senior-level NNSA officials on the
status of major milestones, along with cost and expenditure data for
their programs. In addition, campaign managers are responsible for
conducting technical reviews of the campaigns for which they are
responsible, at least annually, to ensure that campaign activities are
being executed properly and that campaign milestones are being
completed.
However, NNSA campaign managers have not met all of the NNSA
requirements needed to effectively oversee the Primary and Secondary
campaigns. For example, we found that the campaign managers for the
Primary and Secondary campaigns have not established formal
requirements for conducting annual, technical reviews of the
implementation of QMU at the three weapons laboratories. Moreover,
these officials have not established requirements for certifying the
completion of level-2 milestones that relate to QMU. They could not
provide us with documentation showing the specific activities or
outcomes that they expected from the weapons laboratories in order to
certify that the laboratories had completed the level-2 milestones for
QMU. Instead, they relied more on ad hoc reviews of campaign activities
and level-2 milestones as part of their oversight activities for their
campaigns. According to the Primary campaign manager, the officials at
the weapons laboratories are the principal managers of campaign
activities. As a result, he views his role as more of a "sponsor" for
his program and, therefore, does not require any written reports or
evidence from the laboratories to certify that they have completed
specific milestones.
In contrast, we found that the ASC campaign manager has established
formal requirements for a variety of reoccurring technical reviews of
activities associated with the ASC campaign. Specifically, the ASC
campaign relies on semiannual reviews conducted by the ASC Predictive
Science Committee--which provides an independent, technical review of
the status of level-2 milestones--as well as on annual "principal
investigators" meetings that provide a technical review of every
program element within the ASC campaign. The ASC campaign manager told
us that he relies on these technical reviews to oversee program
activities because the quarterly program review meetings are not meant
to help him manage his program but are really a way for senior-level
NNSA officials to stay informed.
In addition, the ASC campaign manager has established detailed, formal
requirements for certifying the completion of level-2 milestones for
the ASC campaign. Specifically, the fiscal year 2006 implementation
plan for the ASC campaign contains a detailed description of what NNSA
expects from the completion of each level-2 milestone, including a
description of completion criteria, the method by which NNSA will
certify the completion of the milestone, and an assessment of the risk
level associated with the completion of the milestone. The ASC campaign
manager told us that, when NNSA officials created the level-2
milestones for the campaigns in 2003, the milestones were really just
"sentences" and lacked the detailed criteria that would enable NNSA
managers to adequately track and document the completion of major
milestones. As a result, the ASC campaign has made a major effort in
recent years to develop detailed, formal requirements to support the
completion of ASC level-2 milestones.
NNSA Has Not Established Adequate Measures to Determine the
Laboratories' Performance in Developing and Implementing QMU:
NNSA uses performance measurement data to inform resource decisions,
improve the management and delivery of products and services, and
justify budget requests. According to NNSA requirements, performance
measurement data should explain in clear, concise, meaningful, and
measurable terms what program officials expect to accomplish for a
specific funding level over a fixed period of time. In addition,
performance measurement data should include annual targets that
describe specific outputs that can be measured, audited, and
substantiated by the detailed technical milestones contained in
documentation such as campaign implementation plans.
With respect to QMU, NNSA has established an overall annual performance
target to measure the cumulative percentage of progress toward the
development and implementation of the QMU methodology. Specifically, in
its fiscal year 2006 budget request to the Congress, NNSA stated that
it expects to complete the development and implementation of QMU by
2010 as follows:
* 25 percent complete by the end of fiscal year 2005,
* 40 percent complete by the end of fiscal year 2006,
* 55 percent complete by the end of fiscal year 2007,
* 70 percent complete by the end of fiscal year 2008,
* 85 percent complete by the end of fiscal year 2009, and:
* 100 percent complete by the end of fiscal year 2010.
According to NNSA, it had progressed 10 percent toward its target of
completing QMU by the end of fiscal year 2004. However, NNSA officials
could not document how they can measure progress toward the performance
target for developing and implementing QMU. Moreover, NNSA officials
could not explain how the 2010 overall performance target for the
completion and implementation of QMU is related to the level-1
milestones for QMU in the Science campaign program plan, which
describes a two-stage process to identify and reduce key uncertainties
in nuclear weapon performance using QMU by 2014. According to one NNSA
official responsible for overseeing the Primary campaign, NNSA created
this annual performance target because the Office of Management and
Budget requires agencies to express some of their annual performance
targets in percentage terms. However, this official said the actual
percentages are not very meaningful, and he does not have any specific
criteria for how to measure progress to justify the use of the
percentages in the budget request.
NNSA has also established broad performance measures to evaluate the
performance of LANL and LLNL. Specifically, in its performance
evaluation plans for LANL and LLNL for fiscal year 2006, NNSA has
established the following three performance measures:
* Use progress toward quantifying margins and uncertainty, and
experience in application, to further refine and document the QMU
methodology.
* Demonstrate application of a common assessment methodology (i.e.,
QMU) in major warhead assessments and the certification of Life
Extension Program warheads.
* Complete the annual assessment of the safety, reliability, and
performance of all warhead types in the stockpile, including reaching
conclusions on whether nuclear testing is required to resolve any
issues.
However, the plan that NNSA uses to evaluate the performance of SNL
does not contain any performance measures or targets specifically
related to QMU, and the performance evaluation plans for LANL and LLNL
do not contain any annual targets that can be measured and linked to
the specific performance measures related to QMU. Instead, the plans
state that NNSA will rely on LLNL and LANL officials to develop the
relevant targets and related dates for each performance measure, as
well as to correlate the level-1 and level-2 milestones with these
measures. When asked why these plans do not meet NNSA's own
requirements, NNSA officials said that they have not included specific
annual performance targets in the plans because to do so would make it
harder for them to finalize the plans and adjust to changes in NNSA's
budget. However, they said that NNSA is planning on implementing more
stringent plans that will include annual performance targets when the
next contract for LANL and LLNL is developed. In addition, NNSA
officials told us that they recognize the need to develop performance
measures related to QMU for SNL and anticipate implementing these
changes in the fiscal year 2007 performance evaluation plan.
NNSA officials told us that they have used more specific measures, such
as the completion of level-2 milestones, in their assessment of the
weapons laboratories' performance since fiscal year 2004. However, we
also found problems with the way NNSA has assessed the performance of
the weapons laboratories in implementing QMU. For example, in NNSA's
annual performance appraisal of LANL for fiscal year 2004, NNSA states
that LANL had completed 75 percent of the work required to develop "QMU
logic" for the W76 life extension by the end of fiscal year 2004.
However, NNSA officials could not document how they are able to measure
progress toward the development and implementation of QMU logic for the
W76 life extension. Again, an NNSA official responsible for overseeing
the Primary campaign told us that the actual percentages are not very
meaningful, and that he did not have any specific criteria for how to
measure progress to justify the use of the percentage in the appraisal.
In a recent report, we recognized the difficulties of developing useful
results-oriented performance measures for programs such as those geared
toward research and development programs.[Footnote 11] For programs
that can take years to observe program results, it can be difficult to
identify performance measures that will provide information on the
annual progress they are making toward achieving program results.
However, we also recognize that such efforts have the potential to
provide important information to decision makers.
NNSA officials told us that they recognize the need for developing
appropriate measures to ensure that adequate progress is being
maintained toward achieving the goals and milestones of the campaigns.
However, according to NNSA, very few products of the scientific
campaigns involve the repetition of specific operations whose costs can
be monitored effectively as a measure of performance. As a result, the
best measure of progress for the scientific campaigns is through
scientific review by qualified technical peers at appropriate points in
the program. However, NNSA has not established any performance measures
or targets for implementing QMU that require periodic scientific peer
reviews or define what is meant by "appropriate" points in the program.
Conclusions:
Faced with an aging nuclear stockpile, as well as an aging workforce,
NNSA needs a methodologically rigorous, transparent, and explainable
approach for how it will continue to assess and certify the safety and
reliability of the nation's nuclear weapons stockpile, now and into the
foreseeable future, without underground testing. After over a decade of
conducting stockpile stewardship, NNSA's selection of QMU as its
methodology for assessment and certification represents a positive step
toward a methodologically rigorous, transparent, and explainable
approach that can be carried out by a new cadre of weapons designers.
However, important technical and management details must be resolved
before NNSA can say with certainty that it has a sound and agreed upon
approach.
First, NNSA must take steps to ensure that all three nuclear weapons
laboratories--not just LANL and LLNL--are in agreement about how QMU is
to be defined and applied. While we recognize that there will be
methodological differences between LANL and LLNL in the detailed
application of QMU to specific weapon systems, we believe that it is
fundamentally important that these differences be understood and, if
need be, reconciled, to ensure that QMU achieves the goal of a common
methodology with rigorous, quantitative, and explicit criteria, as
envisioned by the original 2003 white paper on QMU. More importantly,
we believe that SNL has an important role in the development and
application of QMU to the entire warhead, and we find the continuing
disagreement over the application of QMU to areas outside of the
nuclear explosive package to be disconcerting. There have been several
recommendations calling for a new, technical paper defining QMU, as
well as the establishment of regular forums to further develop the QMU
methodology and reconcile any differences in approach. We believe the
NNSA needs to fully implement these recommendations.
Second, NNSA has not made effective use of its current planning and
program management structure to ensure that all of the research needed
to support QMU is integrated and that scarce scientific resources are
being used efficiently. We believe that NNSA must establish an
integrated management approach involving planning, oversight, and
evaluation methods that are all clearly linked to the overall goal of
the development and application of QMU. In particular, we believe that
NNSA needs clear, consistent, and realistic milestones and regular,
technical reviews of the development of QMU in order to ensure sound
progress. Finally, while we support the development of QMU and believe
it must be effectively managed, we also believe it is important to
recognize and acknowledge that the development and application of QMU,
especially the complexities involved in analyzing and combining
uncertainties related to potential failure modes and performance
margins, represents a daunting research challenge that may not be
achievable in the time constraints created by an aging nuclear
stockpile.
Recommendations for Executive Action:
To ensure that the weapons laboratories will have the proper tools in
place to support the continued assessment of the existing stockpile or
the certification of redesigned nuclear components under the RRW
program, we recommend that the Administrator of NNSA take the following
two actions:
* Require the three weapons laboratories to formally document an agreed
upon, technical description of the QMU methodology that clearly
recognizes and reconciles any methodological differences.
* Establish a formal requirement for periodic collaboration between the
three weapons laboratories to increase their mutual understanding of
the development and implementation of QMU.
To ensure that NNSA can more effectively manage the development and
implementation of QMU, we recommend that the Administrator of NNSA take
the following three actions:
* Develop an integrated plan for implementing QMU that contains (1)
clear, consistent, and realistic milestones for the development and
implementation of QMU across the weapons complex and (2) formal
requirements for certifying the completion of these milestones.
* Establish a formal requirement for conducting annual, technical
reviews of the scientific research conducted by the weapons
laboratories that supports the development and implementation of QMU.
* Revise the performance evaluation plans for the three weapons
laboratories so that they contain annual performance targets that can
be measured and linked to specific milestones related to QMU.
Agency Comments and Our Evaluation:
We provided NNSA with a draft of this report for their review and
comment. Overall, NNSA agreed that there was a need for an agreed-upon
technical approach for implementing QMU and that NNSA needed to improve
the management of QMU through clearer, long-term milestones and better
integration across the program. However, NNSA stated that QMU had
already been effectively implemented and that we had not given NNSA
sufficient credit for its success. In addition, NNSA raised several
issues about our conclusions and recommendations regarding their
management of the QMU effort. The complete text of NNSA's comments on
our draft report is presented in appendix I. NNSA also made technical
clarifications, which we incorporated in this report as appropriate.
With respect to whether QMU has already been effectively implemented,
during the course of our work, LANL and LLNL officials showed us
examples of where they used the QMU methodology to examine specific
issues associated with the stockpile. At the same time, during our
discussions with laboratory officials, as well as with the Chairs of
the JASON panel on QMU, the Office of Defense Programs Science Counsel,
and the Strategic Advisory Group Stockpile Assessment Team of the U.S.
Strategic Command, there was general agreement that the application of
the QMU methodology was still in the early stages of development. As
NNSA pointed out in its letter commenting on our report, to implement
QMU, the weapons laboratories need to make a number of improvements,
including techniques for combining different kinds of uncertainties, as
well as developing better models for a variety of complex processes
that occur during a nuclear weapon explosion. In addition, the
successful implementation of QMU will continue to rely on the expert
judgment and the successful completion of major scientific facilities
such as the National Ignition Facility. We have modified our report to
more fully recognize that QMU is being used by the laboratories to
address stockpile issues and to more completely characterize its
current state of development. At the same time, however, because QMU is
still under development, we continue to believe that NNSA needs to make
more effective use of its current planning and program management
structure.
NNSA raised several specific concerns about our conclusions and
recommendations. First, NNSA disagreed with our conclusion and
associated recommendations that NNSA take steps to ensure that all
three nuclear weapons laboratories are in agreement about how QMU is to
be defined and applied. NNSA stated that we overemphasized the
differences between LANL and LLNL in implementing QMU and that,
according to NNSA, LANL and LLNL have a "common enough" agreement on
QMU to go forward with its implementation. Moreover, NNSA stated that
our recommendations blur very clear distinctions between SNL and the
two nuclear design labs. According to NNSA, QMU is applied to issues
regarding the nuclear explosive package, which is the mission of LANL
and LLNL.
While we believe that some of the technical differences between the
laboratories remain significant, we have revised our report to more
accurately reflect the nature of the differences between LANL and LLNL.
With respect to SNL, we would again point out that SNL officials are
still required to certify the performance of nuclear weapon components
under the conditions of a nuclear explosion and, thus, use similar
elements of the QMU methodology. Therefore, we continue to believe that
all three laboratories, as well as NNSA, would benefit from efforts to
more formally document the QMU methodology and regularly meet to
increase their mutual understanding. As evidence of the benefits of
this approach, we would note that LLNL and LANL are currently
developing a revised "white paper" on QMU, and that in discussions with
one of the two authors, he agreed that inclusion of SNL in the
development of the draft white paper could be beneficial.
Second, NNSA made several comments with respect to our recommendation
that NNSA develop an integrated plan for implementing QMU that contains
clear, consistent, and realistic milestones. For example, NNSA stated
that they expect to demonstrate the success of the implementation of
QMU and the scientific campaigns by the performance of a scientifically
defensible QMU analysis for each required certification problem. In
addition, NNSA stated that the 2010 budget target and the 2014
milestone were developed for different purposes and measure progress at
different times. According to NNSA, the 2010 target describes
developing QMU to the point that it can be applied to certification of
a system (e.g., the W88) without underground testing, while the 2014
milestone is intended to be for the entire Science campaign effort.
However, as we state in our report, and as acknowledged by NNSA
officials responsible for the Primary and Secondary campaigns, there
continue to be problems with the milestones that NNSA has established
for implementing QMU. Among these problems is the fact that these
milestones are not well-defined and conflict with other performance
measures that NNSA has established for QMU. Moreover, in its comments
on our report, NNSA agreed that better integration and connectivity of
milestones between various program elements would improve the
communications of the importance of program goals and improve the
formality of coordination of program activities, "which is currently
accomplished in an informal and less visible manner." Given this
acknowledgment by NNSA, we continue to believe that an integrated plan
for implementing QMU, rather than NNSA's current ad hoc approach, is
warranted.
Third, NNSA made several comments regarding our recommendation that
NNSA establish a formal requirement for conducting annual, technical
reviews of the scientific research conducted by the weapons
laboratories that supports the development and implementation of QMU.
NNSA stated that it believes the ad hoc reviews it conducts, such as
the JASON review, provide sufficient information on scientific
achievements, difficulties, and required redirection to manage these
programs effectively. As a result, NNSA stated that it has not selected
a single review process to look at overall success in the
implementation of QMU but expects to continue to rely on ad hoc
reviews.
We agree that reviews, such as the JASON review, are helpful, and we
relied heavily on the JASON review, as well as other reviews as part of
our analysis. However, as we point out in the report, the issue is that
the campaign managers for the Primary and Secondary campaigns do not
meet all of NNSA's own requirements for providing effective oversight,
which include the establishment of formal requirements for conducting
technical reviews of campaign activities. Therefore, we believe that
NNSA needs to take steps to implement its own policies. In addition, we
believe that the ASC campaign provides a good role model for how the
Primary and Secondary campaigns should be managed.
Finally, NNSA made several comments with respect to our recommendation
for NNSA to revise the performance evaluation plans for the
laboratories so that they contain annual performance targets that can
be measured and linked to specific milestones related to QMU.
Specifically, NNSA stated that the implementation of QMU is an area
where it is difficult to establish a meaningful metric. According to
NNSA, since QMU is implicitly evaluated in every review of the
components of the science campaign, NNSA does not believe it is
necessary to formally state an annual QMU requirement. However, as we
point out in the report, the current performance evaluation plans for
LANL and LLNL do not meet NNSA's own requirements for the inclusion of
annual performance targets that can be measured and linked to the
specific performance measures related to QMU. More fundamentally, since
NNSA has placed such emphasis on the development and implementation of
QMU in the years ahead, we continue to believe that NNSA needs to
develop more meaningful criteria for assessing the laboratories'
progress in developing and implementing QMU.
We are sending copies of this report to the Administrator, NNSA; the
Director of the Office of Management and Budget; and appropriate
congressional committees. We also will make copies available to others
upon request. In addition, the report will be available at no charge on
the GAO Web site at [Hyperlink, http://www.gao.gov].
If you or your staff have any questions about this report or need
additional information, please contact me at (202) 512-3841 or
[Hyperlink, aloisee@gao.gov]. Contact points for our Offices of
Congressional Relations or Public Affairs may be found on the last page
of this report. GAO staff who made major contributions to this report
are listed in appendix II.
Signed by:
Gene Aloise:
Director, Natural Resources and Environment:
Appendixes:
Appendix I: Comments from the National Nuclear Security Administration:
Department of Energy:
National Nuclear Security Administration:
Washington, DC 20585:
JAN 10 2006:
Mr. Gene Aloise:
Director:
Natural Resources and Environment:
U.S. Government Accountability Office:
Washington, D.C. 20548:
Dear Mr. Aloise:
The National Nuclear Security Administration (NNSA) appreciates the
opportunity to review the Government Accountability Office's (GAO)
draft report, GAO-06-261, "NUCLEAR WEAPONS: NNSA Needs to Refine and
More Effectively Manage Its New Approach for Assessing and Certifying
Nuclear Weapons." NNSA understands that the House Strategic Forces
Subcommittee, Committee on Armed Services, originally requested GAO to
determine how NNSA currently defines the scientific research portion of
its campaign that is intended to provide a safe and reliable stockpile.
During the course of this audit, the scope of the audit evolved into a
review of the Quantification of Margins and Uncertainties (QMU)
methodology for assessing and certifying the stockpile.
While NNSA agrees that there must be an agreed-upon technical approach
to QMU implementation and that NNSA should always strive to improve the
management of QMU implementation, we believe that QMU has been
implemented as an effective approach to stockpile certification. The
present implementation of QMU is highly effective in bringing science
to stockpile issues and used for weapons certification issues across
the stockpile, as well as being the basis for the Laboratory Directors'
recommendations in the annual assessment reports on the stockpile.
Ad hoc scientific reviews conducted by panels such as JASONs, the
University of California Science and Technology Panel, and the
Strategic Commands' Strategic Advisory Group Stockpile Assessment Team
(SAGSAT) are appropriate fora for assessing scientific programs in
areas depending on the implementation of QMU. Those reviews have
demonstrated the steady and rapid progress in the application of QMU to
weapons certification since the initial 2003 white-paper on QMU
implementation. The success in the development of QMU is an
accomplishment resulting from a decade of scientific progress since the
establishment of Stockpile Stewardship in 1995. Continued progress in
key science areas in primary and secondary physics, materials science,
and high energy density physics, including the National Ignition
Campaign, and computational advances are required to sustain future
certification requirements.
We have enclosed two documents for GAO's consideration prior to the
publication of the final report. The first document addresses the
background for QMU and what the Program believes to be the maturity
level of the QMU process. The second document is detailed technical
comments for your consideration.
Should you have any questions related to this response, please contact
Richard Speidel, Director, Policy and Internal Controls Management.
Sincerely,
Signed by:
Michael C. Kane:
Associate Administrator for Management and Administration:
Enclosures:
cc: Deputy Administrator for Defense Programs:
Senior Procurement Executive:
Director, Service Center:
NNSA Response to the GAO report, GAO-06-261, "NUCLEAR WEAPONS: NNSA
Needs to Refine and More Effectively Manage Its New Approach for
Assessing and Certifying Nuclear Weapons."
Executive Summary:
Because of a successful record of progress in the development of the
Quantification of Margins and Uncertaes (QMU) approach for certifying
nuclear lear warheads, the National Nuclear Security Administration
(NNSA) is not seeking further refinements beyond the currently
envisioned program of work. NNSA will, however, seek management
improvements in implementing this approach.
Despite the conclusions of the GAO audit of NNSA's QMU program, the
NNSA has already implemented QMU as an effective approach to stockpile
certification.
* The present implementation of QMU is highly effective in bringing
science to stockpile issues.
* QMU is now used for weapons certification issues across the stockpile
and is the basis for the Laboratory Directors' recommendations in the
annual assessment reports on the stockpile.
* The ad hoc scientific reviews conducted by panels such as JASONs,
University of California Science and Technology Panel, and the
Strategic Commands' Strategic Advisory Group Stockpile Assessment Team
(SAGSAT) are appropriate fora for assessing scientific programs in
areas depending on the implementation of QMU.
* These reviews have demonstrated steady and rapid progress in the
application of QMU to weapons certification since the initial 2003
white paper on QMU implementation.
* The success in developing QMU is a key accomplishment resulting from
a decade of outstanding scientific progress since the establishment of
Stockpile Stewardship in 1995.
* Continued progress in key science areas in primary and secondary
physics, materials science, and high energy density physics, including
the National Ignition Campaign, and computational advances will be
required to sustain future certification requirements.
NNSA recognizes that design and certification of a Reliable Replacement
Warhead (RRW) as well as transformation of the nuclear weapons complex
to meet newly identified responsive infrastructure goals pose new
challenges. These will require a careful review of science campaign
priorities and will require better integration across NNSA activities.
The recent completion of the revised Work Breakdown Structure for
Advanced Simulation and Computing (ASC) and the completion of the
Primary Assessment Plan are initial steps in that process. The QMU
approach can be managed as an integrating influence across program
components.
THE NATIONAL NUCLEAR SECURITY ADMINISTRATION (NNSA) HAS IMPLEMENTED
QMU:
Despite the contention of the GAO that because of management
shortcomings NNSA is likely to have difficulties in implementing QMU,
the Department of Energy maintains that it has successfully implemented
QMU. This methodology constitutes the framework by which the Directors
of the Nuclear Weapons National Laboratories, through the Secretaries
of Energy and Defense, execute their statutory responsibility to assure
the President of the United States of the safety, security and
reliability of the U.S. nuclear deterrent. It has visibility, oversight
and management from the highest levels of the government, the national
laboratories, and the august scientific bodies that provide advice to
the Administration, to its agencies, and to the Congress. Not
appreciating the demonstrated success in the implementation of QMU has
led to unfounded conclusions that because of management failings QMU is
likely to fail in the future and that important efforts to transform
the stockpile may be at risk.
QMU is a framework for connecting the scientific method to a variety of
questions regarding assessment of the stockpile and for presenting the
results. Review of QMU requires a scientific evaluation of progress in
provng objective, technically based answers to complex questions that
arise in the prediction of the performance, safety and reliability of
the stockpile. Its utility is best judged not in the abstract but in
the context of the ability to solve specific problems, in this case,
the set of specific issues that must be settled in order to certify
specific devices.
The report is critical of ad hoc reviews to measure the progress in
QMU. Because the value of QMU is most meaningfully weighed by
evaluating technical progress in specific applications, however, NNSA
relies on ad hoc expert reviews of the application of QMU to specific
problems as the best review mechanism. The JASON review of QMU and an
ongoing JASON review on pit lifetimes, which is an application of QMU
in a vital area, are both examples of such reviews. Several Strategic
Advisory Group Stockpile Assessment Team (SAGSAT) reviews of specific
stockpile certification issues are additional examples of useful
reviews.
Each year the NNSA and other organizations conduct numerous reviews
that cover the broad gamut of efforts within the science campaign and
in particular on subjects where QMU plays a vital role. While there are
a number of drivers to conduct reviews, NNSA is cognizant of the high
programmatic costs on the laboratories to support these and is hesitant
to add to their number unless given good justification. NNSA believes
that the reviews it conducts and those of which it has cognizance
provide sufficient information on scientific achievements, difficulties
and required redirection to manage these programs effectively. This GAO
audit itself relies in part on the results of those same ad hoc
reviews, but appears to undervalue them.
Despite the characterization by the GAO report, the development of QMU
and its present application to the broad range of certification issues
facing the national laboratories is a significant and vital
accomplishment. It represents progress brought about through a
sustained decade long effort in implementing the charge of the FY 1994
National Defense Authorization Act which directed the Secretary of
Energy to "establish a stewardship program to ensure the preservation
of the core intellectual and technical competencies of the United
States in nuclear weapons, including weapons design, system
integration, manufacturing, security, use control, reliability
assessment, and certification."
In response, DOE developed the 1995 Science Based Stockpile Stewardship
program, which set out the vision that DOE has subsequently followed,
with few modifications. Important efforts included the establishment of
Accelerated Strategic Computing Initiative (ASCI, now ASC),
revitalization of the Inertial Confinement Fusion (ICF) program
including high energy density physics, efforts in hydrodynamic
experiments and facilities, and a variety of experimental efforts to
improve understanding of materials properties crucial to prediction of
weapons performance.
In order to better organize the program, establish more specific goals,
track progress, and provide a level of transparency to its sponsors,
DOE created the campaign program management structure in 1999, creating
the six science campaign efforts that are the subject of this GAO
report. Efforts begun in response to the 1995 program had by 2002
achieved substantial improvements in capabilities. These included; the
development of primary and secondary bum codes and the improved
computational capability provided by Advanced Simulation and Computing
(ASC); improved understanding of underlying phenomenology through
experimental successes in hydrotesting and the subcritical experiments
program; successes in the area of high energy density physics; improved
understanding of the properties and aging of nuclear weapon materials
and components; and, improved analysis of historical underground
nuclear tests.
The progress in these underlying capabilities enabled the development
of QMU requested by NNSA and described in the seminal 2003 QMU paper by
Dr. Bruce Goodwin and Dr. Ray Juzaitis referred to in the report.
Subsequent progress has been rapid, from the partial level of
application of QMU shown in the NNSA Science Council Review (the 2004
Friendly Review), the increased progress shown in the 2004 JASON review
of QMU, finally, to the application in 2005 of QMU to all annual
weapons systems assessment and certifications.
In the implementation of the underlying 1994 Congressional charge and
1995 program, QMU represents a transformation from certification based
on the individual judgment of designers grounded in the success of the
underground test program, to more quantitative and objective results.
As the report notes, "QMU seeks to quantify (1) how close each critical
factor is to the point at which it would fail to perform as designed
(i.e., the margin to failure) and (2) the uncertainty that exists in
calculating the margin, in order to ensure that the margin is
sufficiently larger than the uncertainty." It is in the formal
development and presentation of quantitative analyses that QMU enables
the articulation of institutional conclusions where results can be
analyzed, repeated, and any differences can be reconciled through inter-
laboratorypeer review processes.
While this statement of the basics of QMU is relatively simple and
easily understood, the complexity occurs in the detailed application to
specific weapons systems and performance issues because, for each
weapons system, the potential failure modes for the system must be
identified, margins for each established, and a thorough analysis of
uncertainties in establishing those margins must be applied. The
uncertainty analysis is complex because it must convolve information
about manufacturing variability, uncertainties of physical
understanding of complex physical phenomenon, and the limitations of a
sparse set of underground and aboveground test data.
GAO claims that "absent the prompt resolution of remaining
disagreements over the definition and implementation of QMU, it is
unclear whether the weapons laboratories will have the common framework
they say they need to support the continued assessment of the existing
stockpile or the certification of redesigned nuclear components under
the RRW program." NNSA believes this statement is incorrect. The
laboratories have a common enough agreement on QMU, and the
definitional differences are well enough appreciated that any
difficulties in certifying an RRW, or any other system will be the
result of a currently unforeseen fundamental technical issue rather
than a definitional dispute. The outcome of a scientific debate does
not depend upon the definition of words, but evidence developed to
answer a question. NNSA's observation at working sessions with the
laboratories is that scientists now quickly get down to the hard work
of understanding fundamental differences in the outcome of experiments
or predictions of computer models rather than argue over approaches.
To be precise, the Directors of the Nuclear Weapons National
Laboratories certify that a nuclear warhead will meet the "Military
Characteristics" under a specified "Stockpile to Target Sequence."
While it is nuclear weapons that are certified and not individual
components, QMU within the context of the six science campaign efforts
that were the subject of the GAO report, is applied to issues regarding
the nuclear explosives package, which is the mission of the nuclear
design laboratories, Lawrence Livermore National Laboratory and Los
Alamos National Laboratory. While Sandia National Laboratories has its
own applications of the QMU methodology, and communications and sharing
of techniques occurs among all three laboratories, the GAO
recommendations blur very clear distinctions between Sandia and the two
nuclear design laboratories.
Nevertheless, the Sandia approach is reconcilable with the approach
used by the nuclear design laboratories, as is to be expected since the
Sandia manager who was responsible for developing their certification
methodologies is now at LANL responsible for the development and
implementation of LANL's certification methodology. Nevertheless, the
specific problems to be solved are different, using different codes,
models and experimental tools in very different physical regimes. In
those areas where interchange can usefully occur, it happens and will
continue, such as the development of statistical techniques.
In overemphasizing the level of difference, the GAO also underplays the
role of complementary approaches, the importance of which is the
rationale for a national policy decision to maintain two nuclear design
laboratories. First, as noted, weapons certification involves
scientific research where the outcome is not a prori known and
confidence in the correctness of the result is improved by achieving
the same answer through multiple approaches. The implication that lack
of an identical approach motivates a need to "refine" the approach,
suggests a misunderstanding of the nature of scientific approaches to
complex problems. Generally, a scientific result is accepted not
because of uniformity of methods, but because multiple researchers can
reproduce the same result using different techniques and approaches.
Likewise, GAO misunderstands LLNL's approach in finding that the LLNL
combines QMU ratios for different failure modes into a single ratio for
the entire warhead. LLNL does not do this. (LLNL combines uncertainties
within a given failure mode.) In fact, all labs treat each failure mode
separately. Inadequate margin for any failure mode represents a risk
that the weapon will fail to meet requirements. Excess margin in one
area, e.g., primary yield cannot generally compensate for inadequate
margin in another area, e.g., one-point safety.
To be sure, NNSA continues to perform research because there are vital
questions that are unanswered and vital capabilities that must be
improved to ensure the long-range health of the deterrent as the
stockpile ages or as replacement systems are introduced. Nevertheless,
the GAO statement that "the weapons laboratories face extraordinary
technical challenges in successfully implementing a coherent and
credible analytical method based on QMU" is without context. Without
noting the significant achievements to date, the statement leaves the
reader with the conclusion that success in this vital area is unlikely
and efforts to certify an RRW will likely fail despite the noted
successes to date. NNSA disagrees.
There is no question that improving the ability to meet future
certification requirements will require further improvements in a
number of areas. The development of QMU is only one aspect of the
science campaign effort that includes, importantly, the underlying
scientific effort to improve physical understanding and reduce
uncertainties resulting from data and models. Methods for combining
differing kinds of uncertainties constrained by the sparse data set
from underground testing is groundbreaking research at the forefront of
statistical science. Models for boost, mix, the high-pressure equation
of state of plutonium, the behavior of dense plasmas, and a range of
other physical phenomena require refinement. Numerical methods of
computation require improvement, the 2nd axis of DARHT must be
commissioned and NIF ignition must be achieved. Certification
requirements are a significant consideration in decisions regarding the
future of LANSCE. To state that current certification methods require
refinement is to state the obvious. But it is wrong to portray the
program as lacking clear goals, or technically defensible standards for
success.
Success is demonstrated by the performance of a scientifically
defensible QMU analysis for each required certification problem. While
there can be other measures of program efficiency, there is no other
comparable measure to determine whether or not the program is achieving
its scientific goals.
Although the GAO report is focused on QMU, this grew out of an audit of
six science efforts, including the science campaign, ASC, and the ICF
program. These programs have multiple goals and achievements beyond the
specifc focus of QMU although they are supportive of that goal. One of
the functions of QMU, as it is further applied will be to identify
research priorities within the science campaign. As the JASON QMU
report cautions, however, "prioritization of efforts has to be
modulated by the need to maintain expertise across the entire weapon
system, and its processes. That is, a baseline of effort needs to be
maintained across all activities, including those judged to be of lower
priority." Of course, less effort should be put into lower-priority
activities (i.e., those bearing on processes with higher margins
relative to uncertainties), but there needs to be enough ongoing
activity even regarding "reliable" (high- margin) processes in order to
maintain expertise and to allow for the possibility of revising
previous estimates of reliability (and responding to those revisions)
or to address unforeseen conditions (e.g., significant findings in
surveillance).
The United States has, since the inception of the Manhattan Project,
relied upon world-class science to support confidence in the nation's
nuclear deterrent, and is likely to continue to do so. QMU is the
current framework for the application of that science base to
establishing confidence. Questions and issues regarding safety,
performance and reliability of the stockpile will, so far as one can
foresee, continue to occur, and therefore the continued development and
refinement of QMU or some follow-on certification methodology will
continue to be required. Therefore, the impression that one can define
certain milestones, the achievement of which will indicate that the
development of QMU is finished, is misleading.
The NNSA fundamentally disagrees with the methodology of trying to
measure scientific progress through an audit largely reliant on review
of administrative documents and disputes some of the conclusions
derived therefrom. The QMU methodology has already proven successful
and is unlikely to be the source of future failings in the program.
However, more detailed responses to a few of the specific findings and
management recommendations, not already covered, are provided.
RESPONSE TO THE GAO FINDINGS ON MANAGEMENT OF SCIENTIFIC EFFORTS:
The GAO report criticizes the management tools and methods used to
administer the Science Campaign, leading one to conclude that the
Science Campaign lacks clear goals and has lacked substantial
achievements. In fact, the Science Campaign Program Plan has had clear
statements of long-term goals that have remained largely unchanged
since the inception of the Science Campaign, and important progress has
been made in key areas. For instance, the campaign has achieved a
vastly improved understanding of plutonium properties under extreme
conditions resulting from the subcritical experiments program, and
increased accuracy of plutonium equation of state data obtained from
the recently commissioned JASPER experiment. Significant new insight
has been gained on an important problem in understanding the energy
balance in nuclear weapons. Understanding of mix sensitivities has been
vastly improved, and these insights will provide direction for future
experimental and modeling efforts. New materials damage models have
been developed and implemented into ASC codes and experimental data is
being acquired to establish important parameters in that model.
Kinetics models for high explosives performance have been developed and
implemented into weapons codes.
The underlying assumption that the science campaigns should respond to
and be measured by a directed set of milestones provides an incomplete
picture. By way of contrast, however, the GAO correctly states, "The
Primary and Secondary campaigns were established to analyze and
understand the different scientific phenomena that occur in the primary
and secondary stages of a nuclear weapon during detonation. As such,
the Primary and Secondary campaigns are intended to support the
development and implementation of the QMU methodology and to set the
requirements for the computers, computer models, and experimental data
needed to assess and certify the performance of nuclear weapons." While
these campaigns have long-term goals towards which they are making
progress they also perform required research to determine the
comprehensive requirements for other elements of the program.
Before this audit had begun, the NNSA had identified that in view of
the recent progress in areas such as QMU, long-term goals need to be
refined and restated, and better integration across the program is
required. Therefore the NNSA had begun the development of the Primary
Assessment Plan referred to in the GAO report. This plan identifies key
level 1 milestones that must be supported by primary certification
capabilities, and indicates priorities for achieving improvements in
those science areas that will be required to support those goals. A key
focus is on Reliable Replacement Warhead certification. The next step
will be to identify those level 2 milestones that are needed to support
the long-term goals, though in the immediate future, these are unlikely
to change to a significant degree from present goals.
NNSA has the following responses to some of the specific report
findings:
Finding: First. the planning documents that NNSA has established for
the Primary and Secondary campaigns do not adequately integrate the
scientific research currently conducted that supports the development
and implementation of OMU.
Response: The NNSA agrees with this statement; this is the motivation
for the development of the Primary Assessment Plan and the subsequently
planned Secondary Assessment Plan. In addition, NNSA will develop
further guidance to the program on science integration associated with
QMU.
Finding: Second. NNSA has not developed a clear, consistent set of
milestones to guide the development and implementation of QMU. For
example. while one key campaign plan envisions a two-stage path to
complete the development of OMU by 2014. the performance measures in
NNSA's fiscal year 2006 budget request call for the completion of OMU
by 2010.
Response:
NNSA agrees that better integration and connectivity of milestones
between various program elements would improve the communications of
the importance of program goals and improve the formality of
coordination of program activities, which is currently accomplished in
an informal and less visible manner. This will be done in part through
more careful coordination of level one and level two milestones. An
NNSA Headquarters team will provide additional program guidance on
science integration supporting QMU and will seek to clarify PART
measures.
At the same time, the GAO analysis of the milestones shown in Table 4
of the report is not entirely accurate. The table shows the level-one
milestones for Science Campaign for the period from 2007 to 2014. These
are milestones are not just for QMU but for the entire science
campaign, of which QMU is only a part. For instance GAO cites the FY
2014 milestone "accounting for simulation and experimental
uncertainties, reassess the ability to reproduce the full underground
test data sets for a representative group of nuclear tests with a
consistent set of models. "To meet this milestone, NNSA must have
completed the development of a full set of improved physics models,
including improved mix and boost models, improved plutonium damage and
equation of state models, and improved models for secondary
performance. These models must have been validated and incorporated
into ASC codes. This also requires developing techniques, under QMU, to
perform the required uncertainty analysis. The milestone anticipates
success and integration of all of these factors.
The GAO also claim that NNSA lacks milestones for the development and
application of QMU, but the report itself lists level-two milestones
for the development of certification plans for the W76 and W88 based on
QMU, milestones of national significance, which have recently been
completed.
The 2010 milestone and 2014 milestone were developed for different
purposes and measure progress at different times. The 2010 milestone
was developed to respond to a requirement of the Office of Management
and Budget (OMB) under the government-wide Performance and Rating Tool
(PART) system to establish and report on a few programmatically
significant long-term milestones. A list of accomplishment was
developed with annual progress goals and a completion date of 2010 as
directed by the OMB. The PART target describes developing QMU to the
point that it can be applied to certification of a system without
underground testing (e.g. LANL manufactured W88 pit). The 2014
milestone refers, however, to a more complete development and more
complex application of this approach for a series of weapons tests.
Therefore, saying that the OMB PART target would be completed in 2010
is a target distinct from the statement that the broader Science
Campaign milestone would be completed in 2014.
Finding: Third. NNSA has not established formal requirements for
conducting annual. technical reviews of the implementation of OMU at
the three weapons laboratories or for certifying the completion of OW-
related milestones.
The issue of ad hoc reviews has been addressed in the overview. The
programs at the national laboratories are reviewed on a frequent basis
established to meet a wide variety of customer requirements, and QMU is
integral to most of those reviews. Relevant periodic reviews include
the University of California Division Review Committees, the Strategic
Command Strategic Advisory Committee's Stockpile Assessment Team
(SAGSAT), periodic reviews of the W76 LEP, and W88 pit certification. A
recent review of the Subcritical Experiments Advisory Committee to
ensure the subcriticality of the proposed Unicorn experiment was a
review of the QMU methodology applied to this important safety question
and noted excellent progress in the application of the QMU methodology.
The GAO cites with approval the "Predictive Science Panel" chartered
under the ASC program, which is a panel of outside experts, not NNSA
staff. The purview of this panel encompasses exactly those parts of
both ASC and the science campaigns that are relevant to the development
of tools, models and methods that support the development of predictive
capabilities, and therefore QMU.
NNSA has not selected a single review process to look at overall
success in the implementation of QMU but expects to continue to rely on
ad hoc reviews.
Finding: Finally. NNSA has not established adequate performance
measures to determine the progress of the laboratories in developing
and implementing QMU.
Response: The NNSA has established level 1 milestones in the Primary
Assessment Plan which incorporate, implicitly QMU goals. The extensive
set of external reviews discussed on page 2 of this response provide
ample opportunity to determine the progress in implementing QMU.
Finding: According to NNSA, very few products of the scientific
campaigns involve the repetition of specific operations whose costs can
be monitored effectively as a measure of performance. As a result, the
best measure of progress for the scientific campaigns is through
scientific review by qualified technical peers at appropriate points in
the program. However. NNSA has not established any performance measures
or targets for implementing QMU that require periodic scientific peer
reviews or define what is meant by "appropriate" points in the program.
Response: Scientific peer reviews will be continued to evaluate
progress in addressing scientific issues. One weighs the scientific
information that has been developed against the problem to be solved.
As stated, NNSA does have targets for accomplishing certain specific
tasks, such as writing certification plans. But to have a metric or
quantifiable targets suggests that one already has an answer or enough
of one that one can define a meaningful measurable outcome.
For those things that can be scheduled and usefully counted, the
Science Campaign already does so. For instance the NNSA has established
a detailed plan for completing the DARHT 2°a axis with well-defined
milestones. NNSA tracks the operating days at LANSCE, again, because
this is an important indicator of facility operating efficiency. NNSA
tracks the number of experiments performed on JASPER and the costs
thereof because this bears on the productivity of the facility and also
is a surrogate for the rate of progress in accumulating important
plutonium equation of state data. In none of these cases, however, does
the metric substitute for an actual evaluation of scientific knowledge
gained.
The implementation of QMU is one of those examples where it is
difficult to establish a meaningful metric. NNSA chartered a review, in
this case the JASON QMU review, to examine the application of QMU in
specific instances, evaluate its adequacy, look at weakness and suggest
future directions. A future additional review by JASON will be
considered. Since QMU is implicitly evaluated in every review of the
components of the science campaign, NNSA does not view it as necessary
to formally state an annual QMU requirement.
In summary, NNSA believes that it has achieved substantial progress to
date in developing both QMU and meeting other goals of the science
campaign, through appropriate management focus and oversight. At the
same time, NNSA agrees and has recognized that the growing immediacy of
meeting new requirements for both the Reliable Replacement Warhead and
responsive infrastructure require a reevaluation of the level of
coordination and integration of goals and milestones across all NNSA
programs. The completion of the Primary Assessment Plan was one step in
a number of envisioned efforts to reassess priorities and improve the
level of coordination.
[End of section]
Appendix II: GAO Contact and Staff Acknowledgments:
GAO Contact:
Gene Aloise (202) 512-3841:
Staff Acknowledgments:
In addition to the individual named above, James Noel, Assistant
Director; Jason Holliday; Keith Rhodes; Peter Ruedel; and Carol
Herrnstadt Shulman made key contributions to this report.
(360508):
FOOTNOTES
[1] The National Defense Authorization Act for Fiscal Year 1994, Pub.
L. No. 103-160, § 3135 (1993), directed DOE to establish the Stockpile
Stewardship Program.
[2] Modern nuclear weapons have two stages: the primary, which is the
initial source of energy, and the secondary, which is driven by the
primary and provides additional explosive energy.
[3] JASON is a group of nationally known scientists who advise
government agencies on defense, energy, and other technical issues.
[4] The terms "nuclear warhead" and "nuclear weapon" have different
technical meanings. For example, a nuclear weapon, in the case of a
reentry vehicle, includes the warhead and certain Department of Defense
components, such as fuses and batteries. However, for purposes of this
report, we often use the terms "warhead" and "weapon" interchangeably.
[5] The Defense Authorization Act for Fiscal Year 2003, Pub. L. No. 107-
314, § 3141 (2002), established a statutory requirement for annual
stockpile assessments.
[6] GAO, Nuclear Weapons: Preliminary Results of Review of Campaigns to
Provide Scientific Support for the Stockpile Stewardship Program, GAO-
05-636R (Washington, D.C.: Apr. 29, 2005).
[7] National Nuclear Security Administration Advisory Committee,
"Science and Technology in the Stockpile Stewardship Program," Mar. 1,
2002.
[8] LLNL first applied QMU in its certification of the life extension
of the W87, which was completed in November 2004.
[9] NNSA Defense Programs Science Council, "Report on the Friendly
Reviews of QMU at the NNSA Laboratories," March 2004.
[10] JASON, The MITRE Corporation, Quantification of Margins and
Uncertainties (QMU), JSR-04-330, Feb. 17, 2005.
[11] GAO, Performance Budgeting: PART Focuses Attention on Program
Performance, but More Can Be Done to Engage Congress, GAO-06-28
(Washington, D.C.: Oct. 28, 2005).
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: