National Nuclear Security Administration
Additional Actions Needed to Improve Management of the Nation's Nuclear Programs
Gao ID: GAO-07-36 January 19, 2007
In response to security and management weaknesses, in 1999 the Congress created the National Nuclear Security Administration (NNSA), a separately organized agency within the Department of Energy (DOE). NNSA is responsible for the nation's nuclear weapons, nonproliferation, and naval reactors programs. Since its creation, NNSA has continued to experience security problems, such as unauthorized access to a NNSA unclassified computer system, and cost and schedule overruns on its major projects, such as the National Ignition Facility. GAO reviewed the extent to which NNSA has taken steps to (1) improve security at its laboratories and plants and (2) improve its management practices and revise its organizational structure. To carry out its work, GAO reviewed legislation; NNSA policies, plans and budgets; and interviewed current and former NNSA and DOE officials.
Although NNSA has begun to build an effective security organization, it still cannot demonstrate that all of its security program objectives are being met at all of its sites. Specifically, the results of internal and independent security oversight assessments have identified weaknesses in physical security at several NNSA sites, including the Nevada Test Site, Sandia National Laboratories and the Y-12 National Security Complex, and weaknesses in cyber security throughout NNSA. The following factors have contributed to this situation: Weak organization of headquarters security. Until recently, NNSA did not have consistent leadership or direction at the headquarters level for its security program. Security staffing shortages at NNSA site offices. Since NNSA became operational, five of its six site offices have not been staffed at the required levels, according to GAO's analysis. Site offices oversee NNSA contractor security operations. Inadequate security staff training. NNSA has not implemented a training program that provides federal security officials with the skills needed to effectively oversee contractor security programs. Incomplete security data. DOE's database for tracking security deficiencies identified by security oversight assessments is incomplete, and, as a result, NNSA lacks a comprehensive understanding of the overall effectiveness of its security program. NNSA has taken several actions to improve its management practices, including developing a planning, programming, budgeting and evaluation process. However, management problems continue, in part, because NNSA and DOE have not fully agreed on how NNSA should function within the department as a separately organized agency. This lack of agreement has resulted in organizational conflicts that have inhibited effective operations. GAO also identified the following areas where additional management improvements are needed: Project management. NNSA has not developed a project management policy, implemented a plan for improving its project management efforts, and fully shared project management lessons learned between its sites. Program management. NNSA has not identified all of its program managers and trained them to a certified level of competency. Financial management. NNSA has not established an independent analysis unit to review program budget proposals, confirm cost estimates, and analyze budget alternatives.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-07-36, National Nuclear Security Administration: Additional Actions Needed to Improve Management of the Nation's Nuclear Programs
This is the accessible text file for GAO report number GAO-07-36
entitled 'National Nuclear Security Administration: Additional Actions
Needed to Improve Management of the Nation's Nuclear Programs' which
was released on January 31, 2007.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to the Subcommittee on Strategic Forces, Committee on Armed
Services, House of Representatives:
January 2007:
National Nuclear Security Administration:
Additional Actions Needed to Improve Management of the Nation's Nuclear
Programs:
GAO-07-36:
GAO Highlights:
Highlights of GAO-07-36, a report to the Subcommittee on Strategic
Forces, Committee on Armed Services, House of Representatives
Why GAO Did This Study:
In response to security and management weaknesses, in 1999 the Congress
created the National Nuclear Security Administration (NNSA), a
separately organized agency within the Department of Energy (DOE). NNSA
is responsible for the nation‘s nuclear weapons, nonproliferation, and
naval reactors programs. Since its creation, NNSA has continued to
experience security problems, such as unauthorized access to a NNSA
unclassified computer system, and cost and schedule overruns on its
major projects, such as the National Ignition Facility. GAO reviewed
the extent to which NNSA has taken steps to (1) improve security at its
laboratories and plants and (2) improve its management practices and
revise its organizational structure. To carry out its work, GAO
reviewed legislation; NNSA policies, plans and budgets; and interviewed
current and former NNSA and DOE officials.
What GAO Found:
Although NNSA has begun to build an effective security organization, it
still cannot demonstrate that all of its security program objectives
are being met at all of its sites. Specifically, the results of
internal and independent security oversight assessments have identified
weaknesses in physical security at several NNSA sites, including the
Nevada Test Site, Sandia National Laboratories and the Y-12 National
Security Complex, and weaknesses in cyber security throughout NNSA. The
following factors have contributed to this situation:
* Weak organization of headquarters security. Until recently, NNSA did
not have consistent leadership or direction at the headquarters level
for its security program.
* Security staffing shortages at NNSA site offices. Since NNSA became
operational, five of its six site offices have not been staffed at the
required levels, according to GAO‘s analysis. Site offices oversee NNSA
contractor security operations.
* Inadequate security staff training. NNSA has not implemented a
training program that provides federal security officials with the
skills needed to effectively oversee contractor security programs.
* Incomplete security data. DOE‘s database for tracking security
deficiencies identified by security oversight assessments is
incomplete, and, as a result, NNSA lacks a comprehensive understanding
of the overall effectiveness of its security program.
NNSA has taken several actions to improve its management practices,
including developing a planning, programming, budgeting and evaluation
process. However, management problems continue, in part, because NNSA
and DOE have not fully agreed on how NNSA should function within the
department as a separately organized agency. This lack of agreement has
resulted in organizational conflicts that have inhibited effective
operations. GAO also identified the following areas where additional
management improvements are needed:
* Project management. NNSA has not developed a project management
policy, implemented a plan for improving its project management
efforts, and fully shared project management lessons learned between
its sites.
* Program management. NNSA has not identified all of its program
managers and trained them to a certified level of competency.
* Financial management. NNSA has not established an independent
analysis unit to review program budget proposals, confirm cost
estimates, and analyze budget alternatives.
What GAO Recommends:
GAO is recommending that the Secretary of Energy and the Administrator,
NNSA, (1) improve NNSA‘s security program, (2) develop and implement
standard operating procedures for conducting business and resolving
conflicts between DOE and NNSA, and (3) institute a series of
initiatives to improve project, program, and financial management. NNSA
generally agreed with the report‘s findings and corresponding
recommendations.
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-07-36].
To view the full product, including the scope and methodology, click on
the link above. For more information, contact Gene Aloise at (202) 512-
3841 or aloisee@gao.gov.
[End of Section]
Contents:
Letter:
Results in Brief:
Background:
Additional Action Needed to Improve NNSA's Security Program:
Several Management Issues Need to be Resolved for NNSA to Become Fully
Effective:
Conclusions:
Recommendations for Executive Action:
Agency Comments and Our Evaluation:
Appendixes:
Appendix I: Comments from the National Nuclear Security Administration:
Appendix II: GAO Contact and Staff Acknowledgments:
Tables:
Table 1: Safeguards and Security Program Areas Covered in NNSA Surveys:
Table 2: Safeguards and Security Program Areas in Office of Independent
Oversight Inspections:
Table 3: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Overall Security Performance, Fiscal Years
1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Table 4: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Physical Security Performance, Fiscal
Years 1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Table 5: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Cyber Security Performance, Fiscal Years
1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Table 6: NNSA Site Office Security Staffing Shortages, Fiscal Years
2001 through 2005:
Table 7: Staffing Levels within NNSA and DOE Offices:
Table 8: Major Projects Not Included in PARS:
Abbreviations:
DNS: Office of Defense Nuclear Security:
DOD: Department of Defense:
DOE: Department of Energy:
FTE: full time equivalent:
NIST: National Institute of Standards and Technology:
NNSA: National Nuclear Security Administration:
NSSP: Nuclear Safeguards and Security Program:
OMB: Office of Management and Budget:
PARS: Project Assessment and Reporting System:
PPBE: planning, programming, budgeting, and evaluation process:
SSIMS: Safeguards and Security Information Management System:
January 19, 2007:
The Honorable Ellen O. Tauscher:
Chairman:
The Honorable Terry Everett:
Ranking Minority Member:
Subcommittee on Strategic Forces:
Committee on Armed Services:
House of Representatives:
During the late 1990s, the Department of Energy (DOE) experienced
management difficulties with its nuclear weapons programs--
particularly with the lack of clear management authority and
responsibility--that contributed to security problems at the nation's
nuclear weapons laboratories and significant cost overruns on major
projects.[Footnote 1] According to a June 1999 report by the
President's Foreign Intelligence Advisory Board (the Board), DOE's
management of the nuclear weapons laboratories, while representing
"science at its best," also embodied "security at its worst" because of
"organizational disarray, managerial neglect, and —a culture of
arrogance."[Footnote 2]
The Board urged the Congress to create a new organization to oversee
the nuclear weapons complex and to insulate it from DOE. The Board
stressed that the new organization, whether established as an
independent agency or a semi-autonomous entity within DOE, should have
a clear mission, streamlined bureaucracy, and drastically simplified
lines of authority and accountability. Responding to the Board's
recommendations, in 1999, the Congress created the National Nuclear
Security Administration (NNSA) under Title 32 of the National Defense
Authorization Act for Fiscal Year 2000--the NNSA Act--to correct the
problems identified.[Footnote 3]
The NNSA Act established NNSA as a "separately organized agency" within
DOE and made NNSA responsible for the management and security of the
nation's nuclear weapons, nuclear nonproliferation, and naval reactor
programs.[Footnote 4] To implement its programs, NNSA relies on
contractors to manage day-to-day site operations and to adhere to DOE
policies when operating the laboratories, production plants, and other
facilities within the complex. Because many NNSA sites handle special
nuclear material, including nuclear weapons, plutonium, and highly
enriched uranium, effective federal oversight is critical to ensuring
that national security, human health and safety, and the environment
are not harmed.
The NNSA Act also established the position of DOE Under Secretary for
Nuclear Security, who was also designated as the Administrator of NNSA.
The Secretary of Energy and the Deputy Secretary of Energy were allowed
to establish policy for NNSA and to give direction to NNSA through the
Administrator; however, other DOE employees were prohibited from
directing the activities of individual NNSA employees. Furthermore, the
NNSA Act established, within NNSA, deputy administrators for defense,
nonproliferation, and naval reactors programs. Finally, the NNSA Act
required that, among other things, NNSA develop a planning,
programming, and budgeting process in order to ensure that the
administration operated under sound financial management principles.
Since its inception, however, NNSA has continued to experience both
security and overall management problems. For example, with respect to
security, in 2003 we reported that NNSA had not been fully effective in
managing its safeguards and security program.[Footnote 5] Specifically,
we found that NNSA had not fully defined the roles and responsibilities
of officials in its security program and that NNSA had shortfalls in
security staff at the site offices that oversee its contractors. In
addition, two NNSA studies commissioned in July 2003 found ongoing
problems with NNSA's security program, including weaknesses in its
security culture, organization, and staffing and training.[Footnote 6]
Finally, DOE's Office of Inspector General found security problems with
NNSA's contractors, including inadequate control of firearms
inventories and improprieties in the testing of the officers who
protect NNSA's sites. In terms of the management of major projects, the
National Ignition Facility and the Dual Axis Radiographic Hydrodynamic
Test Facility--two major facilities needed to support NNSA's nuclear
weapons programs--experienced major delays and cost overruns because of
problems with project management and are still not complete.[Footnote
7]
To implement the NNSA Act and in response to these problems, NNSA has
initiated a series of improvements, including recently creating a new
NNSA headquarters unit to oversee security across the nuclear weapons
complex; establishing a planning, programming, budgeting, and
evaluation process; and putting in place a policy to improve program
management. In addition, in December 2002, the Administrator
reorganized NNSA. Specifically, the Administrator (1) removed a layer
of management by abolishing NNSA's three operations offices located in
New Mexico, Nevada, and California; (2) removed some functions from a
fourth operations office at Oak Ridge, Tennessee, and established the Y-
12 site office; (3) strengthened the oversight role of its site
offices; (4) consolidated business and technical support functions,
such as procurement and contracting, into a single "service center"
organization located in Albuquerque, New Mexico; and (5) reduced NNSA's
staff by about 17 percent by the end of fiscal year 2004.
In this context, you asked us to evaluate the extent to which NNSA has
taken steps to (1) improve security at its laboratories and plants and
(2) improve its management practices and revise its organizational
structure.
To evaluate the extent to which NNSA has taken steps to improve
security at its laboratories and plants, we reviewed the security
requirements contained in DOE directives. We met with security
officials from NNSA headquarters and the NNSA Service Center in
Albuquerque; and at the Lawrence Livermore National Laboratory, Nevada
Test Site, Sandia National Laboratories, Los Alamos National
Laboratory, Pantex and Y-12 National Security Complex site offices. We
also met with security contractor officials at the Livermore, Nevada,
Sandia, Los Alamos, and Y-12, as well as with officials from DOE's
Office of Health, Safety, and Security (formerly the Office of Security
and Safety Performance Assurance). In addition, we met with former DOE
and NNSA officials, including the Director of NNSA's Security Workforce
Panel, and the Director of the NNSA Security Panel. In reviewing the
extent to which NNSA has taken steps to improve security at its
laboratories and plants, we collected and analyzed performance ratings
contained in security oversight reports, issued by DOE's Office of
Health, Safety and Security and NNSA site offices, from fiscal years
1996 through 2005. We used these performance ratings because there was
wide agreement among NNSA and DOE security officials that these ratings
represented the best available information on the overall performance
of NNSA's safeguards and security program. To assess the reliability of
security oversight report ratings, we (1) reviewed existing information
about the methodology and timeframes used in making performance rating
determinations; (2) interviewed knowledgeable agency officials; and (3)
based on document reviews, ensured that we had received all available
performance information. We determined that the data were sufficiently
reliable for the purposes of this report. We also used reports on
NNSA's security efforts prepared by GAO, the DOE Inspector General, and
outside groups, such as a 2005 report commissioned by NNSA.[Footnote 8]
To address the extent of security staffing shortfalls, we collected and
analyzed staffing data provided by NNSA site offices. We relied
primarily on staffing data from the NNSA site offices, as this data was
the most comprehensive and because these offices are responsible for
processing and managing security staffing requirements. To assess the
reliability of these data, we interviewed NNSA security officials who
were responsible for compiling these data and performed basic
reasonableness checks of the data. We determined that the data were
sufficiently reliable for the purposes of our report. Finally, we
reviewed documents related to NNSA's efforts to improve the
organization and operation of its security oversight program.
To evaluate the extent to which NNSA has taken steps to improve its
management practices and revise its organizational structure, we
reviewed the NNSA Act as well as two House of Representatives reports
in 2000 on the act's implementation.[Footnote 9] Because the
establishment of NNSA as a separately organized agency in DOE was a key
provision of the NNSA Act, we met with officials from NNSA
headquarters; the NNSA Service Center; the NNSA site offices; and DOE
offices where NNSA and DOE need to interact, including DOE's Office of
Intelligence and Counterintelligence (formerly the Office of
Counterintelligence), Chief Financial Officer, Chief Information
Officer, General Counsel, and Human Capital Management. To understand
how NNSA and DOE were intended to interact, we interviewed officials
and reviewed documents, such as DOE's January 2000 implementation plan
for NNSA. We also interviewed officials with the Department of
Commerce's National Oceanic and Atmospheric Administration, the Defense
Advanced Research Projects Agency, the Defense Threat Reduction Agency,
and the Department of Transportation's Federal Aviation Administration
to obtain their views on the reporting relationships that need to be in
place for an entity designated as a "separately organized agency" to
succeed. We contacted the first two agencies cited because they were
identified in the Board's June 1999 report as good models of a
separately organized agency. We contacted the latter two agencies cited
after consultation with staff in your offices. We also interviewed
former NNSA and DOE officials, including the first and second
Administrators and Deputy Secretary of Energy, who helped establish
NNSA, to get their perspective on the difficulties involved in creating
a separately organized agency within a department.
With respect to other management improvements, we met with officials
from NNSA headquarters, the NNSA Service Center, NNSA site offices, and
NNSA contractors; and various DOE offices, including the Office of
Engineering and Construction Management. To develop information on the
status of NNSA's efforts to improve project and program management, we
analyzed DOE's annual performance and accountability reports;
construction project data from NNSA's budget requests for fiscal years
1995 through 2005; and DOE program management performance information.
Finally, we reviewed documents on NNSA's planning, programming,
budgeting, and evaluation process and human capital efforts.
We conducted our work from March 2005 to January 2007 in accordance
with generally accepted government auditing standards, which included
an assessment of data reliability and internal controls.
Results in Brief:
Although NNSA has improved its headquarters oversight of security
across the nuclear weapons complex, NNSA still cannot demonstrate that
all of its security program objectives are being met at all of its
sites, according to our review of NNSA site offices' surveys and DOE's
independent inspections. We examined the results of the surveys and
inspections because NNSA and DOE security officials generally agreed
that the survey and inspection results represented the best available
information on the overall performance of NNSA's security program. The
surveys and inspections identify weaknesses with physical security at
several NNSA sites, including the Nevada Test Site, Sandia National
Laboratories, and the Y-12 National Security Complex, and weaknesses
throughout NNSA in the cyber security area. We identified, through
information review and interviews of various NNSA, DOE, and contractor
officials, the following four factors that have contributed to problems
with NNSA's security program:
* Lack of consistent leadership and direction for its security
activities. For several years, the NNSA headquarters security
organization experienced turnover in the position of Chief of the
Office of Defense Nuclear Security (DNS). Specifically, four
individuals have occupied the position since NNSA's creation, often in
an acting capacity. In addition, these chiefs have reported to
different levels within the organization. The current Chief is a
permanent appointee, reporting directly to the NNSA administrator, and
he has taken a number of steps to develop an effective headquarters
security organization.
* Security personnel staffing shortages at site offices. Having
sufficient staff to oversee the security programs of its contractors
continues to be a problem. For example, since NNSA became operational,
some site offices have experienced staffing shortfalls. As a result,
sites are limited in their ability to effectively oversee contractors'
security activities.
* Lack of adequate training resources and opportunities for site office
security staff. NNSA has not implemented a training program that
provides NNSA federal security officials with the skills needed to
effectively oversee contractor security programs. In addition, NNSA
site offices often do not have all the resources needed to meet
training needs. For example, according to site office officials, the
Los Alamos Site Office did not receive training funds for fiscal year
2006 and the Nevada Site Office received a minimal training budget for
its security staff.
* Lack of data to gauge program effectiveness. NNSA does not have
complete data for tracking security deficiencies identified by security
oversight reviews and, as a result, does not have information regarding
the overall effectiveness of its safeguards and security program. NNSA
officials told us that while they believe security across the weapons
complex has improved, NNSA does not have sufficient data to support
this assertion. In addition, NNSA has not implemented a formal process
for sharing best practices or lessons learned to guide security
improvements. While best practices and lessons learned have been
communicated informally, a formal process could help ensure that
previously identified security deficiencies, such as, for example, the
retrieval of badges from terminated employees at one NNSA site, are
reviewed and corrected as necessary at other NNSA field locations.
NNSA has taken several actions to improve its management practices,
including issuing a program management policy; however, almost 7 years
after its creation, NNSA and DOE have not developed procedures that
govern how NNSA should function as a separately organized agency within
the department. Most notably, DOE and NNSA's counterintelligence
offices, whose missions were outlined in the original NNSA Act,
continually disagreed over the scope and direction of the
counterintelligence program. While DOE wanted to focus primarily on
investigating counterintelligence breaches after they occur, NNSA
sought to prevent intelligence leaks. Recent changes to the NNSA Act
are intended to address these conflicts by placing all
counterintelligence activities under the department.[Footnote 10] We
also identified the following four other management areas where
additional NNSA actions could improve its effectiveness in managing the
nuclear weapons complex:
* Human capital. NNSA has made progress in developing a human capital
strategy. However, DOE and NNSA have not conducted a systematic,
detailed analysis of how many staff NNSA needs in relation to DOE. As a
result, we identified areas where potential staff imbalances have
affected NNSA's ability to operate separately from DOE. For example,
NNSA's Office of General Counsel has 35 attorneys, including the
General Counsel, to provide NNSA legal analysis, while the rest of DOE
has 277 attorneys. According to NNSA's General Counsel, his office
would need 15 to 20 additional attorneys to fully handle NNSA's legal
workload with minimal assistance from DOE. Currently, NNSA relies on
DOE's Office of General Counsel to perform a significant portion of its
legal work.
* Project management. While both DOE and NNSA have initiated efforts to
improve project management, NNSA reported in November 2006 that about
16 percent of NNSA projects were at risk of breaching their cost
baseline, schedule baseline or both. We identified seven areas for
improvement that would foster a stronger culture for effective project
management. For example, while NNSA staff has proposed the development
of a new plan to further improve project management, NNSA management
indicated that it was not aware of this proposal; NNSA has not acted on
this proposal. Furthermore, DOE's Project Assessment and Reporting
System--a Web-based system for keeping DOE senior managers apprised of
the performance of projects costing more than $5 million--does not
include four major NNSA projects, estimated to cost over $100 million
each. Consequently, these projects do not receive the senior management
oversight that can be provided through that system.
* Program management. NNSA program managers are responsible for
completing a set of activities by employing a working knowledge of such
diverse areas as contracting, budgeting, and engineering. Recognizing
the important role of program managers, NNSA has taken several actions,
such as developing a program management policy. However, NNSA has yet
to identify all of its program managers or train them to a certified
level of competency. Indeed, DOE's most recent performance and
accountability report for fiscal year 2006 showed that NNSA fully met
only about 52 percent of its program goals while the rest of DOE
achieved about a 79-percent success rate.
* Financial management. NNSA has made significant progress in
implementing the planning, programming, budgeting, and evaluation
process (PPBE) over the last 4 years, as mandated by the NNSA Act.
However, several areas of improvement still have not been fully
addressed. For example, NNSA has issued policy letters on PPBE, but
some of these letters are still in draft form because, in part, NNSA is
waiting to obtain DOE's views on certain matters. In addition, NNSA's
PPBE mechanism for centralized resource allocation relies on collegial
decision making among senior NNSA managers, with the Administrator
resolving disputes and deciding on the final resource allocation.
However, the Administrator does not have an independent group to review
program proposals, confirm cost estimates, and analyze alternatives.
According to a 2003 DOE Inspector General report, most senior managers
believe that such an analytical group would be of value.[Footnote 11]
While NNSA has taken some action in this direction, it is not clear
when such a group will be established.
To improve the management of NNSA and its ability to oversee the
nuclear weapons complex, we are making a series of recommendations to
the Secretary of Energy and the Administrator, NNSA to, among other
things, (1) improve various aspects of NNSA's security oversight
program; (2) clearly define NNSA's status as a separately organized
agency within DOE; and (3) institute a series of NNSA initiatives to
improve project and program management, and the agency's planning,
programming, budgeting, and evaluation process.
We provided NNSA with a copy of this report for their review and
comment. NNSA generally agreed with the report and its corresponding
recommendations. NNSA noted that it considers the agency to be a
success but acknowledged that there was considerable work yet to be
accomplished. NNSA and DOE also provided technical comments, which we
have incorporated in this report as appropriate.
Background:
NNSA conducts its activities at research and development laboratories,
production plants, and other facilities (collectively referred to as
the nuclear weapons complex). Specifically, NNSA operates three
national laboratories that design nuclear weapons--Lawrence Livermore
National Laboratory, California; Los Alamos National Laboratory, New
Mexico; and the Sandia National Laboratories, New Mexico and
California; and four nuclear weapons production sites--the Pantex
Plant, Texas; the Y-12 National Security Complex, Tennessee; the Kansas
City Plant, Missouri; and parts of the Savannah River Site, South
Carolina; as well as the Nevada Test Site.
To implement its programs, NNSA received about $9.1 billion for fiscal
year 2006, including almost $6.4 billion for its nuclear weapons
activities, about $1.6 billion for its defense nuclear nonproliferation
programs, and about $782 million for the Naval Reactors program. NNSA's
appropriation also included about $766 million to provide security at
its sites. NNSA requested over $9.3 billion for fiscal year 2007,
including $6.4 billion for its nuclear weapons activities, $1.7 billion
for its defense nuclear nonproliferation programs, and $795 million for
the Naval Reactors program. According to NNSA's Future Years Nuclear
Security Program plan, between fiscal years 2007 and 2011, NNSA is
proposing to spend almost $48.5 billion on its nuclear weapons, nuclear
nonproliferation, and naval reactors programs.
For several years before NNSA was established, external studies found
problems with the organization and operation of what is now NNSA's
principal organization--DOE's Office of Defense Programs. These studies
cited continuing problems in the areas of overall management,
organization, priority setting, and maintenance of a viable
infrastructure and workforce. For example, the Institute for Defense
Analysis's March 1997 study, often called the "120-Day Study," cited
ambiguities and overlaps between the roles of headquarters and the
field as a primary source of inefficiencies and conflict.[Footnote 12]
Most influential in the creation of NNSA was the study conducted by a
Special Investigative Panel of the President's Foreign Intelligence
Advisory Board.[Footnote 13] Prepared in response to a series of
security problems, including public access to classified documents at
the Los Alamos National Laboratory, the Board found that DOE was a
dysfunctional bureaucracy incapable of reforming itself and that
reorganization was clearly warranted to resolve security and
counterintelligence problems. As noted earlier, the Board urged the
Congress to create a new organization that, whether established as an
independent agency or a semi-autonomous entity within DOE, should have
a clear mission, streamlined bureaucracy, and drastically simplified
lines of authority and accountability. To correct the problems
identified by the Board and others, in 1999, the Congress created the
NNSA.
For the last several years, we have monitored various high-risk areas
impacting NNSA operations and NNSA actions to implement the NNSA Act.
In January 2001, and again in January 2003, we identified strategic
human capital management as a governmentwide, high-risk area. We found
that the lack of attention to strategic human capital planning had
created a risk to the federal government's ability to perform its
missions economically, efficiently, and effectively.[Footnote 14] In
that context, we have stated that strategic workforce planning is
needed to address two critical needs: (1) aligning an organization's
human capital program with its current and emerging mission and
programmatic goals and (2) developing long-term strategies for
acquiring, developing, and retaining staff to achieve programmatic
goals.[Footnote 15]
In April 2001, we testified that NNSA's efforts to establish a new
organization looked promising.[Footnote 16] However, we highlighted the
need for NNSA to clearly define the roles and responsibilities of
headquarters and field staff and to establish clear lines of authority
between NNSA and its contractors, among other things. In May 2001, NNSA
announced plans to reorganize its headquarters operations. In December
2001, we noted that NNSA had set several important goals for its
overall reorganization efforts, including establishing clear and direct
lines of communication, clarifying the roles and responsibilities of
NNSA's headquarters and field offices, and integrating and balancing
priorities across NNSA's missions and infrastructure.[Footnote 17]
However, we found that NNSA's plans for the headquarters reorganization
did not contain a clear definition of the roles and responsibilities of
the headquarters organizational units.
In addition to reorganizing its headquarters, in February 2002, NNSA
proposed reorganizing its entire operation to solve important, long-
standing issues. Specifically, NNSA proposed a new organizational
structure that would (1) remove a layer of management by converting
existing operations offices to one support office, (2) locate NNSA
operational oversight close to laboratories and plants by strengthening
its site offices, and (3) streamline federal staff and hold federal
staff and contractors more accountable. In February 2002, we testified
that, with the proposed new organizational structure, resolution of
NNSA's long-standing organizational issues appeared to be within NNSA's
grasp.[Footnote 18] However, we noted that NNSA's lack of a long-term
strategic approach to ensuring a well-managed workforce precluded it
from identifying its current and future human capital needs, including
the size of the workforce; its deployment across the organization; and
the knowledge, skills, and capabilities needed to fulfill its mission.
In December 2002, the Administrator of NNSA implemented the proposed
reorganization.
Our May 2003 report on the management of NNSA's security program
identified similar concerns about NNSA's security organization and
management.[Footnote 19] Specifically, we found that NNSA (1) had not
fully defined clear roles and responsibilities for its headquarters and
site security operations and (2) had shortfalls at its site offices in
the total number of staff and in expertise, which could make it more
difficult for the site offices to effectively oversee security
activities. We therefore concluded that NNSA could not be assured that
its contractors were working to maximum advantage to protect critical
facilities and material from individuals seeking to inflict damage.
In June 2004, we found that NNSA's reorganization had addressed some
past problems by better delineating lines of authority and improving
communication. However, we also found that NNSA's reorganization had
not ensured that the agency had sufficient staff with the right skills
in the right places because it had downsized its federal workforce by
about 17 percent without first determining the critical skills and
capabilities needed to meet its mission and program goals.[Footnote 20]
Recently, in response to a July 2005 report by the Secretary of
Energy's Advisory Board's Nuclear Weapons Complex Infrastructure Task
Force on the nuclear weapons complex of the future, NNSA changed the
reporting relationship of the site office managers.[Footnote 21] To
facilitate the transformation of the nuclear weapons complex, the Task
Force recommended that the site office managers begin reporting to the
Deputy Administrator for Defense Programs instead of to the
Administrator. The Task Force believed that the Deputy Administrator
was the primary line manager with mission responsibility for the work
conducted at the sites and that this change would better align
responsibility and management of the program. In May 2006, the
Administrator made this change.
Additional Action Needed to Improve NNSA's Security Program:
Although NNSA has improved its headquarters oversight of security
across the nuclear weapons complex, NNSA continues to experience
difficulties in developing a safeguards and security program that
provides reasonable assurance that all protection objectives are being
met, especially at certain locations and in several topical areas, most
notably cyber security, according to our review of NNSA site office
surveys and independent inspections by DOE's Office of Independent
Oversight. We analyzed the results of the surveys and inspections
because NNSA and DOE security officials generally agreed that the
survey and inspection results represented the best available
information on the overall performance of NNSA's security program.
While the results of the surveys and inspections are not entirely
consistent, they do identify weaknesses with physical security at
several NNSA sites, including the Nevada Test Site, Los Alamos National
Laboratory, and the Y-12 National Security Complex, and weaknesses
throughout NNSA in the cyber security area. Over the last several
years, NNSA's security program has been hampered by several factors,
including (1) the lack of an effective headquarters organization, (2)
security staffing shortages at its site offices, (3) the lack of
adequate training resources and opportunities for its site office
security staff, and (4) the lack of metrics to gauge program
effectiveness.
NNSA Is Experiencing Difficulty in Meeting All Security Program
Objectives:
According to DOE Manual 470.4-1, Safeguards and Security Program
Planning and Management, NNSA must develop a safeguards and security
program that ensures NNSA has the necessary protections in place to
protect its security interests against malevolent acts, such as theft,
diversion, sabotage, modification, compromise, or unauthorized access
to nuclear weapons, nuclear weapons components, special nuclear
materials, or classified and unclassified information. NNSA's DNS is
responsible for developing the administration's security program.
To determine the overall effectiveness of NNSA's safeguards and
security program, NNSA and DOE principally rely on two types of
evaluations. Specifically, NNSA site offices conduct surveys of
contractors' performance while the Office of Independent Oversight
conducts periodic inspections, typically every 18 months. These
evaluations identify the strengths and weaknesses of NNSA's security
program and provide a basis for line management to decide on program
implementation activities, including allocating resources and
correcting security deficiencies.
Site Office Surveys. NNSA's site offices assess the extent to which day-
to-day security activities at its contractor-operated laboratories and
production facilities ensure that program objectives are met. According
to DOE's Safeguards and Security Program Planning and Management
manual, the frequency of these surveys varies from 12 months to 24
months, depending on the site's importance rating and type of materials
stored at that location. The site office surveys cover 8 topical areas
and 33 subtopical areas and are based on observations of performance,
including compliance with DOE and NNSA security directives. Table 1
shows the 8 topical and 33 subtopical safeguards and security areas
covered.
Table 1: Safeguards and Security Program Areas Covered in NNSA Surveys:
Safeguards and security program areas: Program management and support;
* Protection management program;
* Safeguards and security planning and procedures;
* Management control;
* Program-wide support;
Program definition: Establishes a standardized approach for protection
program planning that will provide an information baseline for use in
integrating departmental safeguards and security considerations,
facilitating management evaluation of program elements, determining
resources for needed improvements, and establishing cost-benefit bases
for analyses and comparison.
Safeguards and security program areas: Protective forces;
* Management;
* Training;
* Duties;
* Facilities and equipment;
Program definition: Protect security interests from terrorist or other
adversarial actions that could have major national security
consequences.
Safeguards and security program areas: Physical security;
* Access controls;
* Intrusion detection and assessment systems;
* Barriers and delay mechanisms;
* Testing and maintenance;
* Communications;
Program definition: Protect departmental assets from malevolent acts
such as theft, diversion, and sabotage events, such as civil disorder,
by considering site and regional threats, protection planning
strategies, and protection measures.
Safeguards and security program areas: Information protection;
* Basic requirements;
* Technical surveillance and countermeasures;
* Operations security;
* Classification guidance;
* Classified matter protection and control;
Program definition: Protect and control classified and controlled
sensitive matter that is generated, received, transmitted, used,
stored, reproduced, or destroyed.
Safeguards and security program areas: Cyber security;
* Classified cyber security;
* Telecommunications security;
* Unclassified cyber security;
Program definition: Protect automated information systems against
unauthorized access to or modification of information, whether in
storage, processing, or transit, against loss of accountability for
information and user actions, and against the denial of service to
authorized users, including those measures necessary to protect
against, detect, and counter such threats.
Safeguards and security program areas: Personnel security;
* Access authorizations;
* Human reliability program;
* Control of classified visits;
* Safeguards and security awareness;
Program definition: Ensure that granting an individual access to
classified matter, special nuclear material, or both would not endanger
the common defense and security and would be clearly consistent with
the national interest.
Safeguards and security program areas: Unclassified visits and
assignments by foreign nationals;
* Sponsor program management and administration;
* Counterintelligence requirements;
* Export controls/ technology transfer requirements;
* Security requirements;
* Approvals and reporting;
Program definition: Document and track visits and assignments by
foreign nationals to DOE sites or involving DOE information or
technologies.
Safeguards and security program areas: Control and accountability of
nuclear materials;
* Program administration;
* Material accountability;
* Materials control;
Program definition: Provide information on, control of, and assurance
of the presence of nuclear materials, including those systems necessary
to establish and track nuclear material inventories, control access to
and detect loss or diversion of nuclear material, and ensure the
integrity of those systems and measures.
Source: DOE orders and manuals.
[End of table]
The surveys assign the following color-coded performance ratings to
communicate the extent of performance and compliance with required
procedures and practices, as well as the impact of any deficiencies
along with other mitigating factors:
* Satisfactory (Green). The element being evaluated meets protection
objectives or provides reasonable assurance that protection objectives
are being met.
* Marginal (Yellow). The element being evaluated partially meets
protection objectives or provides questionable assurance that
protection objectives are met.
* Unsatisfactory (Red). The element being evaluated does not meet
protection objectives or does not provide adequate assurance that
protection objectives are being met.
Office of Independent Oversight Inspections. The Office of Independent
Oversight independently evaluates DOE sites, facilities, organizations,
and operations in the areas of safeguards and security, cyber security,
emergency management, and environment, safety, and health programs. The
Office of Independent Oversight inspects both federal and contractor
operations at a site and identifies findings, issues, and opportunities
for improvement. It also performs follow-up reviews to ensure
corrective actions are effective and that weaknesses in the safeguards
and security program are appropriately addressed. As shown in table 2,
the Office of Independent Oversight's inspections cover eight topical
and 36 subtopical safeguards and security areas.
Table 2: Safeguards and Security Program Areas in Office of Independent
Oversight Inspections:
Safeguards and security program areas: Protection program management;
* Planning process;
* Organization and staffing;
* Budget process;
* Program direction;
* Control systems;
* Survey program;
Program definition: Programs are designed to ensure that security
interests are provided an appropriate degree of protection from hostile
acts. It is an iterative process, whereby activities related to
planning, staffing, budget, direction, and feedback are integrated with
overall strategic and near-term operational planning.
Safeguards and security program areas: Physical security systems;
* Intrusion detection assessment;
* Entry and search controls;
* Badges, passes, and credentials;
* Barriers;
* Communications;
* Testing and maintenance;
Program definition: Programs are designed to use intrusion detection
and assessment, entry and search control, barriers, communications,
testing and maintenance, and supporting systems and interfaces to
deter, detect, annunciate, assess, delay, and communicate an
unauthorized activity.
Safeguards and security program areas: Protective forces;
* Management;
* Training;
* Equipment and facilities;
* Duties;
Program definition: Programs are designed to protect both DOE security
interests from theft, sabotage, and other hostile acts that may
adversely impact national security or the health and safety of the
public, as well as life and property at DOE facilities.
Safeguards and security program areas: Classified matter protection and
control;
* Program management;
* Control of secret and confidential documents;
* Control of top secret documents;
* Control of classified materials;
* Special programs;
* Operations security;
* Technical surveillance countermeasures;
* Classification management;
Program definition: Programs are designed to provide protection against
unauthorized disclosure of classified matter and for sensitive
information from its inception to destruction or decontrol.
Safeguards and security program areas: Material control and
accountability;
* Administration;
* Accounting;
* Measurements;
* Inventories;
* Containment;
Program definition: Programs are designed to provide an information and
control system for nuclear material. These programs encompass those
systems and measures necessary to establish and track nuclear material
inventories, control access, detect loss or diversion of nuclear
material, and ensure the integrity of those systems and measures.
Safeguards and security program areas: Personnel security;
* Management;
* Personnel clearance program;
* Security education program;
* Visitor control program;
* Human reliability program;
Program definition: Programs are designed to ensure that access to
sensitive information, classified matter, and special nuclear material
will be granted only after it has been determined that such access will
not endanger security and is consistent with the national interest.
Safeguards and security program areas: Cyber security--classified and
unclassified;
Program definition: Programs are designed to prevent deliberate or
inadvertent unauthorized disclosure, acquisition, manipulation,
modification, or loss of information contained within computer networks
and systems, as well as measures designed to prevent denial of
authorized use of the system.
Source: Office of Independent Oversight inspector's guides.
[End of table]
As shown in table 2, the Office of Independent Oversight examines areas
that are comparable to those that NNSA site offices assess through
their surveys. Similarly, the Office of Independent Oversight's rating
scale is comparable to the one NNSA site offices use. Specifically, the
Office of Independent Oversight may assign the following color-coded
ratings to overall programs, selected topical areas, or both:
* Effective performance (Green). The system inspected provides
reasonable assurance that the identified protection needs are met. That
is, (1) the protection meets all applicable standards, (2) the
protection did not meet all standards, but other systems or
compensatory measures provide equivalent protection, or (3) the failure
to fully meet a standard does not significantly degrade the protection.
Line managers are expected to effectively address any specific
deficiencies identified.
* Needs improvement (Yellow). The system inspected only partially meets
identified protection needs or provides questionable assurance that the
identified protection needs are met. That is, one or more applicable
standards are not met and are only partially compensated for by other
systems, and the resulting deficiencies degrade the effectiveness of
the inspected system. Line managers would be expected to significantly
increase their attention to the identified areas of weakness.
* Significant weakness (Red). The system inspected does not adequately
assure that identified protection needs are met. That is, one or more
applicable standards are not met, no compensating factors reduce the
impact on system effectiveness, and the deficiencies seriously degrade
the system's effectiveness. Line managers are expected to apply
immediate attention, focus, and resources to the deficient program
areas.
In addition to ratings, NNSA site office surveys and Office of
Independent Oversight inspection reports develop findings that identify
validated security deficiencies associated with a contractor's
safeguards and security program. In response to these findings, DOE
requires that contractors develop corrective actions to correct the
security deficiencies. These corrective actions include risk
assessment, root cause and cost-benefit analyses. To facilitate the
tracking of security findings and the development of corrective actions
by the responsible organization, DOE Manual 470.4-1 requires site
office and the Office of Independent Oversight staff to ensure that
findings are entered into the Safeguards and Security Information
Management System (SSIMS) --a centralized information management
system--in a timely manner.
NNSA Site Office Survey and Office of Independent Oversight Inspection
Results, While Not Entirely Consistent, Identify Several Problem Areas:
To evaluate the extent to which NNSA has addressed security issues at
its laboratories and plants, we reviewed and compared the results of
NNSA's site office surveys and DOE's Office of Independent Oversight
inspections for the 5-year period before the administration became
operational--fiscal years 1996 through 2000--and the 5-year period
after it became operational--fiscal years 2001 through 2005. While the
results of these surveys and inspections are not entirely consistent,
they do indicate that NNSA continues to experience difficulties in
implementing a safeguards and security program that provides reasonable
assurance that all protection objectives are being met, especially at
certain locations and in several topical areas, most notably cyber
security.
Initially, we totaled the ratings given to each topical area shown in
table 2 and determined the cumulative percent of ratings that resulted
in either a satisfactory, marginal, or unsatisfactory rating. With
respect to the site office surveys, the cumulative percentage of
topical areas that received a satisfactory rating declined slightly
between the two periods. Specifically, as shown in table 3, 86 percent
of the areas covered by the survey received a satisfactory rating for
the period before NNSA became operational, while 81 percent received a
satisfactory rating after NNSA was established.
Table 3: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Overall Security Performance, Fiscal Years
1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Ratings: Satisfactory;
Fiscal years 1996 through 2000: Site office surveys: 86%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 82%;
Fiscal years 2001 through 2005: Site office surveys: 81%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 52%.
Ratings: Marginal;
Fiscal years 1996 through 2000: Site office surveys: 12%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 16%;
Fiscal years 2001 through 2005: Site office surveys: 17%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 39%.
Ratings: Unsatisfactory;
Fiscal years 1996 through 2000: Site office surveys: 2%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 2%;
Fiscal years 2001 through 2005: Site office surveys: 2%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 9%.
Ratings: Total number of assigned ratings;
Fiscal years 1996 through 2000: Site office surveys: 209;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 50;
Fiscal years 2001 through 2005: Site office surveys: 221;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 74.
Source: GAO analysis of NNSA site office survey and Office of
Independent Oversight inspection reports.
[End of table]
The results for our analysis of the overall Office of Independent
Oversight results showed a different picture. Specifically, while 82
percent of the areas covered by the Office of Independent Oversight
inspections received a satisfactory rating for the period before NNSA
became operational, only 52 percent received a satisfactory rating
after NNSA was established. Most notably, the percentage of areas
receiving a marginal rating increased from 16 percent for the period
1996 through 2000 to 39 percent for 2001 through 2005.
Because of (1) the importance of physical security in response to the
terrorist attacks of September 11, 2001, and (2) ongoing concerns about
the vulnerability of DOE computer systems to cyber security attacks,
and to identify where the differences were occurring between the NNSA
site office survey and the Office of Independent Oversight inspection
results, we analyzed the results for the physical security and cyber
security topical areas in the NNSA site office surveys and the Office
of Independent Oversight inspections for the 5-year period before the
administration became operational--fiscal years 1996 through 2000--and
the 5-year period after it became operational--fiscal years 2001
through 2005.[Footnote 22]
With respect to physical security, as shown in table 4, for the site
office surveys, the cumulative percentage of physical security topical
areas that received a satisfactory rating remained relatively unchanged
between the two periods for the seven sites we reviewed. In each
period, slightly more than 80 percent of the assessments resulted in a
satisfactory rating, even as physical security requirements have been
strengthened in response to the September 11, 2001 attacks.
Table 4: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Physical Security Performance, Fiscal
Years 1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Ratings: Satisfactory;
Fiscal years 1996 through 2000: Site office surveys: 84%;
Fiscal years 1996 through 2000: Office of Independent Oversight
Inspections: 80%;
Fiscal years 2001 through 2005: Site office surveys: 82%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 49%.
Ratings: Marginal;
Fiscal years 1996 through 2000: Site office surveys: 14%;
Fiscal years 1996 through 2000: Office of Independent Oversight
Inspections: 17%;
Fiscal years 2001 through 2005: Site office surveys: 15%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 41%.
Ratings: Unsatisfactory;
Fiscal years 1996 through 2000: Site office surveys: 2%;
Fiscal years 1996 through 2000: Office of Independent Oversight
Inspections: 3%;
Fiscal years 2001 through 2005: Site office surveys: 3%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 10%.
Ratings: Total number of assigned ratings;
Fiscal years 1996 through 2000: Site office surveys: 163;
Fiscal years 1996 through 2000: Office of Independent Oversight
Inspections: 36;
Fiscal years 2001 through 2005: Site office surveys: 168;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 59.
Source: GAO analysis of NNSA site office survey and Office of
Independent Oversight inspection reports.
[End of table]
While the site surveys showed relatively little change between the
periods reviewed, the Office of Independent Oversight's evaluation of
physical security showed a pronounced decline. Specifically, as shown
in table 4, while about 80 percent of the Office of Independent
Oversight's physical security assessments resulted in a satisfactory
rating for the period between fiscal years 1996 and 2000, only about 49
percent of the Office of Independent Oversight's physical security
ratings were satisfactory for the period of fiscal year 2001 through
2005.
Several factors contributed to the differences between the NNSA site
office ratings and the Office of Independent Oversight inspections for
physical security. First, while more site offices surveys were
performed in fiscal years 2001 through 2005 than the Office of
Independent Oversight inspections, proportionately more Office of
Independent Oversight ratings were marginal and unsatisfactory. In
particular, we found that at three sites--the Nevada Test Site; Sandia
National Laboratories, New Mexico; and the Y-12 National Security
Complex--over half of the Office of Independent Oversight ratings were
either marginal or unsatisfactory. In contrast, only the Y-12 National
Security Complex's site office surveys had more than 50 percent of the
ratings as marginal or unsatisfactory over the same period. Moreover,
the Office of Independent Oversight inspections had a higher percentage
of marginal and unsatisfactory ratings in the Physical Security
Systems, Classified Matter Protection and Control, and Materials
Control and Accountability topical areas than did the site office
surveys.
With respect to cyber security, both the site office surveys and the
Office of Independent Oversight inspections show declines.
Specifically, as shown in table 5, 92 percent of the cyber security
assessments in NNSA site office surveys resulted in a satisfactory
rating for both classified and unclassified cyber security before NNSA
was operational, but only 77 percent of the assessments resulted in a
satisfactory rating after NNSA became operational.
Table 5: Site Office Survey and Office of Independent Oversight
Inspection Results of NNSA's Cyber Security Performance, Fiscal Years
1996 through 2000, Compared with Fiscal Years 2001 through 2005:
Ratings: Satisfactory;
Fiscal years 1996 through 2000: Site office surveys: 92%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 86%;
Fiscal years 2001 through 2005: Site office surveys: 77%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 60%.
Ratings: Marginal;
Fiscal years 1996 through 2000: Site office surveys: 4%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 14%;
Fiscal years 2001 through 2005: Site office surveys: 23%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 33%.
Ratings: Unsatisfactory;
Fiscal years 1996 through 2000: Site office surveys: 4%;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 0%;
Fiscal years 2001 through 2005: Site office surveys: 0%;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 7%.
Ratings: Total number of assigned ratings;
Fiscal years 1996 through 2000: Site office surveys: 46;
Fiscal years 1996 through 2000: Office of Independent Oversight
inspections: 14;
Fiscal years 2001 through 2005: Site office surveys: 53;
Fiscal years 2001 through 2005: Office of Independent Oversight
inspections: 15.
Source: GAO analysis of NNSA site office survey and Office of
Independent Oversight inspection reports.
[End of table]
Similarly, Office of Independent Oversight inspection results also
indicated that cyber security performance had declined. Before NNSA
became operational, 86 percent of cyber security assessments resulted
in satisfactory ratings, compared with only 60 percent after NNSA
became operational.
With respect to the individual sites, the Office of Independent
Oversight inspections gave the most marginal or unsatisfactory ratings
for cyber security to four sites: Los Alamos National Laboratory;
Pantex; Sandia National Laboratories, New Mexico; and the Y-12 National
Security Complex. In contrast, the NNSA site office surveys assigned
marginal ratings to every NNSA site.
Concerns about cyber security were also raised in one of the two
independent analyses NNSA commissioned in July 2003 and by NNSA
officials with whom we spoke. Specifically, one analysis, NNSA
Security: An Independent Review, found that NNSA cyber security
policies, procedures and practices were less mature than their
counterparts in other security disciplines. Furthermore, the review
found that NNSA was "resource poor" and cyber security implementation
varied widely throughout NNSA because of inadequate funding,
insufficient cyber security personnel and expertise, and inadequate
collaboration between DOE and NNSA cyber security organizations.
According to current and former NNSA officials, and several site office
security directors, the cyber security program has received inadequate
attention and was poorly implemented because NNSA lacks a risk-based
approach to physical and cyber security. In this regard, in June 2006,
DOE officials revealed that a hacker had attacked an unclassified
computer system at NNSA's Service Center in Albuquerque, New Mexico,
and had gained access to a file containing the names and social
security numbers of over 1,500 NNSA employees.
More recently, in September 2006, DOE's Inspector General reported on
the weaknesses associated with the department's cyber security
programs, including NNSA's. According to the Inspector General, while
DOE has made progress in addressing some cyber-related problems,
ongoing problems continue to place DOE's critical information systems
and data at risk of compromise. The report cited weaknesses in the
following areas: systems inventory, certification, and accreditation;
contingency planning; access controls; and configuration management and
change controls. These problems occurred, in part, because program and
field offices did not always implement departmental and federal cyber
security requirements. For example, the Inspector General reported,
NNSA site officials stated that they were required to comply with NNSA
cyber security policy, rather than standards established by the
National Institute of Standards and Technology (NIST). However, the DOE
Inspector General also found that NNSA sites had not fully implemented
NNSA's cyber security policy and that many NNSA field sites were
permitted to follow a less thorough certification and accreditation
process that did not include all NIST or NNSA requirements.
Some of the differences between the site office survey and the Office
of Independent Oversight inspection results may be due to continued
differences in how each organization conducts its evaluations. We and
other oversight organizations have reviewed this issue over the last 6
years. For example, in February 2000, we reported that from 1994
through 1999, the Los Alamos National Laboratory and the Lawrence
Livermore National Laboratory received inconsistent oversight
ratings.[Footnote 23] During a given year, Los Alamos National
Laboratory received ratings ranging from marginal to excellent,
depending on the DOE organization conducting the assessment. Similarly,
the Lawrence Livermore National Laboratory received ratings ranging
from marginal to far exceeds expectations. We reported that this
inconsistency resulted, in part, from various organizations' use of
different criteria as well as the timing of the rating. Similarly, in
May 2000, DOE's Office of Inspector General reported on various factors
affecting the safeguards and security ratings and found that certain
security ratings were changed without a documented rationale.[Footnote
24]
More recently, in 2005, the Office of Independent Oversight issued a
report on the inconsistencies in safeguards and security
ratings.[Footnote 25] Specifically, the report noted that the Office of
Independent Oversight tended to identify significant safeguards and
security program deficiencies that the site office surveys did not
detect. Furthermore, the report identified several major weaknesses
commonly found in survey programs, including a lack of
comprehensiveness in conducting surveys; insufficient, or
insufficiently rigorous, performance testing; a reluctance to assign
formal findings when deficiencies were identified; and the lack of
sufficient detail to justify, or even to explain, the assigned ratings
in some survey reports.
Several Factors Have Contributed to Problems in NNSA's Safeguards and
Security Program:
Because many NNSA sites handle special nuclear material, including
nuclear weapons, plutonium, and highly enriched uranium, effective
federal oversight is critical to ensuring that the national security is
not harmed. While NNSA has recently improved its headquarters oversight
of security, it has been hampered over the last 7 years by several
factors, including (1) the lack of an effective headquarters
organization, (2) security staffing shortages at its site offices, (3)
lack of adequate training resources and opportunities for its site
office security staff, and (4) the lack of a complete security
information database and framework for evaluating security review
results.
NNSA Has Had Difficulty Developing an Effective Headquarters Security
Organization:
Under GAO's Standards for Internal Control in the Federal Government,
federal agencies must establish and maintain an effective system of
internal controls over their operations.[Footnote 26] Such a system is
the first line of defense in safeguarding assets and preventing and
detecting errors. Under our standards, managers should, among other
things, ensure that (1) the organizational structure clearly defines
key areas of authority and responsibility, (2) staff have the required
skills to meet organizational objectives, (3) progress can be
effectively measured, and (4) operations can be effectively monitored.
However, during much of its history, NNSA's security program has
experienced the following organizational issues:
* Initial organizational arrangements were not productive. The NNSA Act
established the position of Chief of DNS to advise the NNSA
Administrator on security-related matters and implement the security
policies directed by the Secretary and the Administrator. NNSA also
established the Nuclear Safeguards and Security Program (NSSP) within
NNSA's Facilities and Operations organization to provide assistance to
NNSA's site offices on such issues as implementing the Design Basis
Threat and developing budget requests for security.[Footnote 27]
However, the heads of the two offices had strained relations and often
disagreed on how to improve physical security at NNSA facilities. In
September 2002, the NSSP Director was replaced and his position
eventually terminated, and other NSSP staff either retired or left
NNSA. As a result, the effectiveness of the NSSP program was greatly
reduced and the field lost a source of guidance and leadership in NNSA
headquarters.
* Leadership turned over several times. NNSA's headquarters security
organization experienced considerable turnover in the position of Chief
of DNS and changes in the reporting level of that position within NNSA.
Specifically, four individuals have occupied that position since NNSA's
creation, often in an acting capacity. In addition, these Chiefs have
reported to different levels within NNSA. For example, the original
Chief reported directly to the Administrator of NNSA. However, when he
retired in 2003, his replacement occupied two positions--Chief of DNS
and Associate Administrator for Facilities and Operations--a practice
often referred to as "dual-hatting." When NNSA reorganized its
headquarters security organization in 2004, the Chief of DNS once again
reported directly to the Administrator. The second Administrator, who
resigned in January 2007, told us that he regretted having dual-hatted
the Chief of DNS position because it lowered the position's visibility
within NNSA. He believed that it was correct to elevate the current
Chief to serve as an Associate Administrator reporting directly to the
Administrator.[Footnote 28]
* Without effective leadership, important oversight activities were
ignored. The third Chief of DNS employed only a small staff and took
the position that virtually all headquarters oversight responsibility
rested with the Office of Independent Oversight. This approach was
strongly criticized in the April 2005 NNSA Security, An Independent
Review, which noted that the Office of Independent Oversight had
little, if any, direct responsibility, authority, and accountability
for day-to-day supervision of the NNSA complex and that the Office of
Independent Oversight's principal role was one of periodic independent
performance assessment.
To his credit, the current Associate Administrator for DNS has taken
numerous steps to build an effective headquarters security
organization. Specifically, he has:
* issued a program management plan to establish an integrated, NNSA-
wide, performance-based management approach for implementing work and
defining security performance expectations among DNS, NNSA site
offices, and their contractors to ensure full integration of security
requirements with the programmatic missions of NNSA;
* issued specific requirements for NNSA's Performance Assurance
Program, which provides a multi-tiered system of self-assessments and
other reviews of security performance and is aimed at assuring
comprehensive assessments of safeguards and security interests;
* improved interaction between NNSA headquarters and site office
security elements. For example, the Associate Administrator established
the Defense Nuclear Security Leadership Coalition, which comprises all
safeguards and security directors and meets regularly to address
security implementation challenges. He also conducts monthly
teleconferences with both federal and contractor security officials at
NNSA sites;
* established the Office of Security Oversight to assess safeguards and
security performance, serve as the point of contact for correcting
security deficiencies identified during surveys and independent
inspections, ensure the quality and consistency of federal oversight of
contractor activities, and identify areas for improvement throughout
the weapons complex; and:
* adopted a more proactive approach to security through stronger
accountability. For example, within the last 2 years, NNSA has replaced
three federal security directors at the Los Alamos, Nevada and Y-12
site offices.
NNSA Continues to Experience Security Staffing Shortfalls:
Having sufficient staff to oversee the security programs of its
contractors has been a persistent problem since NNSA's creation. For
example, in our May 2003 report on the management of NNSA's security
program, we found, among other things, that NNSA had shortfalls at its
site offices in the total number of security staff and in security
expertise, which may have made it more difficult for the site offices
to effectively oversee security activities. Given this situation, we
concluded that NNSA could not be assured that its contractors were
working to maximum advantage to protect critical facilities and
material from individuals seeking to inflict damage.[Footnote 29] We
raised a similar concern in our June 2004 report on NNSA's
reorganization. Specifically, we found that NNSA's reorganization had
not ensured that the agency had sufficient staff with the right skills
in the right places because it had downsized its federal workforce
without first determining the critical skills and capabilities needed
to meet its mission and program goals. For example, we noted that some
site offices, namely Pantex, Y-12, and Los Alamos, were having
difficulty filling critical skills in the safety and security
disciplines.[Footnote 30]
Complementing our work was an independent analysis commissioned by
NNSA. Specifically, in a March 2004 report, the "Chiles Commission"
identified a number of challenges that NNSA faces in recruiting,
training, and retaining enough skilled federal staff members to fill
the current and future security positions needed to effectively oversee
the security of its weapons complex.[Footnote 31] These challenges
included identifying the (1) number and skills mix of the security
staff, (2) near-term retirement eligibility of a significant percentage
of the security workforce, (3) lack of professional development and
training programs, and (4) lack of adequate resources. The commission
found that federal security personnel were managed on a site-specific
basis, noting that NNSA lacked a comprehensive human capital plan to
identify the areas of needed experience, ensure timely hiring, and
manage the skills mix of the federal security workforce across the
complex.
As part of this review, we assessed the continuing extent of security
staffing shortfalls at NNSA site offices. Specifically, we requested
information on the number of full-time equivalent (FTE) security
positions required to conduct effective oversight versus the actual
number of FTEs on board for fiscal years 2001 through 2005. As table 6
shows, three of the six site offices--the Los Alamos, Livermore and
Sandia Site Offices--did not report information on the required level
of staffing needed to effectively oversee contractors' security
performance for at least 3 of the 5 years for which we requested data.
According to site office security officials at these locations, they
could not find the records showing official staffing shortfalls.
Table 6: NNSA Site Office Security Staffing Shortages, Fiscal Years
2001 through 2005:
Site office: Los Alamos;
Fiscal year 2001: Required FTEs: [A];
Fiscal year 2001: Actual FTEs: 8;
Fiscal year 2002: Required FTEs: [A];
Fiscal year 2002: Actual FTEs: 4;
Fiscal year 2003: Required FTEs: [A];
Fiscal year 2003: Actual FTEs: 6;
Fiscal year 2004: Required FTEs: [A];
Fiscal year 2004: Actual FTEs: 8;
Fiscal year 2005: Required FTEs: 18;
Fiscal year 2005: Actual FTEs: 9.
Site office: Livermore;
Fiscal year 2001: Required FTEs: [A];
Fiscal year 2001: Actual FTEs: 27;
Fiscal year 2002: Required FTEs: [A];
Fiscal year 2002: Actual FTEs: 27;
Fiscal year 2003: Required FTEs: [A];
Fiscal year 2003: Actual FTEs: 26;
Fiscal year 2004: Required FTEs: 20;
Fiscal year 2004: Actual FTEs: 18;
Fiscal year 2005: Required FTEs: 20;
Fiscal year 2005: Actual FTEs: 19.
Site office: Nevada;
Fiscal year 2001: Required FTEs: 25;
Fiscal year 2001: Actual FTEs: 23;
Fiscal year 2002: Required FTEs: 25;
Fiscal year 2002: Actual FTEs: 20;
Fiscal year 2003: Required FTEs: 23;
Fiscal year 2003: Actual FTEs: 15;
Fiscal year 2004: Required FTEs: 20;
Fiscal year 2004: Actual FTEs: 13;
Fiscal year 2005: Required FTEs: 21;
Fiscal year 2005: Actual FTEs: 14.
Site office: Pantex;
Fiscal year 2001: Required FTEs: 8;
Fiscal year 2001: Actual FTEs: 7;
Fiscal year 2002: Required FTEs: 8;
Fiscal year 2002: Actual FTEs: 7;
Fiscal year 2003: Required FTEs: 8;
Fiscal year 2003: Actual FTEs: 7;
Fiscal year 2004: Required FTEs: 8;
Fiscal year 2004: Actual FTEs: 7;
Fiscal year 2005: Required FTEs: 8;
Fiscal year 2005: Actual FTEs: 8.
Site office: Sandia;
Fiscal year 2001: Required FTEs: [A];
Fiscal year 2001: Actual FTEs: 7;
Fiscal year 2002: Required FTEs: [A];
Fiscal year 2002: Actual FTEs: 7;
Fiscal year 2003: Required FTEs: [A];
Fiscal year 2003: Actual FTEs: 9;
Fiscal year 2004: Required FTEs: [A];
Fiscal year 2004: Actual FTEs: 10;
Fiscal year 2005: Required FTEs: [A];
Fiscal year 2005: Actual FTEs: 11.
Site office: Y-12;
Fiscal year 2001: Required FTEs: 11;
Fiscal year 2001: Actual FTEs: 6;
Fiscal year 2002: Required FTEs: 11;
Fiscal year 2002: Actual FTEs: 10;
Fiscal year 2003: Required FTEs: 10;
Fiscal year 2003: Actual FTEs: 12;
Fiscal year 2004: Required FTEs: 11.5;
Fiscal year 2004: Actual FTEs: 11;
Fiscal year 2005: Required FTEs: 14;
Fiscal year 2005: Actual FTEs: 13.
Source: Analysis of NNSA site office staffing data.
[A] Required FTE numbers were unknown at these sites.
[End of table]
According to NNSA's data, five of the six site offices have experienced
security staffing shortfalls during at least one fiscal year. For
example, since fiscal year 2001, the Nevada Site Office has been
increasingly understaffed, and in fiscal year 2005, the Los Alamos Site
Office had half of the required staff it needed to conduct oversight.
Consequently, according to site office security officials, most site
offices have only one staff member responsible for a safeguards and
security topical area and do not have additional staff on board to fill
in if necessary. Site offices have compensated in various ways. For
example, the Los Alamos Site Office has recently obtained detailees
from other site office elements to assist in conducting oversight.
According to one Safeguards and Security Director, security personnel
have to work excessive overtime to complete assigned tasks. Finally,
because it does not have a dedicated survey team, the Nevada Site
Office has had to request survey assistance from other site offices to
conduct surveys. According to a former NNSA official involved in the
administration's initial organization, NNSA did not tailor security
staffing numbers to actual needs and could have better planned its
allocation of staff to site offices during the reorganization.
To address the lack of security staff, NNSA has taken two actions.
First, it has implemented the Defense Nuclear Security Intern Program.
This program, which was established in 2004 and is administered by the
Pacific Northwest National Laboratory, is designed to attract and train
individuals to fill critical NNSA security positions and support the
development of a qualified, experienced pool of security professionals
to supplement site office staffing. Participants are assigned to a
specific NNSA site for 12 to 15 months. Successful interns may have an
opportunity for permanent placement as federal employees within NNSA's
safeguards and security program if they successfully complete the
program. To date, four participants have done so, with three of the
four converted to permanent positions at the Pantex, Y-12, and Nevada
site offices. According to site office officials, the most significant
challenge associated with successfully implementing the intern program
is obtaining clearances in a timely manner.
Second, in April 2005, NNSA held a staffing summit to determine site
offices' needs. Headquarters and site offices provided their end-of-
year staffing levels, the number of additional FTE staff needed to meet
critical needs based on available resources, and an estimate of the
optimal FTE levels. Based on available resources, NNSA site offices
were granted authority for six FTE security staff. However, according
to site office officials, site offices remain understaffed for
security.
Training for Security Staff Can be Improved:
NNSA has not implemented an appropriate training program to meet the
needs of its security personnel. As a result, the number of security
staff with adequate training and the right skills continues to be a
problem. According to our Standards for Internal Control in the Federal
Government, managers should ensure that their personnel possess and
maintain a level of competence needed to accomplish their assigned
duties. Specifically, skill needs must be continually assessed, and
training should be aimed at developing and retaining the skills needed
to achieve changing organizational goals. In addition, DOE requires
that critical training needs for federal staff are identified and met
and that a systematic mechanism for providing training be established.
Training needs are both organizationally and individually driven. Site
offices are required to identify performance gaps for newly hired
employees and specialized training for subsequent years and are
responsible for ensuring that training requirements are met. For their
part, employees are required to develop an individual development plan
and discuss their training needs with their supervisors, who are
responsible for authorizing all training.
In March 2004, the Chiles Commission reported that NNSA's federal
security workforce was forgoing needed training because of the press of
urgent business and a shortage of resources. The Commission found that
over 50 percent of NNSA personnel surveyed believed that the existing
training and education of the federal security workforce was not
adequate. In addition, the Commission reported that staffing plans did
not have the built-in flexibility needed to allow for professional
development and training. NNSA officials told us that problems
associated with site office workloads resulted from the 2002
reorganization where all oversight responsibilities formerly at the
operations offices were transferred to the newly formed site offices
without sufficient allocation of staff. In addition, site office
security directors told us training budgets do not address the needs of
their staff. For example, according to NNSA officials, the Los Alamos
Site Office did not receive training dollars for fiscal year 2006, and
the Nevada Site Office received only a minimal training budget for its
security staff. According to a human resource official at the NNSA
Service Center, there is an "enormous disconnect" between training
needs and budgets, noting that training needs are not addressed
systematically. To address this issue, Safeguards and Security
directors at each site office have recently developed site-specific
technical qualifications programs for their staffs. While these efforts
are important, we found no NNSA-wide formal training program that
provides NNSA federal security officials with the skills needed to
conduct effective oversight.
NNSA Lacks Security Metrics to Gauge the Effectiveness of Its
Safeguards and Security Program and Has Not Effectively Communicated
Lessons Learned to Guide Security Improvements:
NNSA does not have comprehensive metrics regarding the overall
effectiveness of its safeguards and security program and is not
communicating lessons-learned to guide security improvements across the
complex. According to our Standards for Internal Control, for an agency
to run and control its operations, it must have relevant, accurate
information that is recorded and communicated to management in a form
and within a time frame that enables managers to carry out their
operational responsibilities. According to NNSA officials, NNSA has not
developed a framework for evaluating the results of independent and
internal security reviews. NNSA officials told us that while they
believe that security across the weapons complex has improved, NNSA
does not have good metrics to support the assertion that security has
improved. DNS is currently working to develop performance metrics that
will allow NNSA to track processes and make decisions about its
safeguards and security program, but NNSA does not plan to develop an
overall indicator of its security status to assist in efforts to gauge
program effectiveness.
In addition, NNSA does not have complete data for tracking security
deficiencies identified by security oversight reviews. According to
NNSA security officials, not all security deficiencies have been
entered into SSIMS because (1) the database is very difficult to use,
(2) connections to it are slow because of the lack of high-speed
encrypted modems, and (3) it often crashes once a connection is
established. Therefore, some site office officials questioned SSIMS's
usefulness as a management tool and suggested that it be replaced. As a
result, most site offices have begun to use their own tracking systems,
or those developed by contractors, to track identified security
deficiencies. For example, the Los Alamos National Laboratory uses
REDBOOK, a custom-designed classified system operating on an ACCESS
platform to internally track the status of security deficiencies and
associated corrective action plans. Previously, Los Alamos had read-
only access to SSIMS, but lost its connection after the 2004 stand-down
of operations because of missing computer disks. Similarly, during the
summer of 2004, the Y-12 Site Office developed PEGASUS--an interactive
tracking system that is tailored to the site office's oversight efforts
and links all oversight-related correspondence. According to Y-12 site
office officials, the system informs site office security staff
electronically about their daily oversight tasks, reduces the amount of
paperwork associated with resolving security deficiencies, and allows
the security director to track whether federal security personnel are
performing their oversight duties. In addition, PEGASUS, which is
updated weekly, has the ability to automatically identify recurring
security deficiencies and provides better quality data than SSIMS.
According to one security contractor official, the difficulty
associated with logging onto SSIMS results in security deficiencies
being reported 6 to 12 months after they are resolved.
In commenting on our report, officials in the Office of Health, Safety,
and Security indicated that NNSA sites offices have not reported to the
SSIMS helpdesk any logon problems that would result in a 6-to 12-month
reporting delay. In addition, the officials said that (1) a
significantly high percentage of the officials who have taken SSIMS
training have indicated that SSIMS is one of the best, if not the best,
computer database experiences they have had, (2) since July 2004, SSIMS
has offered high speed encrypted connectivity to all NNSA sites, and
(3) NNSA sites would not lose their SSIMS connection if the sites
installed and used this high speed encrypted connectivity. The Director
of the Office of Security Technology and Assistance in the Office of
Health, Safety, and Security agreed that the SSIMS database was not
complete and therefore not useful for providing oversight; however, he
felt the problem was with how the NNSA site offices were using the
database.
Finally, NNSA has not implemented a formal process for sharing best
practices or lessons learned to guide security improvements across the
weapons complex. According to site office security officials, lessons
learned have been informally shared in some instances, including
monthly meetings with NNSA safeguards and security directors,
teleconferences with contractor security officials, and the review of
results from the Office of Independent Oversight inspections conducted
at other sites. According to the Associate Administrator, he has taken
the initiative to informally communicate lessons learned throughout the
complex, but these lessons have yet to be fully implemented. For
example, as early as February 2005, the DOE Inspector General
identified problems with the retrieval of badges from terminated
employees at the Los Alamos National Laboratory.[Footnote 32] After the
report was issued, the Associate Administrator sent the findings to
NNSA site offices and requested that they review their respective
processes. Nevertheless, the DOE Inspector General identified similar
problems at other NNSA sites in subsequent reviews of their respective
badge retrieval processes.[Footnote 33]
While best practices and lessons learned have been communicated
informally, it is important that NNSA sites have a formal means for
communicating these lessons. Absent updated information in SSIMS, or a
sufficient connection to it, it is difficult to determine whether
proven corrective actions for security deficiencies identified have
been implemented across the complex. Utilizing the existing SSIMS
capability to identify how other sites address security deficiencies
would significantly help alleviate redundancy of effort in addressing
similar problems identified throughout the weapons complex.
Several Management Issues Need to be Resolved for NNSA to Become Fully
Effective:
NNSA and DOE have not yet addressed all of the problems that need
attention to improve the management of the nation's nuclear programs.
Problems continue, in part, because NNSA and DOE have not yet fully
agreed on how NNSA should function within the department. In addition
to addressing this underlying issue, NNSA could strengthen its ability
to manage the nuclear weapons complex if it took additional action in
four other areas: (1) human capital, (2) project management, (3)
program management, and (4) financial management.
DOE and NNSA Have Not Yet Fully Determined How NNSA Should Operate as a
Separately Organized Agency within DOE:
NNSA has focused considerable attention on reorganizing its internal
operations, but it and DOE continue to struggle with establishing how
NNSA should operate as a separately organized agency within the
department. Several factors have contributed to this situation. First,
DOE and NNSA did not have a useful model to follow for establishing a
separately organized agency in DOE. Second, the January 2000
implementation plan, required by the NNSA Act, did not define how NNSA
would operate as a separately organized agency within DOE. As a result,
although some NNSA programs have set up procedures for interacting with
DOE, other programs have not, resulting in organizational conflict.
Even where formal procedures have been developed, interpersonal
disagreements have hindered effective cooperation. In this environment,
concerns about NNSA's organizational status persist, with a January
2006 report by the Defense Science Board calling for the creation of a
new, independent National Nuclear Weapons Agency. However, former
senior DOE and NNSA officials with whom we spoke generally did not
favor removing NNSA from DOE.
Other Federal Agencies Did Not Provide a Useful Model for How a
Separately Organized Agency Should Be Structured:
In its June 1999 report recommending the establishment of NNSA, the
President's Foreign Intelligence Advisory Board suggested that the NNSA
could be constructed in one of two ways: (1) by remaining an element of
DOE, but becoming semi-autonomous, that is, separately organized within
the department, or (2) by becoming completely independent, with its
administrator reporting directly to the President. The Board cited
several models for either course of action. If the first approach were
chosen, the Board identified as potential models the National Security
Agency, the Defense Advanced Research Projects Agency within the
Department of Defense (DOD), and the National Oceanic and Atmospheric
Administration within the Department of Commerce. If an independent
agency approach were chosen, the Board cited the National Aeronautics
and Space Administration and the National Science Foundation as
potential models.
To determine if the agencies the Board cited could have provided useful
models for how NNSA should be established as a separately organized
agency within DOE, we met with officials from the Defense Advanced
Research Projects Agency and the National Oceanic and Atmospheric
Administration, as well as with officials from the Defense Threat
Reduction Agency in DOD and the Federal Aviation Administration in the
Department of Transportation. None of the officials associated with
these four agencies considered their agency to be separately organized
or believed that their agency's operational methods were transferable
to NNSA. Officials of the respective agencies told us the following:
* Defense Advanced Research Projects Agency: The organization is not a
separately organized agency but simply a part of DOD. The agency's
internal operational staff is kept to a minimum, with considerable
reliance on the department for many activities. For example, the
Defense Finance and Accounting Service provides support for the
agency's financial operations.
* National Oceanic and Atmospheric Administration. The Administration
is considered a bureau within the Department of Commerce with no higher
status than any other departmental bureau. In operational areas such as
budget or personnel, both the Administration and the Department of
Commerce have comparable offices, but it is understood that the
Administration office is subordinate, with the Department of Commerce
office having final decision-making authority.
* Defense Threat Reduction Agency. The agency is part of DOD. For all
operational activities, guidance comes from DOD headquarters, and the
agency is expected to follow the normal military chain of command.
* Federal Aviation Administration. The Administration is an operating
administration under the Department of Transportation. It has no higher
status than any other departmental administration, such as the Federal
Highway Administration or the Federal Railroad Administration. In
operational areas, such as the chief information officer, chief
financial officer, and general counsel, the Administration receives
direction from the cognizant departmental office. However, the Federal
Aviation Administration does have separate authority for human
resources and procurement.
Former and current senior DOE and NNSA officials also told us they had
difficulty identifying any federal role model for a separately
organized agency. Absent such a role model, these officials said, NNSA
was created using their best judgment about how a separately organized
agency should be established.
DOE and NNSA Have Not Systematically Developed Standard Operating
Procedures to Define Their Relationship:
GAO's standards require federal agencies to establish and maintain an
effective system of internal controls over agency operations.[Footnote
34] We have noted that internal controls are particularly important for
managing risk during organizational change, and we have emphasized one
control--an agency's organizational structure--that we believe should
clearly define key areas of authority and responsibility.
Section 3297 of the NNSA Act required DOE to submit a plan implementing
the act's provisions to the relevant congressional committees.[Footnote
35] However, DOE's January 1, 2000, implementation plan did little to
define NNSA's status as a separately organized agency. Reflecting the
opposition of the then DOE senior leadership to the creation of NNSA,
the implementation plan dual-hatted virtually every significant
statutory position in NNSA with DOE officials, including the Director
of NNSA's Office of Defense Nuclear Counterintelligence and General
Counsel. As we testified in April 2001, this practice caused
considerable concern about NNSA's ability to function with the
independence envisioned in the NNSA Act.[Footnote 36] In its October
2000 report on the establishment of NNSA, the Special Oversight Panel
on Department of Energy Reorganization noted that the Secretary, facing
criticism for having impeded NNSA's autonomy, retreated from the dual-
hatting philosophy and appointed separate individuals to NNSA offices.
Dual-hatting was subsequently forbidden by an amendment to the NNSA
Act.[Footnote 37]
In a July 2000 hearing before the Special Oversight Panel, the incoming
NNSA Administrator testified that establishing NNSA as a semi-
autonomous entity would require developing a single, interconnected set
of policies and principles that permeated the entire NNSA and minimized
outside inputs. Nevertheless, DOE and NNSA have not systematically
developed standard operating procedures to define their relationship.
Consequently, although some NNSA programs have set up specific policies
and procedures for how they will interact with their DOE counterparts,
other programs have not. Without formal procedures, several functions-
-budgeting, procurement, information technology, management and
administration, and safeguards and security--have not been effectively
or efficiently implemented, as the following points detail.[Footnote
38]
* Budgeting. DOE and NNSA do not have a formal order defining how DOE's
and NNSA's budgeting requirements would be combined into a practicable
process. As a result, according to NNSA officials, NNSA has had to
follow both DOE's and NNSA's budgeting processes.
* Procurement. Although DOE made a commitment to issuing NNSA-specific
acquisition procedures in its January 2000 implementation plan for
NNSA, it has not done so. According to DOE Office of General Counsel
officials, the department subsequently determined that the proposed
NNSA-specific procedures were inconsistent with Federal Acquisition
Regulations, the NNSA Act, and the January 2000 implementation plan.
According to both DOE and NNSA officials, since 2004 the Department has
blocked NNSA's efforts to issue its own acquisition regulations. As a
result, according to NNSA officials, NNSA has had to issue a series of
deviations to the DOE acquisition regulations to carry out NNSA
acquisition policies in areas such as negotiating a more effective
contract fee arrangement and awarding additional years to a contract's
term.[Footnote 39]
* Information technology. The DOE and NNSA offices of the Chief
Information Officer have yet to reach agreement on a DOE order drafted
in September 2004 to define the relationship between the two offices.
According to DOE and NNSA officials, the draft order has not been made
final because of DOE's concern that the NNSA office would have too
great a role in tasks that should be performed by DOE's chief
information officer. Absent this order, certain problems have
developed. For example, according to a NNSA official, the two
information offices verbally agreed that all direction given to NNSA
contractors about information technology must come from NNSA, not DOE.
However, because this guidance was not formally documented, NNSA
contractors continue to receive dual requests for data from both the
DOE and NNSA offices of the Chief Information Officer. According to an
NNSA official, these dual requests have continually forced the
contractors to rank the requests.
* Management and administration. DOE and NNSA do not have a formal
process for obtaining DOE approval of NNSA-specific procedures,
although the Congress authorized the Administrator to establish NNSA-
specific policies unless disapproved by the Secretary of
Energy.[Footnote 40] However, DOE's Office of General Counsel has
delayed the development of this process because it believes that NNSA
should not be treated any differently than any other departmental
element and should thus be limited to using the flexibility already
available in DOE orders. As a result, NNSA's efforts to promulgate
Administration-specific procedures have become entangled in a
cumbersome review process with no certainty that closure of that review
will occur, according to NNSA officials. NNSA officials cited instances
in which the Administration has attempted to promulgate NNSA-specific
procurement regulations and policies on the roles of the DOE and NNSA
Offices of the Chief Information Officer. In these instances, the
review bogged down because a formal process for establishing NNSA-
specific procedures did not exist.
* Safeguards and security. DOE and NNSA have not established formal
procedures outlining the relative safeguards and security
responsibilities of the DOE and NNSA offices that oversee contractor
security activities. As a result, security officials at contractor
facilities told us they have received safeguards and security oversight
from a number of sources, including DOE; DNS; and the NNSA site offices
that oversee the contractor facilities. Furthermore, the oversight they
receive from these various sources can conflict. For example, DNS
requires contractors to demagnetize electronic media before disposing
of it. However, DOE Manual 470.4-4 requires physical destruction of
electronic media through pulverization or other appropriate methods.
Uncertain about which requirement to follow, the Sandia National
Laboratories requested, and was subsequently granted by the NNSA Sandia
site office manager, an exemption to DOE's physical destruction
requirement. However, the Office of Independent Oversight reported that
Sandia National Laboratories, the NNSA Sandia site office, and DNS had
not fully evaluated the risks associated with following alternative
destruction procedures. Sandia National Laboratories now follows the
DOE requirement.
While the lack of formal procedures has adversely affected some
functions, the presence of formal procedures has not always ensured
that the roles and responsibilities of DOE and NNSA offices are clearly
defined and that the offices operate together effectively and
efficiently. Specifically, the original NNSA Act prescribed roles and
responsibilities for DOE's Office of Counterintelligence and NNSA's
Office of Defense Nuclear Counterintelligence. Under section 3204 of
the NNSA Act, the director of DOE's Office of Counterintelligence was
responsible for establishing policy for counterintelligence programs
and activities at departmental facilities. At the same time, section
3232 of the NNSA Act specified that NNSA's Chief of Defense Nuclear
Counterintelligence was responsible for developing and implementing the
Administration's counterintelligence programs.
To better define the relationship between the two counterintelligence
offices, DOE issued a Secretarial Policy Memorandum in January 2001 and
a DOE order in December 2004. The Secretarial memorandum, in part,
directed that the DOE headquarters counterintelligence program offices
would be a shared resource between the DOE Office of
Counterintelligence and the NNSA Office of Defense Nuclear
Counterintelligence but that the funding needed for those program
offices would come from DOE's Office of Counterintelligence budget. The
December 2004 DOE order contained similar provisions but specified that
the headquarters inspections and polygraph programs would be directed
solely by the Director of DOE's Office of Counterintelligence.
According to DOE and NNSA counterintelligence officials, the
organizational relationship created by the Secretarial memorandum and
the DOE order led to the following problems:
* DOE and NNSA counterintelligence officials have disagreed over the
scope and direction of the counterintelligence program. NNSA
counterintelligence officials would like NNSA's counterintelligence
program to emphasize a multifaceted and proactive approach to prevent
intelligence activity. This program would integrate all of the tools of
the shared counterintelligence program offices--Analysis; Operations
and Investigations; Plans, Policy, Training and Awareness; and
Information and Special Technologies--as well as the DOE-directed
Polygraphs and Inspections program office. However, DOE's Office of
Counterintelligence emphasized a reactive approach to
counterintelligence--one that primarily focused on investigating
counterintelligence incidents that have already occurred. NNSA
counterintelligence officials stated that they would have liked to have
changed the scope and direction of the NNSA counterintelligence
program. However, such changes require the concurrence of DOE's Office
of Counterintelligence, according to NNSA counterintelligence
officials, which did not occur.
* DOE and NNSA counterintelligence officials also disagreed over their
ability to jointly direct the staff in the headquarters
counterintelligence program offices. Funding for these offices was
included in DOE's counterintelligence budget, and DOE's Office of
Counterintelligence senior management provided the program office
directors with their performance rating, with advisory input from the
NNSA counterintelligence office. Consequently, NNSA counterintelligence
officials believed that NNSA had less leverage in assigning tasks to
the program office staff, while DOE counterintelligence officials
believed that the DOE and NNSA counterintelligence offices had equal
leverage.
* DOE and NNSA counterintelligence officials disagreed over the
allocation of counterintelligence resources. The NNSA Chief for Defense
Nuclear Counterintelligence believed that counterintelligence resources
should be allocated according to risk management and priority-setting
principles. She noted that the risk associated with NNSA nuclear
weapons facilities was generally higher than the risk associated with
other DOE facilities because the loss of nuclear weapons information or
materials to foreign intelligence would constitute a serious breach of
national security. Therefore, in an August 2005 memorandum, the NNSA
chief asked the Director of DOE's Office of Counterintelligence to
comprehensively evaluate whether resources could be redirected to
benefit higher-risk counterintelligence targets and to provide guidance
for future decision making about resource allocation. As of November
2006, the DOE director had not responded to the NNSA memorandum.
* DOE and NNSA counterintelligence officials disagreed over
counterintelligence policymaking. For example, from September 2004 to
August 2005, the NNSA counterintelligence office unsuccessfully
attempted to obtain the DOE counterintelligence office's concurrence on
issuing a joint DOE/NNSA policy on personnel security file reviews and
sharing of counterintelligence information. When the DOE
counterintelligence office did not respond, the NNSA
counterintelligence office issued a NNSA-specific policy on August 12,
2005. This issuance without DOE's concurrence sparked a series of
verbal and e-mail exchanges between DOE and NNSA counterintelligence
officials, according to NNSA officials. DOE counterintelligence
officials stated in the e-mails we reviewed that NNSA did not have the
authority to issue NNSA-specific counterintelligence policy without DOE
concurrence. As such, the DOE counterintelligence office initially
demanded that the NNSA office retract the policy. Subsequently, in
April 2006, the DOE counterintelligence office issued a memorandum that
embraced the NNSA policy.
* DOE and NNSA counterintelligence officials disagreed over, or were
confused about, their roles and responsibilities in handling specific
counterintelligence matters. A case in point was the department's
handling of the well-publicized mid-2005 intrusion into an unclassified
NNSA computer system and removal of the names and social security
numbers of 1,502 individuals working for NNSA. The 1,502 NNSA
individuals were not notified of this incident until June 2006, after
the incident was acknowledged during a congressional hearing held
earlier in the month. Subsequently, the Secretary of Energy requested
that the Office of Inspector General examine the department's response
to this incident. In its July 2006 report, the Inspector General
identified multiple factors that contributed to the lack of a timely
response to this incident, including the delay in notifying the 1,502
NNSA individuals that certain personal information had been
compromised. The Inspector General recommended that the department
redefine and clarify the roles and responsibilities of DOE and NNSA
counterintelligence officials.
DOE and NNSA counterintelligence officials told us that they spent an
enormous amount of time resolving these types of problems between the
two offices and that this effort detracted from their focus on
counterintelligence activities. According to NNSA counterintelligence
officials, the arrangement did not meet the intent of the NNSA Act to
create a separately organized counterintelligence program within NNSA.
As a result, these officials added, NNSA's ability to execute a
separate counterintelligence program that meets its unique needs was
impeded and largely limited to what the two counterintelligence offices
could agree on.
Subsequently, the Congress amended the NNSA Act to address the problems
with the NNSA's and DOE's counterintelligence programs. Specifically,
section 3117 of the John Warner National Defense Authorization Act for
Fiscal Year 2007 contained provisions to consolidate the
counterintelligence programs of DOE and NNSA under the Department of
Energy.[Footnote 41] Section 3117 also established the Intelligence
Executive Committee within the Department of Energy, consisting of the
Deputy Secretary of Energy, who would chair the committee, and each
Under Secretary of Energy. The section requires the Secretary of Energy
to use the committee to assist in developing and promulgating the
counterintelligence and intelligence policies, requirements, and
priorities of the entire Department. The section also established
within the staff of the Administrator, NNSA, a position of NNSA
Intelligence and Counterintelligence Liaison who would be responsible
for liaison between the NNSA and the Department's Office of
Intelligence and Counterintelligence. The amendment sunsets on
September 30, 2010, and the NNSA functions, personnel, funds, assets,
and other resources transferred to the Secretary of Energy will be
returned to the control of the Administrator, NNSA.
The conference report stated that the Congress reluctantly provided
this authority to consolidate the program despite the ongoing
skepticism of many in the Congress over the Department's ability to
implement a strong security program for the Department and its
laboratories.[Footnote 42] The conference report noted that with the
recent Departmental initiative to combine the Offices of Intelligence
and Counterintelligence and with the authority provided in this
provision, the intelligence and counterintelligence functions would be
organized as they were when the Department experienced significant
counterintelligence problems. The conference report stated that the
conferees believed that the Department could have addressed many of the
perceived issues associated with having separate counterintelligence
offices for the Department and NNSA by demonstrating greater management
resourcefulness. The report also stated that the conferees were
unpersuaded that the Department fully and faithfully implemented the
counterintelligence program structure called for in the NNSA Act.
Nevertheless, the conference report noted that the Department had
continued--under the leadership of two different Secretaries of Energy
and two different Deputy Secretaries--to identify the
counterintelligence structure of the NNSA Act as an impediment to the
smooth functioning of security operations within DOE. Therefore, the
conferees agreed to include a provision that would provide temporary
authority for the disestablishment of the Office of Defense Nuclear
Counterintelligence within NNSA and the transfer of NNSA
counterintelligence personnel to the Department's Office of
Counterintelligence. DOE officials told us that they are in the process
of developing plans to implement the consolidation of the
counterintelligence programs and expect full implementation in early
2007. According to these officials, the consolidation should be a
seamless process and will provide the leadership necessary to
effectively implement DOE's counterintelligence activities.
External Entities Continue to Express Concern about NNSA's
Organizational Relationship:
Almost 7 years after NNSA was created, concerns persist about the
organization and management of the nuclear weapons complex that led to
NNSA's establishment. For example, in January 2006, DOD's Defense
Science Board Task Force on Nuclear Capabilities reviewed alternative
institutional arrangements that it believed could provide for more
efficient management of the nuclear enterprise. The task force
expressed concern that other elements of DOE were continuing to
micromanage NNSA. To correct this situation, the task force proposed
the following three options:
* Reform the existing structure. The task force would accomplish this
reform by having the Secretary of Energy enforce the requirement that
the Administrator be autonomous and accountable only to the Secretary.
The task force believed that this approach would require a significant
change in DOE culture, but it had little confidence in the prospect for
effective, lasting change within DOE.
* Move NNSA to DOD. The task force believed that there was more
confluence between DOD's core missions and the nuclear weapons complex.
However, the task force noted that, historically, the Congress has been
reluctant to concentrate nuclear weapons authority in a single
department and that, on balance, DOD was not a good fit for the
development and production of nuclear weapons.
* Create a National Nuclear Weapons Agency. The task force supported
this option, believing that the nuclear weapons enterprise was unique
in its missions and the demands placed upon managers and that a unique
organizational approach was therefore warranted. The National Nuclear
Weapons Agency would report through a Board of Directors, comprising
the Secretaries of several cabinet offices, including Energy and
Defense, to the President.
However, former senior DOE and NNSA officials, as well as, one Defense
official with whom we spoke, generally did not favor removing NNSA from
DOE. These officials included the first and second Administrators of
NNSA, the former Deputy Secretary of DOE, and a senior member of the
Defense Science Board Task Force responsible for the January 2006
report. These officials offered the following comments on NNSA's
organization:
* According to the second NNSA Administrator, it would not be a good
idea to make NNSA a totally independent agency. The Administrator
indicated that NNSA is very much entangled in the operations of DOE and
the costs associated with making NNSA totally independent from DOE
would outweigh the benefits.
* One former senior official told us he never knew what it meant to
have NNSA as a separately organized agency. In this official's view,
because DOE and NNSA had not clearly defined the concept, there was
some uncertainty in the relationship between the DOE Secretary and the
Administrator of NNSA. While this official would not be in favor of
establishing NNSA as a totally separate agency, he believes that NNSA
needs to take a half-step away from DOE to gain more independence.
* According to another former senior official, the DOE/NNSA
organizational relationship might have benefited from a more formal
definition. Regarding organizational changes, in this official's view,
the Congress should either make NNSA a totally independent agency or
fold it back into DOE.
* NNSA is not a separately organized agency, a Defense Science Board
Task Force official told us. In this official's view, DOE ensured that
fate by deciding, in its 2000 implementation plan, that certain
functions would not be transferred to NNSA. When NNSA was created, this
official said, the Congress was clear in the enabling legislation about
the type of entity it wanted NNSA to be. However, the Congress has
allowed the problems associated with DOE's implementation to languish.
NNSA Needs to Fully Resolve Its Human Capital Issues:
Under GAO's internal control standards, agencies should ensure that
they have a sufficient number of staff, particularly in managerial
positions, and that staff have the skills required to carry out the
agency's organizational objectives. While NNSA has engaged in a variety
of human capital activities since its creation, DOE and NNSA have yet
to fully determine the appropriate size of their respective workforces.
In its January 2000 NNSA Implementation Plan, DOE outlined a series of
steps it would take to staff NNSA. In order to be prepared when the
NNSA Act took effect on March 1, 2000, DOE indicated that current
employees assigned to those program offices that would become part of
NNSA would be transferred to NNSA. For other employees, such as those
in departmental support offices, DOE stated that a more detailed
analysis would be necessary to determine those employees subject to
transfer.
In February 2002 testimony before the Special Oversight Panel on
Department of Energy Reorganization of the House Armed Services
Committee, the then Administrator stated that NNSA was taking a "phased
approach" to assess the status of each NNSA function and the costs and
benefits of continued reliance on DOE for service. For example, in the
area of human resources, the former Administrator stated that NNSA had
decided to develop an independent capability to select, promote, and
develop its own executive workforce. In other areas, such as the
investigation of serious accidents, NNSA would continue to rely on DOE
personnel. However, DOE and NNSA have never conducted a systematic,
detailed analysis of NNSA's staffing needs in relation to DOE's.
During this review, we identified areas where potential staff
imbalances have affected NNSA's ability to operate as a separately
organized agency within DOE. These imbalances occur even though NNSA's
federal staff deals with about 40 percent of DOE's budget and programs,
according to a June 2006 DOE analysis. The imbalance is most notable in
NNSA's Office of General Counsel. Section 3217 of the NNSA Act
established the position of the NNSA General Counsel to serve as NNSA's
chief legal officer. Initially, DOE assigned the DOE General Counsel to
also serve as the NNSA General Counsel. This dual-hatting was
subsequently prohibited by an amendment to the NNSA Act.[Footnote 43]
NNSA has since hired a General Counsel and 34 other attorneys to
provide NNSA with legal advice and analysis. In contrast, the rest of
DOE has 277 attorneys. According to NNSA's General Counsel, who has
been at NNSA since September 2001, no formal analysis of his office's
staffing needs has ever been done. However, he believes that such an
analysis would show that his office needs about 15 to 20 additional
attorneys to fully handle NNSA's legal workload with minimal assistance
from DOE. Absent additional attorneys, NNSA must rely on DOE's Office
of General Counsel to perform a significant portion of its legal work.
The June 2006 DOE analysis also showed a large difference between the
size of the NNSA federal staff compared with the rest of DOE federal
staff in other areas. NNSA concurred with this analysis. (See table 7.)
Table 7: Staffing Levels within NNSA and DOE Offices:
Function: Public Affairs;
NNSA staffing: 16;
DOE staffing: 74.
Function: Congressional Affairs;
NNSA staffing: 5;
DOE staffing: 23.
Function: Security;
NNSA staffing: 131;
DOE staffing: 181.
Function: Planning, Programming, Budgeting, and Evaluation;
NNSA staffing: 124;
DOE staffing: 555.
Function: Information Technology;
NNSA staffing: 50;
DOE staffing: 522.
Source: DOE Office of Human Capital Management.
[End of table]
According to the first Administrator, because of the difference in
staff size, DOE can control NNSA's agenda by deciding which issues will
get reviewed and when. In his view, this problem began when NNSA was
created and has only gotten worse. He also stated that because NNSA did
not have a plan for how it would fully staff its headquarters
organization, he tried to hire the people he needed one at a time. An
April 2005 NNSA report on its security program raised similar
concerns,[Footnote 44] noting that the disparity in size between NNSA
and DOE headquarters staffs greatly impaired NNSA's effectiveness
within the headquarters bureaucracy. For example, the report said, DOE
dominated the development of security policy because its staff was
significantly larger. In addition, a May 2005 NNSA staffing analysis
noted that the far larger DOE staff are able to devote considerable
effort to "second-guessing" NNSA actions or to excessive requests for
data and that other DOE staff sometimes seek to interfere with NNSA.
Reconciling any disparity between the size of the NNSA and DOE staffs,
and determining the staff NNSA needs to carry out its overall mission,
should take place within the context of a broader, data-driven
workforce plan. In December 2001, we reported that NNSA did not have
the coherent human capital and workforce planning strategies it needed
to develop and maintain a well-managed workforce over the long run. We
therefore recommended that, before NNSA allocated any additional
excepted service positions, it develop a thorough human capital and
workforce planning strategy.[Footnote 45] In June 2004, we reported
that while the workforce plan NNSA developed in December 2003 attempted
to establish a framework for long-term workforce planning, the plan was
of limited use because it did not have current statistics on workforce,
positions, and organizational structures.[Footnote 46] Furthermore, the
downsizing NNSA was completing at the time was taking place in an
unstructured environment. Under these conditions, agencies can
experience significant challenges in deploying people with the right
skills, in the right places, and at the right time, and in ensuring
that they perform their missions economically, efficiently, and
effectively. In NNSA's case, we noted at the time that there were early
indications that the lack of planning was contributing to skill
imbalances such as those imbalances in security staff already
discussed. We further stated that without a functional long-term
workforce plan, NNSA ran the risk of facing more serious staff
shortages or skill imbalances in the future, which would affect its
ability to adequately oversee its contractors and ensure the safety and
security of its various facilities.
To its credit, NNSA has recognized that it needs to improve its human
capital management, and in a November 28, 2005, memorandum to NNSA
senior managers, the Administrator announced a workforce planning
initiative. The Administrator acknowledged that NNSA had no systematic
way to assess current, or forecast future, human capital needs.
Consequently, he asked the NNSA Office of Human Resources to develop
and refine a workload analysis and planning approach. While the
Administrator expected that this approach would be usable by the middle
of fiscal year 2006, according to the Director for NNSA's Office of
Human Resources, he made certain changes to the workload analysis to
ensure that more appropriate data could be collected. As a result of
those changes, the Director said, the workload analysis and planning
approach would take longer to complete.
NNSA Can Take Additional Action to Improve Project Management:
Project management remains a significant concern for NNSA and DOE and,
over the last several years, both have initiated improvement efforts.
For example, NNSA has implemented earned value management systems to
improve project reporting, conducted independent project reviews to
diagnose problem areas before the projects are executed, and, in April
2006, completed training and certifying its project managers. In July
2006, DOE issued a new program and project management order.
Nevertheless, DOE's fiscal year 2006 performance and accountability
report still identified project management as a significant issue.
Furthermore, according to NNSA's November 2006 report on project
management performance, 5 of 31 NNSA projects, or about 16 percent, are
in jeopardy of breaching their cost baseline, schedule baseline or
both. To further improve NNSA's project management performance, we
identified seven areas for additional action.
* Develop a project management policy. Any improvement requires a
policy that clearly articulates management's vision. While NNSA has
been involved in drafting a revised DOE order on project management,
and creating and promulgating DOE project management guides, in an
April 2003 memorandum, the NNSA Administrator committed to issuing an
NNSA-specific project management policy to (1) further assist in the
implementation of the DOE project management manual issued in March
2003 and (2) assure consistent application with the NNSA Act. However,
according to an official in NNSA's Office of Project Management and
Systems Support--which would be responsible for developing a project
management policy--no NNSA project management policy has been issued.
* Prepare a project management improvement plan. Short-and long-range
plans are critical to implementing a project management policy.
However, NNSA developed its last project management improvement plan in
December 2000. According to information presented at NNSA's 2005
Project Management Training Workshop, the Administration successfully
accomplished many of the eight goals in this plan, including using
qualified, experienced project managers to oversee NNSA projects and
establishing formal procedures for obtaining senior manager review and
approval of projects. However, the NNSA staff presenter at the workshop
also noted that NNSA was not completely successful in creating a
cultural shift in NNSA that emphasizes the importance of project
management or creating incentives to improve project managers'
performance. As a result, NNSA has not significantly reduced project
delivery times or costs. The NNSA staff presenter, therefore, proposed
the development of new short-and long-range improvement plans, which
officials from NNSA's Office of Project Management and Systems Support
acknowledged the Administration currently does not have. While these
officials acknowledged the need for such plans, NNSA management
indicated that it was not aware of the staff proposal.
* Prepare project management annual reports. After its creation, NNSA
began preparing an annual status report to the Congress on its
construction project accomplishments. However, it submitted only two
reports--for fiscal years 2000 and 2001. In fiscal year 2000, NNSA
reported, among other things, that about 70 percent of its Defense
Programs' 46 projects were progressing satisfactorily and that Defense
Programs had many success stories during the year, including (1)
reestablishing the quarterly project review process, (2) developing and
implementing a project management improvement plan, and (3) sponsoring
a project management workshop. However, NNSA also reported that the
root causes of project management problems were deeply entrenched and
could not be eliminated without a basic cultural change, which would
require at least 3 years. In fiscal year 2001, NNSA reported that 30,
or about 73 percent, of its 41 projects were progressing satisfactorily
and noted accomplishments in project management. However, NNSA also
reported that (1) management costs for projects were nearly double
those of other organizations and (2) projects took approximately 3
years longer to accomplish than similar projects performed elsewhere.
The Director of NNSA's Office of Project Management and Systems
Support--who assumed his position in 2004--did not know why the reports
were discontinued. Other NNSA officials in the Office of Project
Management and Systems Support told us that the reports were
discontinued because statements, such as those contained in the fiscal
year 2001 report, cast NNSA in too unfavorable a light.
* Include major projects in the Project Assessment and Reporting System
(PARS). In 2001, DOE implemented PARS--a Web-based system for keeping
DOE senior managers apprised about the performance of projects costing
more than $5 million. In response to a February 2004 memorandum from
the Deputy Secretary of Energy citing inadequate management of critical
activities within DOE and NNSA, NNSA indicated that it intended to
treat more of its activities as projects so that the cost and schedule
performance associated with those activities could be improved. While
NNSA has applied project management principles to more of its
activities, we identified four major NNSA projects--estimated to cost
over $100 million each--that are not enrolled in the PARS system and do
not receive senior management oversight through that system (See table
8.) NNSA indicated that these projects are not in PARS because they are
information technology projects or programmatic efforts. However, we
believe each listing in table 8 meets DOE's definition of a project
because it has a defined start and end date. In addition, information
technology projects are not prohibited from inclusion in PARS.
Table 8: Major Projects Not Included in PARS:
Project name: B-61 Life Extension;
Description: NNSA expects this project, designed to extend the service
life of the B-61 bomb, to cost about $287 million between fiscal years
2005 and 2009.[A].
Project name: W-76 Life Extension;
Description: NNSA expects this project, designed to extend the service
life of the W-76 nuclear warhead by replacing components, to cost about
$1.05 billion between fiscal years 2005 and 2011.
Project name: Enterprise Project;
Description: NNSA increased the total cost of this project, which will
replace the accounting and management systems at Los Alamos National
Laboratory, from about $70 million when it was initiated in 2001 to
nearly $160 million.
Project name: Purple and BlueGene/L Supercomputers under the Advanced
Simulation and Computing Program;
Description: NNSA expects this project to cost about $290 million.
Source: GAO analysis of DOE information.
Note: The W-80 Life Extension was also not included as a project in
PARS; however, NNSA recently announced that this effort was being
terminated.
[A] The cost for the B-61 and W-76 life extension projects is contained
in NNSA's fiscal year 2007 budget.
[End of table]
We are particularly concerned that the life extension projects were not
included in the PARS database. In our July 2003 report on NNSA's life
extension program, we recommended that NNSA establish the individual
life extensions as projects and manage them according to DOE's project
management requirements.[Footnote 47] As we reported, the then NNSA
Deputy Assistant Administrator for Military Application and Stockpile
Operations acknowledged that the three life extensions were projects,
and NNSA concurred with our recommendations. During our current review,
however, we could not find any evidence that NNSA had taken any action
to establish the life extensions as projects.
* Complete a lessons learned report database. DOE's project management
manual states that a lessons learned report should be completed,
distributed, and included in a project's permanent file; however, the
manual does not require the establishment of a comprehensive lessons
learned project management database. To its credit, the Administration
has attempted to create such a database on one of its internal Web
pages, but we found that certain key NNSA lessons learned reports were
not included, including, for example, reports for NNSA's Accelerated
Strategic Computing Initiative and W-87 Life Extension. We believe that
these reports provided information worthy of dissemination across the
NNSA complex. For example, the W-87 lessons learned report noted that
the project plan was prepared too late in the development cycle and was
not used as a tool to identify problems and take corrective actions. In
addition, the Accelerated Strategic Computing Initiative report noted
that the Lawrence Livermore and Los Alamos National Laboratories had
mixed success, in part, because they did not look at the lessons
learned from other communities. While officials in the Office of
Project Management and Systems Support acknowledged that the Web page
was incomplete, they said the forthcoming DOE project management order
would reinforce the importance of lessons learned information. However,
the final order, issued in July 2006, did not discuss a lessons learned
database; instead, it required that a lessons learned report be
prepared and submitted to DOE's Office of Engineering and Construction
Management for broader sharing among the DOE project management
community. According to that office, it has yet to develop internal
procedures for sharing this information.
* Develop contractor performance benchmarks. Benchmarking is an
integral part of project management.[Footnote 48] It is used to analyze
the gap between actual and preferred performance in order to identify
opportunities for improvement. We found that while NNSA tracks cost and
schedule performance on individual projects through various reporting
systems, including information contained in the PARS system, neither
the PARS system nor any other departmental system ranks the collective
project management performance of NNSA's contractors. On its own
initiative, however, the Los Alamos National Laboratory contracted in
2004 to have its project management performance benchmarked against
comparable organizations. The August 2004 report found that, on a scale
of 1 to 5, with 1 as the lowest, Los Alamos' project management rated a
2. Los Alamos had project management processes, although such processes
were not considered an organizational standard on all projects.
According to NNSA site office managers we contacted, it would be
helpful to know how their contractor ranked in project management
performance against other NNSA contractors. The Sandia Site Office
Manager added that this ranking, as well as an NNSA national average,
would help determine if project management at her laboratory was doing
well or not. Overall, the managers added, if their contractor's project
management performance was not at a comparable level to other
contractors, they could identify problems and develop solutions.
* Require contractors to receive project manager training. DOE and NNSA
have initiated a program to train and certify all departmental project
managers on projects costing $5 million or more. However, because
departmental projects are usually undertaken by an outside contractor,
the training of the contractor's project manager is equally important.
Several studies conducted at the request of either DOE, NNSA or their
contractors have recommended training and certifying contractors'
project managers. For example, in 1999, the National Research Council
found that the lack of project management skills was a fundamental
cause of poor project performance and recommended that DOE require all
contractors' project managers to be experienced, trained, and qualified
in project management appropriate to the project.[Footnote 49]
Similarly, in 2004, the Civil Engineering Research Foundation reported
that project management competence was a critical factor in project
success and recommended that formal project management training of
contractor project managers be made a contract requirement.[Footnote
50] Despite this support for contractor project manager training and
certification, neither DOE nor NNSA has a policy on contractor project
manager training. According to an NNSA official, the Administration has
taken the position that it will stipulate in its contracts what it
wants accomplished and then leave it up to the contractor to determine
how best to achieve NNSA's requirements.
NNSA Can Also Improve Program Management Through Additional Action:
Program managers are responsible for accomplishing an organized set of
activities directed toward a common purpose or goal. To fulfill that
responsibility, NNSA program managers are expected to have a working
knowledge of such diverse areas as contracting, budgeting, and
engineering. Recognizing the importance of program managers, NNSA has
taken several actions, such as initiating a project and program
management improvement team and developing a program management policy.
However, DOE's performance and accountability reports indicate that
additional improvements in program management are possible. For
example, according to DOE's fiscal year 2006 performance and
accountability report, NNSA fully met only about 52 percent of its
program goals, compared with about 79 percent for the rest of DOE.
Similarly, according to departmental information, NNSA met about 74
percent of its performance targets from fiscal years 2002 through 2006,
compared with 87 percent for the rest of DOE. To further improve its
program management performance, NNSA will need to take the following
actions:
* Prepare status reports on program management improvement. In any
performance-based management system, once a problem has been
identified, a corrective action plan is prepared and reports are
generated to periodically evaluate the status of the corrective action.
In a July 2004 memorandum to the DOE Under Secretary, NNSA indicated
that each NNSA office would prepare a corrective action plan, known as
a program management implementation plan. While all NNSA offices
completed at least a portion of these implementation plans, only the
offices of Defense Programs, Nuclear Nonproliferation, and Defense
Nuclear Security outlined specific tasks, or deliverables, in their
plans that needed to be completed. More importantly, of these offices,
only Defense Programs and Nuclear Nonproliferation identified a due
date for each task. Plans prepared by the offices of Emergency
Operations, and Infrastructure and Environment did not include any
specific tasks or due dates. Moreover, none of the NNSA offices have
prepared periodic reports on the status of their efforts to improve
program management. Therefore, it is unclear how much progress each
NNSA office has made.
* Complete a best practices guide on program management. In response to
the Government Performance and Results Act, agencies have learned that
they can improve performance by emulating the best practices of leading
organizations. In October 2003, an internal review of NNSA's program
management found that, among other things, NNSA needed to have (1) a
common language and terminology about program management, (2)
consistency in program manager functions and qualifications, and (3) a
clear channel of direction and consistent and timely flow of
information from NNSA programs to contractors.[Footnote 51] It
recommended that NNSA develop and issue a best practices guide by
December 1, 2003. The draft guide was never issued in final form and
disseminated within NNSA.[Footnote 52] According to a NNSA official and
a NNSA contractor official who served on the program management team,
the best practices guide floundered after NNSA made certain
organizational and employee changes and the guide no longer had a
person serving as a champion of program management improvement.
* Identify and train program managers. NNSA's August 2004 program
management policy largely superseded earlier NNSA team recommendations
on program manager training.[Footnote 53] According to the August 2004
policy, NNSA's Office of Management and Administration was responsible
for developing and maintaining a database of NNSA program managers to
identify (1) the manager for each program, (2) the training and
certification requirements for each program manager position, (3) the
dates the program manager initially completed various training and
certifications, and (4) when, if applicable, the training and
certification must be renewed. The policy did not stipulate any
deadline for the database's completion. Consequently, according to an
official in NNSA's Office of Management and Administration, this
program manager database has not been created. In contrast, according
to an official in DOE's Office of Engineering and Construction
Management, the department has developed a training curriculum for DOE
program managers with approval of that curriculum by the DOE Deputy
Secretary expected in the next few months. Unlike the NNSA program
management policy, DOE will not initially require that program managers
be certified on the amount of training received.
Certain Financial Management Weaknesses Remain Despite NNSA
Improvements:
Under section 3252 of the NNSA Act, NNSA had to develop a planning,
programming, and budgeting process that comported with sound financial
and fiscal management principles.[Footnote 54] The NNSA Act also
requires NNSA to annually submit to the Congress a Future Years Nuclear
Security Program plan that details NNSA's planned expenditures for the
next 5 years.[Footnote 55] Very early in his tenure, the first NNSA
Administrator indicated that he intended to comply with the NNSA Act by
instituting a planning, programming and budgeting process similar to
DOD's. While DOD's approach has not been without its problems over the
past 40 years, it is generally recognized as a system that, when
properly led and staffed, can provide cost-effectiveness comparisons
and develop the detailed program and budget plans called for in the
NNSA Act. Although NNSA has made significant progress in implementing
its PPBE process over the last 4 years, it still needs to (1) complete
its PPBE policy guidance, (2) establish an independent analysis unit to
assist the Administrator in making budget allocation decisions, (3)
develop criteria for compiling an NNSA-wide priorities list, (4)
develop an up-to-date schedule of those programs requiring budget
validation, and (5) establish a DOE-compliant budget validation
process.
* Complete NNSA's PPBE policy. As recognized in Office of Management
and Budget (OMB) Circular A-123, an important part of internal controls
over financial reporting is having policies and procedures in place to
ensure that management directives are carried out and that management's
assertions in its financial reporting are valid. To this end, NNSA has
issued 10 policy letters providing guidance to NNSA staff on various
aspects of the PPBE process: 6 of these 10 letters are final, 3 have
been in draft for more than 1 year, and 1 has been in draft for more
than 2 years. According to the director of NNSA's Office of Planning,
Programming, Budgeting, and Evaluation, the four draft policy letters
have not been made final, in part, because NNSA has been waiting to see
DOE's position on issues in the draft letters. However, we believe that
waiting 2 years for DOE's views is inconsistent with NNSA's operation
as a separately organized agency with its own PPBE budgeting process.
NNSA indicated that it is continuing to evaluate aspects of the PPBE
budgeting process and expects to finalize at least three of the draft
policy letters during fiscal year 2007.
* Establish an independent analysis unit. PPBE provides a mechanism for
centralized resource allocation decisions. In NNSA, that mechanism
consists of having senior NNSA managers develop a budget for their
program and then collegially review, through a management council, the
budgets from the other programs, including the programs' justification
for additional resources to meet requirements they cannot fund within
their target allocation. The council can either decide to shift funding
between programs to meet such requirements or determine whether the
issue should be presented to the Administrator for resolution. The
Administrator reviews the council's decisions, resolves issues
identified by the council, and makes final resource allocation
decisions. Such a mechanism places considerable pressure on the
Administrator alone to make difficult funding decisions.
While the NNSA PPBE process was modeled after the process DOD uses, it
differs markedly in one aspect. Unlike DOD, NNSA has not established an
independent group responsible for reviewing program proposals,
verifying cost estimates, and analyzing alternatives. In August 2003,
DOE's Office of Inspector General reported that most NNSA senior
managers interviewed believed such an analytical group could be of
value.[Footnote 56] For example, the then NNSA Acting Associate
Administrator for Management and Administration said that he supported
the formation of such a group to provide objective analyses to the
Administrator when he had to move resources between programs. Several
senior managers also stated that resource allocation decisions among
program offices were likely to become more frequent in future years
and, therefore, NNSA's need for independent analytical services was
likely to grow. Although NNSA agreed with the Inspector General's
recommendation regarding the need for an independent analytical support
group, it did not believe that such a group could be created before the
end of calendar year 2005, after downsizing and reorganization had been
completed.
NNSA indicated that its management is fully supportive of the
importance of independent analysis at both the NNSA and DOE level, but
mindful that the resources needed to create this capability must be
balanced with all of the competing needs within NNSA in a highly
constrained funding and staffing environment. However, in our view, in
such an environment, the need for an independent budget analysis unit
to ensure that appropriate budgetary decisions are made becomes even
more important.
* Develop criteria for compiling an NNSA-wide priorities list. In order
to make trade-off decisions among organizations and programs within an
organization, DOE formally requires each of its organizations to
annually submit a report to its Office of Budget that ranks its
programs for the budget year and provides a rationale for the ranking.
As we noted in July 2003, NNSA did not submit these ranking reports for
fiscal years 2002 through 2004.[Footnote 57] It submitted a list for
the first time for the fiscal year 2007 budget, after the Secretary of
Energy specifically requested it. According to an NNSA budget official,
each NNSA program, such as Defense Programs and Nuclear
Nonproliferation, prepared its list individually. Afterwards, this
official stated, the NNSA Administrator compiled an overall priority
list without using any formal criteria.
According to an official in DOE's Office of Budget, his office had some
questions about the ranking of activities on the NNSA list. For
example, the impact associated with not funding a particular activity
was not always sufficiently clear to explain NNSA's ranking rationale.
While NNSA subsequently answered all of the Office of Budget's
questions satisfactorily, he noted that NNSA might have benefited from
preparing and following some formal criteria for developing its
priority ranking list, as do other DOE organizations, such as the
Office of Environmental Management. Formal criteria, in this official's
view, would consider the impact of not funding an activity plus the
probability of success if additional funding occurred.
NNSA indicated that it annually issues strategic planning guidance and
program and fiscal guidance, holds internal NNSA office programming
sessions, and convenes a senior NNSA leadership meeting to discuss the
various NNSA office proposals. NNSA added that, considering all these
actions and balancing competing priorities, the Administrator makes his
decisions which are documented in the Administrator's Final
Recommendations report. In our view, having formalized criteria to
follow would assist the Administrator in making his decisions and
render those decisions more transparent and defensible to outside
review.
* Develop an up-to-date schedule of those programs requiring budget
validation. Under the PPBE process, once NNSA establishes the relative
priority of its programs, the budget associated with those programs
must be developed and validated. During validation, NNSA evaluates the
need for the program's proposed activities and the cost estimates in
support of those activities. In a May 2004 memorandum, NNSA's Office of
Management and Administration announced the Administration's 5-year
schedule for conducting budget validations for fiscal years 2005
through 2009. NNSA indicated that it would select programs for
validation on the basis of several factors, including programs
receiving OMB Performance Assessment Rating Tool reviews, program
manager requests, Administrator direction, significant external
interest, and high program visibility. Using those factors, for fiscal
year 2006, NNSA planned to validate the budgets for the Office of the
Administrator, Science Campaign, Readiness Campaign, Safeguards and
Security, International Nuclear Materials Protection and Cooperation,
Elimination of Weapons-Grade Plutonium Production, and Naval Reactors
Operations and Maintenance, and for the fiscal year 2007 budget, it
would validate Directed Stockpile Work/Stockpile Services, Nuclear
Weapons Incident Response, Nonproliferation and Verification research
and development, and Highly-Enriched Uranium Transparency
Implementation.[Footnote 58]
While NNSA followed the May 2004 schedule for fiscal years 2006 and
2007, it had not updated this schedule as of November 2006 to reflect
program changes. For example, new major program initiatives--such as
the Reliable Replacement Warhead and the efforts to transform the
nuclear weapons complex, called Responsive Infrastructure--are not
listed on the May 2004 validation schedule for fiscal years 2005
through 2009. While the amounts requested in the fiscal year 2007
budget for these two initiatives are low compared with other program
efforts, both initiatives are considered centerpieces in the NNSA
budget, will require significant additional funding in succeeding
fiscal years, and meet NNSA's selection criteria for program validation
on the basis of significant external interest and high program
visibility. Because the May 2004 schedule has not been updated, NNSA
does not know when certain programs like the Reliable Replacement
Warhead and Responsive Infrastructure would be validated, according to
an official in NNSA's Office of Planning, Programming, Budgeting, and
Evaluation. This official added that NNSA is contemplating updating the
program validation schedule later in 2006.
NNSA indicated that it would not be beneficial to add the Reliable
Replacement Warhead and Responsive Infrastructure programs to its
budget validation schedule until DOE and the Office of Management and
Budget have made decisions to proceed with these activities. However,
both of these programs are ongoing and are expected to grow; therefore,
in our view, they should be considered when NNSA updates its budget
validation schedule.
* Establish a DOE-compliant budget validation process. NNSA's May 2004
draft budget validation policy also outlined the Administration's
approach to budget validation. Specifically, the Administration will
follow a two-phased validation approach. During phase one, NNSA will
assess selected programs' conformance with strategic guidance, program
plans, and performance data and seek insight into how each organization
formulates its budget. During phase two, NNSA will choose a few of the
programs from the phase one review and conduct a detailed review that
examines the reasonableness of the cost estimates supporting the
program's budget. Following this two-phased approach, NNSA performed a
detailed validation review on only approximately 12 percent of the
Administration's budget submitted for fiscal years 2006 and 2007.
However, in our view, this approach does not adhere to DOE's guidance
on that subject. According to the director of DOE's Budget Operations
Division, commencing in fiscal year 2001 as a performance measure, DOE
has required that at least 20 percent of the budget undergo a detailed
review, so that 100 percent of the budget would be evaluated every 5
years. In our 2003 report on NNSA's Stockpile Life Extension Program,
we also found that NNSA was not conducting proper budget validation
reviews and, consequently, recommended that NNSA validate its stockpile
life extension budget request in accordance with DOE
directives.[Footnote 59] NNSA indicated that its budget validation
review focuses on the process used in developing budget requests and is
not an audit of the resulting budget estimates. NNSA further indicated
that it does not have the capabilities to audit and instead relies on
DOE's Office of Inspector General and GAO for those reviews. While such
reviews are helpful in evaluating budget estimates, we believe these
reviews do not abrogate NNSA's responsibility to adhere to DOE
requirements.[Footnote 60]
NNSA also noted that the results of its progress in implementing PPBE
are reflected in (1) the average scores received through the Office of
Management and Budget's Program Assessment Rating Tool scores which are
among the highest in DOE and the U.S. Government, (2) NNSA management
and operating contractors working without serious interruption through
multiple continuing resolutions in the past four years, and (3) NNSA
transitioning smoothly to the DOE budget formulation process in the
summer of 2006.
Conclusions:
Prior to the creation of NNSA, DOE's management of the nuclear weapons
complex was widely considered, in the words of the President's Foreign
Intelligence Advisory Board as "science at its best" and "security at
its worst." Underpinning this characterization was a long departmental
history of organizational disarray and poor security, project and
program management. Consequently, producing a well-organized and
effective agency out of what was widely considered a dysfunctional
enterprise has been a considerable challenge. In some areas, NNSA
should be viewed as a success. Most notably, through its internal
reorganization efforts, NNSA has addressed many major organizational
issues and has made important progress in establishing critical
management systems, especially PPBE.
However, without a workable model on how a separately organized agency
should function, NNSA and DOE officials have had to develop their
working relationships largely on a case-by-case basis, often with
limited success. The fact that the Congress felt it necessary to step
in and modify the NNSA Act to address continuing organizational
conflict within the counterintelligence program demonstrates, in our
view, that DOE and NNSA need to take a more active approach to clearly
defining DOE and NNSA's working relationships and determining how
conflict will be resolved. While there have been continuing calls for
removing NNSA from DOE and establishing it as a separate agency, we do
not believe that such drastic change is necessary to produce an
organization that can provide effective oversight of the nation's
nuclear weapons complex.
Beyond addressing organizational issues, if NNSA is ultimately to
provide comprehensive oversight of the operation and security of the
nation's nuclear weapons programs, it must address a variety of
lingering, often unrelated, but important management issues. These
issues include providing sufficient, qualified staff to conduct program
and operational oversight, especially in the security area, and
developing and implementing improvements needed to support effective
project, program, and financial management.
Recommendations for Executive Action:
To improve NNSA's efficiency and effectiveness, we are making 21
recommendations in five areas--organization, security, project
management, program management, and financial management.
To ensure that NNSA functions as a separately organized agency, we
recommend that the Secretary of Energy and the Administrator, NNSA,
take the following two actions:
* clearly define NNSA's status as a separately organized agency within
the department; and:
* develop and implement standard operating procedures for conducting
business and resolving conflicts between DOE and NNSA sister offices.
To improve security oversight, we recommend that the Administrator,
NNSA, take the following four actions:
* develop a human capital strategy that includes standards for
determining long-term staffing needs;
* implement a professional development program for security staff to
ensure the completion of training needed to effectively perform
oversight responsibilities;
* develop a framework to evaluate results from security reviews and
guide security improvements; and:
* establish formal mechanisms for sharing and implementing lessons
learned across the weapons complex.
To fully implement NNSA's management reforms, we recommend that the
Administrator, NNSA, institute a series of initiatives to improve
project management, program management, and financial management.
With respect to project management, we recommend that NNSA take the
following seven actions:
* establish an NNSA-specific project management policy to ensure
application of the DOE project management manual;
* prepare a project management improvement plan;
* reinstitute annual reporting to the Congress on project management
accomplishments;
* include major projects, such as the Stockpile Life Extension
refurbishments, in DOE's Project Assessment and Reporting System;
* complete a comprehensive database of all projects' reports on
management lessons learned to improve project management throughout
NNSA;
* develop benchmark data on individual contractors' project management
performance to assist managers in improving contractor performance;
and:
* require that contractors' project managers receive project manager
training and attain project manager certification.
With respect to program management, we recommend that NNSA take the
following three actions:
* prepare periodic reports that show the status of its program
management improvement efforts to determine how much progress is being
made;
* complete a best practices guide for program management; and:
* identify, train, and certify all program managers in accordance with
NNSA's program management policy.
With respect to financial management, we recommend that NNSA take the
following five actions:
* complete all NNSA PPBE policy guidance;
* establish an independent budget analysis unit;
* develop criteria for compiling a NNSA-wide priorities list to assist
NNSA in supporting its budget request;
* update the schedule of those NNSA programs requiring budget
validation to insure that recent program initiatives like the Reliable
Replacement Warhead are included; and:
* establish a DOE-compliant budget validation process.
Agency Comments and Our Evaluation:
We provided NNSA with a copy of this report for their review and
comment. NNSA generally agreed with the report and its corresponding
recommendations. NNSA noted that it considers the agency to be a
success but acknowledged that there was considerable work yet to be
accomplished. NNSA and DOE also provided technical comments, which we
have incorporated in this report as appropriate. NNSA's comments on our
draft report are presented in appendix I.
As arranged with your office, unless you publicly announce its contents
earlier, we plan no further distribution of this report until 30 days
after the date of this report. We are sending copies of this report to
appropriate congressional committees, the Secretary of Energy; the
Administrator, NNSA; and the Director of the Office of Management and
Budget. We will also make copies available to others upon request. In
addition, the report will be available at no charge on the GAO Web site
at [Hyperlink, http://www.gao.gov].
If you or your staff have any questions about this report or need
additional information, please contact me at (202) 512-3481 or
aloiseg@gao.gov. Contact points for our Offices of Congressional
Relations and Public Affairs may be found on the last page of this
report. GAO staff who made major contributions to this report are
listed in appendix II.
Signed by:
Gene Aloise:
Director, Natural Resources and Environment:
[End of section]
Appendix I Comments from the National Nuclear Security Administration:
Department of Energy:
National Nuclear Security Administration:
Washington, DC 20585:
December 21, 2006:
Mr. Gene Aloise:
Director, Natural Resources And Environment:
U.S. Government Accountability Office:
Washington, DC 20548:
Dear Mr. Aloise:
The National Nuclear Security Administration (NNSA) appreciates the
opportunity to review the Government Accountability Office's (GAO)
draft report: GAO-07-36, "National Nuclear Security Administration:
Additional Actions Needed to Improve Management of the Nation's Nuclear
Programs." We understand that the House's Chairman and Ranking Minority
Member, Strategic Forces Subcommittee, Committee on Armed Services
asked GAO to review the steps NNSA has taken to improve security at its
laboratories and plants and to improve its management practices and
revise its organizational structure.
NNSA generally agrees with the report and its corresponding
recommendations. We note that we consider NNSA a success but we
acknowledge that there is considerable work yet to be accomplished.
NNSA will provide detailed responses to the recommendations during the
Management Decision phase after the Final Report has been issued. In
the interim, I have attached a copy of the draft report with annotated
comments for edification and clarification. Should you have any
questions about this response, please contact Richard Speidel,
Director, Policy and Internal Controls Management.
Sincerely,
Signed by:
Michael C. Kane:
Associate Administrator for Management and Administration:
Attachment:
cc: Chief Operating Officer:
Director, Service Center:
Senior Procurement Executive:
[End of section]
Appendix II: GAO Contact and Staff Acknowledgments:
GAO Contact:
Gene Aloise (202) 512-3841:
Staff Acknowledgments:
In addition to the individual named above, James Noel, Assistant
Director; Robert Baney; Preston Heard; Jonathan Ban; Rebecca Shea;
Cynthia Daffron; Greg Marchand; and Carol Herrnstadt Shulman made key
contributions to this report.
(360556):
FOOTNOTES
[1] GAO, National Nuclear Security Administration: Key Management
Structure and Workforce Planning Issues Remain as NNSA Conducts
Downsizing, GAO-04-545, (Washington, D.C.: June 25, 2004).
[2] A special investigative panel, President's Foreign Intelligence
Advisory Board, Science at its Best, Security at its Worst: A Report on
Security Problems at the U.S. Department of Energy (Washington, D.C.:
June 1999).
[3] Pub. L. No. 106-65, 113 Stat. 512, 953 (1999).
[4] The Office of Naval Reactors is managed as a separate entity within
NNSA reporting to both NNSA and the U.S. Navy.
[5] GAO, Nuclear Security: NNSA Needs to Better Manage Its Safeguards
and Security Program, GAO-03-471 (Washington, D.C.: May 30, 2003).
[6] See Strengthening NNSA Security Expertise: An Independent Analysis,
Henry G. Chiles et al., March 2004 and NNSA Security: An Independent
Review, LMI Government Consulting, April 2005.
[7] See, for example, GAO, National Ignition Facility: Management and
Oversight Failures Caused Major Cost Overruns and Schedule Delays, GAO/
RCED-00-271 (Washington, D.C.: Aug. 8, 2000).
[8] See NNSA Security, An Independent Review, LMI Government
Consulting, Apr. 2005. This effort was directed by Admiral Richard W.
Mies USN (Retired).
[9] House Armed Services Committee Special Panel on the Department of
Energy Reorganization, 106th Cong., Department of Energy National
Nuclear Security Administration Implementation Plan (Feb. 2000); House
Armed Services Committee Special Panel on the Department of Energy
Reorganization, 106th Cong., Establishing the National Nuclear Security
Administration: A Year of Obstacles and Opportunities (Oct. 13, 2000).
[10] Pub. L. No. 109-364, § 3117 (2006).
[11] National Nuclear Security Administration's Planning, Programming,
Budgeting, and Evaluation Process, DOE/IG-0614, Aug. 2003.
[12] Institute for Defense Analysis, The Organization and Management of
the Nuclear Weapons Program (Washington, D.C.: Mar. 1997).
[13] A special investigative panel, President's Foreign Intelligence
Advisory Board, Science At Its Best, Security At Its Worst: A Report on
Security Problems at the U.S. Department of Energy (Washington, D.C.:
June 1999).
[14] See GAO, High-Risk Series: An Update, GAO-01-263 (Washington,
D.C.: Jan. 2001), and GAO, High Risk Series: An Update, GAO-03-119
(Washington, D.C.: Jan. 2003). Also, see GAO, Performance and
Accountability Series--Major Management Challenges and Program Risks: A
Governmentwide Perspective, GAO-01-241 (Washington, D.C.: Jan. 2001).
In addition, see the accompanying 21 reports (numbered GAO-01-242
through GAO-01-262) on specific agencies.
[15] GAO, Human Capital: Key Principles for Effective Strategic
Workforce Planning, GAO-04-39 (Washington, D.C.: Dec. 11, 2003).
[16] GAO, Department of Energy: Views on the Progress of the National
Nuclear Security Administration in Implementing Title 32, GAO-01-602T
(Washington, D.C.: Apr. 4, 2001).
[17] GAO, NNSA Management: Progress in the Implementation of Title 32,
GAO-02-93R (Washington, D.C.: Dec. 12, 2001).
[18] GAO, Department of Energy: NNSA Restructuring and Progress in
Implementing Title 32, GAO-02-451T (Washington, D.C.: Feb. 26, 2002).
[19] GAO-03-471.
[20] GAO-04-545.
[21] DOE, Secretary of Energy Advisory Board, Report of the Nuclear
Weapons Complex Infrastructure Task Force: Recommendations for the
Nuclear Weapons Complex of the Future (Washington, D.C.: July 13,
2005).
[22] Physical security combines security equipment, personnel, and
procedures to protect facilities, information, documents or material
against theft, sabotage, diversion, or other criminal acts. For the
purposes of analyzing trends in NNSA's physical security ratings we
combined the ratings for the following topical areas: Protection
Program Management; Physical Security Systems; Protective Force;
Classified Matter Protection and Control, Materials Control and
Accountability; and Personnel Security.
[23] GAO, Nuclear Security: Improvements Needed in DOE's Safeguards and
Security Oversight, GAO/RCED-00-62 (Washington, D.C.: Feb. 24, 2000).
[24] Inspection of Allegations Relating to the Albuquerque Operations
Office Security Survey Process and the Security Operations' Self-
Assessment at Los Alamos National Laboratory, DOE/IG-0471, May 2000.
[25] Review of the National Nuclear Security Administration's Federal
Line Management Oversight of Security Operations, Office of Independent
Oversight and Performance Assurance, DOE, 2005.
[26] GAO, Standards for Internal Control in the Federal Government,
GAO/AIMD-00-21.3.1 (Washington, D.C.: Nov. 1999).
[27] The Design Basis Threat is a classified document that identifies
the potential size and capabilities of a terrorist force that DOE
protective forces must be able to defeat.
[28] Section 3232 of the NNSA Act requires that the Chief of DNS report
to the Administrator and have direct access to the Secretary and every
departmental official on security matters. See 50 U.S.C. § 2422.
[29] GAO-03-471.
[30] GAO-04-545.
[31] Strengthening NNSA Security Expertise: An Independent Analysis,
Henry G. Chiles et al., Mar. 2004.
[32] Inspection Report: Security and Other Issues Related to Out-
Processing of Employees at Los Alamos National Laboratory, DOE/IG-0677
(Washington, D.C., Feb. 22, 2005).
[33] Inspection Report: Security Clearance Terminations and Badge
Retrieval at the Lawrence Livermore National Laboratory, DOE/IG-0716
(Washington, D.C., Jan. 19, 2006) and Inspection Report: Badge
Retrieval and Security Clearance Termination at Sandia National
Laboratory--New Mexico, DOE/IG-0724 (Washington, D.C., Apr. 18, 2006).
[34] GAO/AIMD-00-21.3.1.
[35] 50 U.S.C. § 2401 note.
[36] GAO-01-602T.
[37] Pub. L. No. 106-398, § 3157, 114 Stat. 1654, 1654A-468 (2000)
(codified as amended at 50 U.S.C. § 2410).
[38] In commenting on our draft report, DOE's Office of General Counsel
indicated that it believed the prohibitions contained in the NNSA Act
regarding "authority, direction, and control" prevented DOE and NNSA
from better defining their working relationship. Sections 3202 and 3220
of the NNSA Act shield NNSA personnel, except for the NNSA
Administrator and NNSA intelligence and counterintelligence personnel,
from the authority, direction, and control of DOE personnel. See 42
U.S.C. § 7132, 50 U.S.C. § 2410. The Administrator is subject to the
authority, direction, and control of the Secretary or Deputy Secretary
of Energy, who cannot redelegate those powers. NNSA intelligence and
counterintelligence personnel were also recently subjected to DOE
authority, direction, and control, as we describe later in this report.
In the view of DOE's Office of General Counsel, this statutory
structure is the reason for ongoing coordination problems, regardless
of whether formal procedures have been put in place between DOE and
NNSA offices.
[39] Section 3262 of the NNSA Act requires the Administrator to
establish procedures to comply with Federal Acquisition Regulations.
See 50 U.S.C. § 2462.
[40] See 50 U.S.C. § 2402(d).
[41] Pub. L. No. 109-364, § 3117 (2006).
[42] H.R. Conf. Rep. No. 109-702 at 969-971 (Sep. 29, 2006).
[43] Pub. L. 106-398, § 3157, 114 Stat. 1654, 1654A-468 (2000)
(codified as amended at 50 U.S.C. § 2410).
[44] NNSA Security: An Independent Review, April 2005.
[45] GAO-02-93R.
[46] GAO-04-545.
[47] GAO, Nuclear Weapons: Opportunities Exist to Improve the
Budgeting, Cost Accounting, and Management Associated with the
Stockpile Life Extension Program, GAO-03-583 (Washington, D.C.: July
28, 2003).
[48] Measuring Performance and Benchmarking Project Management at the
Department of Energy, National Research Council of the National
Academies, 2005.
[49] Improving Project Management in the Department of Energy, National
Research Council of the National Academies, 1999.
[50] Independent Research Assessment of Project Management Factors
Affecting Department of Energy Project Success, Civil Engineering
Research Foundation, July 12, 2004.
[51] Formally termed the NNSA Program Management Workload Reduction
Team, the team's charter was to arrive at a realistic, uniform, and
timely approach for optimizing and implementing Federal management
processes across NNSA.
[52] NNSA's Office of Defense Programs has issued a best practices
document to assist its program managers in the preparation of program
plans.
[53] Two NNSA program management improvement teams, one established in
2002 and the other in 2003, recommended training program managers. For
example, the 2003 team noted that there was a finite set of officially
designated NNSA program managers who needed to be trained and certified
to a minimum set of qualification standards and recommended an
implementation schedule to have all NNSA program managers named by
January 1, 2004, and a policy setting out program management
qualification standards implemented across NNSA by October 1, 2004.
[54] 50 U.S.C. § 2452.
[55] 50 U.S.C. § 2453.
[56] Audit Report: National Nuclear Security Administration's Planning,
Programming, Budgeting, and Evaluation Process, DOE/IG-0614, Aug. 5,
2003.
[57] GAO-03-583.
[58] The Nonproliferation and Verification Research and Development
Program's mission is to conduct needs-driven research, development,
testing, and evaluation of new technologies that are intended to
strengthen the United States' ability to prevent and respond to
nuclear, chemical, and biological attacks. The Highly-Enriched Uranium
Transparency Implementation Program's mission is to ensure that the
nonproliferation goals of the February 1993 agreement on the U.S.
purchase of uranium from the Russian Federation are met.
[59] GAO-03-583.
[60] According to NNSA Policy Letter NAP-1 dated May 21, 2002, DOE
orders are applicable to NNSA unless or until a NNSA policy letter is
provided.
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: