Defense Acquisitions
Space System Acquisition Risks and Keys to Addressing Them
Gao ID: GAO-06-776R June 1, 2006
On April 6, 2006, we testified before Congress on the Department of Defense's (DOD) space acquisitions. In fiscal year 2007, DOD expects to spend nearly $7 billion to acquire space-based capabilities to support current military and other government operations as well as to enable DOD to transform the way it collects and disseminates information, gathers data on its adversaries, and attacks targets. Despite its growing investment in space, however, DOD's space system acquisitions have experienced problems over the past several decades that have driven up costs by hundreds of millions, even billions, of dollars; stretched schedules by years; and increased performance risks. In some cases, capabilities have not been delivered to the warfighter after decades of development. Within this context, Congress requested that we provide additional comments regarding the need for better program management, space acquisition policy, and DOD's Space Radar and Transformational Satellite Communications System acquisitions.
GAO provided information on the top obstacles to achieving program success from the point of view of program mangers. We found that the top obstacles are funding instability, requirements instability, staffing problems, excessive oversight and inexperienced leadership. Estimated costs have been high, and grown, for DOD's Space Based Infrared System (SBIRS)-High, the Evolved Expendable Launch Vehicle program, the Advanced Extremely High Frequency Satellite (AEHF) Program, the National Polar-orbiting Operational Environmental Satellite System (NPOESS), and the Space Based Infrared System-Low program also had high estimated costs which increased. Additionally, the Space Based Infrared System-Low program and SBIRS-High both overpromised capabilities. DOD has been taking actions to improve cost estimating and we are in the process of addressing these actions. To address the problem of low levels of technological maturity, DOD has committed to delay the development of one new major space program--the Transformational Satellite Communications System (TSAT)--until technology needs are better understood. It has also committed to deliver new space-based capabilities in an incremental fashion so that acquisition efforts can be more executable and the science and technology base can be more engaged in major space programs. DOD has also faced issues regarding the addition of new requirements well into the acquisition phase. Our past reports have pointed to requirements setting problems in the AEHF, NPOESS, and SBIRS-High programs, and noted that DOD could take further steps to strengthen requirements setting by implementing processes and policies, as needed, which stabilize requirements for acquisitions. GAO and DOD have disagreed on what Technological Readiness Levels (TRL) should beat major decision points for space system acquisitions, and will continue to disagree as long as GAO continues to base its reviews of space programs on best practices and DOD continues to use the wide leeway afforded regarding critical technologies and their maturity levels to initiate product development. We identified the main difference between TRL 6 and 7 as the testing environment. For TRL 6, the testing environment would be a laboratory or a simulated operational environment, and for TRL 7, the testing environment would be an operational environment. Achieving TRL 6 or 7 by the critical design review (CDR) is a matter of risk--if the critical technologies in question are supremely important and have no space-based heritage, then it is warranted to test the technologies in space before proceeding through CDR. To ensure its integration efforts are successful, the TSAT program is planning to demonstrate critical technologies at TRL 6 when key integration tests are conducted in fiscal year 2007, use the results of its first round of integration tests to refine testing during the second, more comprehensive round, conduct a series of independent tests to verify results of contractor testing, and assess the results of the main integration tests before making a decision to enter the production development phase. According to GAO's prior work on best practices, to ensure success integration, programs mangers should ensure that (1) the right validation events occur at the right times, (2) each validation event produces quality results, and (3) the knowledge gained from an event is used to improve the product. Finally, the program manager needs assurance that all testing that has been done is reflective of the capabilities that the program is trying to deliver.
GAO-06-776R, Defense Acquisitions: Space System Acquisition Risks and Keys to Addressing Them
This is the accessible text file for GAO report number GAO-06-776R
entitled 'Defense Acquisitions: Space System Acquisition Risks and Keys
to Addressing Them' which was released on June 2, 2006.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
June 1, 2006:
The Honorable Jeff Sessions:
Chairman:
The Honorable Bill Nelson:
Ranking Minority Member:
Subcommittee on Strategic Forces:
Committee on Armed Services:
United States Senate:
Subject: Defense Acquisitions: Space System Acquisition Risks and Keys
to Addressing Them:
On April 6, 2006, we testified before the subcommittee on the
Department of Defense's (DOD) space acquisitions. In fiscal year 2007,
DOD expects to spend nearly $7 billion to acquire space-based
capabilities to support current military and other government
operations as well as to enable DOD to transform the way it collects
and disseminates information, gathers data on its adversaries, and
attacks targets. Despite its growing investment in space, however,
DOD's space system acquisitions have experienced problems over the past
several decades that have driven up costs by hundreds of millions, even
billions, of dollars; stretched schedules by years; and increased
performance risks. In some cases, capabilities have not been delivered
to the warfighter after decades of development.
Within this context, you requested that we provide additional comments
regarding the need for better program management, space acquisition
policy, and DOD's Space Radar and Transformational Satellite
Communications System acquisitions. Your specific questions and our
answers are discussed below.
Question: What are the top obstacles to achieving program success from
the point of view of program managers?
As part of a 2005 review[Footnote 1] on program management best
practices, we surveyed DOD's major weapon program managers, including
some managing space programs, who cited the following as "top"
obstacles to achieving successful outcomes in an open ended question:
² funding instability (about 36 percent),
² requirements instability (13 percent),
² staffing problems (8 percent),
² excessive oversight (7 percent), and:
² inexperienced leadership (7 percent).
Although the majority of respondents to our survey believed that the
initial baselines of their programs were reasonable, a significant
group, about 24 percent, responded that their program parameters were
not reasonable at the start, and 45 program managers responded that
their program had been rebaselined one or more times for cost and
schedule increases. In addition, 18 percent said one or more key
technologies fell below best practice standards for maturity.
Our reviews of space programs are consistent with these views--we have
found technologies to be immature at program start for major space
programs. Further, in delving deeper into the root causes behind these
problems, we have found that competition for funding has incentivized
programs to produce optimistic cost and schedule estimates, over
promise on capability, suppress bad news, and forsake the opportunity
to identify potentially better alternatives. In addition, because DOD
starts more weapons programs than it can afford, it invariably finds
itself in the position of having to shift funds to sustain programs--
often to the point of undermining well-performing programs to pay for
poorly performing ones. We also have found that DOD starts its space
programs too early, that is, before it has assurance that the
capabilities it is pursuing can be achieved within available resources
(time, money technology, people, etc.) and time constraints, and it
allows new requirements to be added well into the acquisition phase, a
course of action that can further stretch technology challenges. This
is encouraged by the funding process, as acquisition programs tend to
attract the majority of research, development, test, and evaluation
(RDT&E) dollars. Many officials working within the space community
agreed that these were key underlying causes of acquisition problems
during a review we conducted last year.[Footnote 2] In addition,
officials we spoke with also cited pressures resulting from having a
diverse array of officials and organizations involved with the space
acquisition process, tensions between the science and technology (S&T)
and acquisition communities as to who is better suited to translate
technology concepts into reality, pressures resulting from short
tenures among staff critical to achieving acquisition success, and
difficulties in overseeing contractors.
Question: Do you believe that the Air Force is addressing these
obstacles?
The Air Force has recently taken steps to put is Transformational
Satellite Communications System (TSAT) program on a more executable
track by reducing its expectations in the level of sophistication for
the first two satellites so that it can meet its schedule goals. It is
also holding off entering product development of the first increment
until critical technologies are proven. If the Air Force adheres to
this commitment for TSAT and applies it to Space Radar, as it has also
informally committed to do, then it would be addressing some of the
obstacles noted above. For example, it would reduce the risk of funding
instability since cost estimates would be more realistic. In addition,
the Air Force has committed to estimating cost and funding new
acquisitions to an 80-percent confidence level, strengthening systems
engineering, and strengthening the acquisition workforce. And for some
specific programs, the Air Force has applied additional mechanisms to
regulate requirements. These actions could also remove obstacles, if
effectively implemented.
However, as we testified, such actions should be accompanied by an
investment strategy for space, and ultimately DOD's entire weapons
portfolio, to separate wants from needs and to alleviate long-standing
pressures associated with competition within DOD to win funding. DOD
could also instill the best practices it is now embracing into its
space acquisition policy. In addition, we have recommended that DOD, as
a whole, take steps to hold people and programs accountable when best
practices are not pursued. This will require DOD to empower program
managers to make decisions related to funding, staffing, and moving
into subsequent phases and to match program manager tenure with
delivery of a product. It may also require DOD to tailor career paths
and performance management systems to provide incentives for longer
tenures. By embracing a model that incorporates all these elements, DOD
can achieve better outcomes for its space programs. By not doing so,
there will still be incentives and allowances to overpromise
capability, underestimate cost and schedule, and to start programs
prematurely, which, in turn, can eventually undo other improvement
efforts.
Question: DOD starts more space and weapons programs than it can
afford, which, according to GAO, "pressures programs to underestimate
costs and over promise capabilities." Can you provide a few examples of
this problem in space programs and [say] if and how the problem is
being addressed?
Actual costs for nearly every major space acquisition we review each
year as part of our annual weapon system assessment have greatly
exceeded earlier estimates--a clear indication that programs
consistently underestimate costs. For example, the Space Based Infrared
System (SBIRS)-High cost estimate climbed from about $4 billion as of
October 1996 to over $10 billion in September 2005, and costs are
expected to increase further. Estimated costs for the Evolved
Expendable Launch Vehicle (EELV) program have climbed from about $15
billion in October 1998 to $27 billion in August 2005 with 43 fewer
launches to be purchased than anticipated. Estimated costs for the
Advanced Extremely High Frequency Satellite program (AEHF) increased
from $5.6 billion as of October 2001 to $6.2 billion as of August 2005,
with quantities decreasing from five to three satellites. Estimated
costs for the National Polar-orbiting Operational Environmental
Satellite System (NPOESS) grew from $5.9 billion in August 2002 to
nearly $8 billion in September 2005. Our past reports have also
identified cases where programs have overpromised capabilities. For
example, the Space Based Infrared System (SBIRS)-Low program started
under the assumption that the satellites would be able to detect and
track multiple objects and differentiate a threatening warhead from
decoys, even though that technology challenge was exceedingly high. In
fact, the program was never able to achieve this capability. It was
eventually shut down in the face of cost and schedule overruns that
came with addressing technology challenges. The SBIRS-High program
began with the assumption that there would be four satellites in
geosynchronous orbit, but more than 10 years later, DOD plans to reduce
the number of satellites it will procure and still does not have the
assurance it needs that the missile detection capability can be
achieved in time to replace the existing detection system. In addition,
DOD has initiated efforts to develop a parallel competing capability
with the SBIRS-High program. Similarly, the NPOESS program is now
considering dropping some of its planned capability because of
technology and design-related challenges.
DOD has been taking actions to improve cost estimating and we are in
the process of assessing these actions. As mentioned above, for
example, it has committed to estimating cost and funding new
acquisitions to an 80-percent confidence level. In addition, the Air
Force is requiring the use of independent cost estimates--rather than
estimates produced by a program office or a contractor. It is also
committed to strengthening its cost-estimating capabilities--in terms
of people, methodologies, and tools. In regard to the issue of
overpromising capability, the Air Force has deferred pursuing some of
its more ambitious capabilities on its TSAT program, so that the
program can be better positioned to meet its schedule. We do not know
at this point whether it will be doing the same for its new Space Radar
program. As we underscored in our testimony, it is important that these
and other individual actions be made within a framework of broader,
systemic improvements to DOD's overall acquisition process, the
acquisition workforce, and an overall investment strategy.
Question: The second problem is that DOD "starts its space programs too
early, that is, before it is sure that the capabilities it is pursuing
can be achieved within available resources and time constraints." Can
you provide a few examples of this problem in space programs and
[indicate] if and how the problem is being addressed?
Many of our annual reviews of major space acquisitions show that
programs have started with relatively low levels of technology
maturity--meaning DOD does not have assurance that the technologies can
work as intended. This includes, AEHF, NPOESS, SBIRS-High, and SBIRS-
Low--now known as the Space Tracking and Surveillance System.
Exceptions include the Navy's Mobile User Objective System, or MUOS
(though the program later added two additional technologies that did
not meet best practices standards for maturity) and the Global
Positioning System Block IIF. At times, we have found that key sensors
to be included in new satellites were not fully tested, or even
prototyped, before being included in a program. In other cases,
technologies used to support the health of the overall satellite, such
as cooling systems, were immature. And in other cases, software needs
were vastly underestimated. In the case of AEHF, technical resources to
support security needs were underestimated.
Many programs we have studied felt the need to start the acquisition
process before such needs were better understood because acquisition
programs tend to attract more funding than science and technology
efforts. In addition, in the case of space, programs have historically
attempted to satisfy all requirements in a single step, regardless of
the design challenge or the maturity of the technologies to achieve the
full capability. While this is partly attributable to a desire to speed
delivery of capability, it has perversely slowed down programs, since
programs were at increased risk of facing costly and disruptive
technical and design problems.
As noted previously, DOD has committed to delay the development of one
new major space program--TSAT--until technology needs are better
understood. It has also committed to deliver new space-based
capabilities in an incremental fashion so that acquisition efforts can
be more executable and the science and technology base can be more
engaged in major space programs. It has not taken such action yet on
other new programs, notably Space Radar, though it has informally
committed to. In addition, DOD's space acquisition policy still allows
major acquisitions to begin without demonstrating that technology can
work as intended.
Question: A third issue is that DOD has "allowed new requirements to be
added well into the acquisition phase." I would also add that sometimes
the original requirements may be unrealistic or unaffordable and that
this may be part of the problem. Can you provide a few examples of the
requirements problem in space programs and [indicate] if and how the
problem is being addressed?
Our past reports have pointed to requirements setting problems in the
AEHF, NPOESS, and SBIRS-High programs. In the case of SBIRS-High, we
pointed to problems related to not adequately defining requirements up
front. These were further detailed in subsequent DOD studies, including
those by the SBIRS-High Independent Review Team and the Defense Science
Board. Both noted that the acquisition approach the Air Force was
following, known as Total System Performance Responsibility, placed too
much responsibility on the part of the contractor to negotiate
requirements, and that the process eventually broke down. In the case
of NPOESS, we reported in the early phases of the program that the Air
Force and the National Oceanic and Atmospheric Administration had
difficulty resolving diverging requirements. In the case of AEHF, we
reported that DOD substantially and frequently altered requirements and
design in the early phases of the program. While considered necessary,
some changes increased costs by hundreds of millions of dollars and
caused scheduling delays on a program that DOD was trying to accelerate
in order to address a potential capability gap. DOD has since rejected
the acquisition approaches that led to requirements-setting problems on
both SBIRS-High and AEHF. It has also instituted control mechanisms to
regulate requirements on SBIRS-High. In our testimony, we noted that
DOD could take further steps to strengthen requirements setting by
implementing processes and policies, as needed, which stabilize
requirements for acquisitions, like NPOESS, that are being shared with
other agencies.
We have also reported on programs that took on unrealistic or
potentially unaffordable requirements. The SBIRS-Low program's pursuit
of discrimination capability is an older example of such a program.
More recently, we have pointed to affordability and feasibility issues
related to Space Radar and the TSAT programs, which together, have been
preliminarily estimated to cost about $40 billion. Specifically, we
have stated that DOD was planning to start these acquisitions even when
many of their critical technologies were still immature, and it was
pursuing a highly ambitious path in terms of the technology push. Given
that these systems were among the most complex programs ever undertaken
for space, they were being counted on to enable wider DOD
transformation efforts, and DOD was already contending with highly
problematic space efforts, we believed that DOD could not afford to
pursue such risky approaches for TSAT and Space Radar. As noted
earlier, DOD has taken steps to ensure it is pursuing realistic
requirements for TSAT, and it has informally committed to do the same
for Space Radar.
Question: Is there a clear definition of each Technology Readiness
Level (TRL) that all of you agree on (GAO and DOD) and that exists in
writing and that clearly applies to space programs?
The National Aeronautics and Space Administration (NASA) developed the
original ranking and definitions of technology maturity levels. GAO and
DOD agree on the TRL definitions--in its reports, GAO continues to
reference the TRL scale for assessing critical technologies from DOD's
Interim Defense Acquisition Guidebook (app 6, dated October 30, 2002).
However, for space system acquisitions, GAO and DOD have disagreements
on what the TRLs should be at major decision points. According to our
work on best practices, product development should be initiated after
critical technologies have been incorporated into a system prototype
and tested in an operational environment--meaning the cold-radiated
vacuum of space. Our prior reports have recognized that space systems
are uniquely difficult to test in a true operational environment.
However, DOD has found ways to test sensors and other critical
technologies on experimental satellites. Nonetheless, DOD continues to
stand up formal space system acquisitions too early--before critical
technologies have been tested in operational or relevant environments-
-that is, before DOD has assurance that the capabilities it is pursuing
can be achieved. This causes DOD to extend technology invention to its
acquisitions, which have reverberating effects and require large
amounts of time and money to fix. In these cases, DOD points to its
National Security Space Acquisition Policy, which allows it to take
such an approach--unlike DOD's acquisition policy for non-space
acquisitions, where TRL 7 (testing in an operational environment) is
preferred before product development is initiated (TRL 6 is required).
As long as GAO continues to base its reviews of space programs on best
practices and DOD continues to use the wide leeway afforded in its
space acquisition policy regarding critical technologies and their
maturity levels to initiate product development, GAO and DOD will
continue to have disagreements in this area.
Question: What is the difference between TRL 6 and 7 and what is the
advantage or disadvantage of being at level 6 or 7 at the [Critical]
Design Review?
The main difference between TRL 6 and 7 is the testing environment. For
TRL 6, the testing environment would be a laboratory or a simulated
operational environment, and for TRL 7, the testing environment would
be an operational environment--meaning in space. According to GAO's
work on best practices, achieving a high level of technology maturity
at program start is an important indicator of whether available
resources in terms of knowledge, time, money and capacity match the
customer's requirements. In addition, the key measure for a successful
critical design review (CDR) is when 90-percent of the design drawings
have been submitted to manufacturing. When space programs reach CDR and
TRLs are below 6, it is unlikely that a high percentage of design
drawings would have been released to manufacturing, thereby increasing
program risk at this juncture. Another key point to remember is that
CDR is the point at which programs begin ordering long-lead parts to
build the first few satellites. This investment in hardware is at risk
if the technologies do not prove out to work as intended. Achieving TRL
6 or 7 by CDR is a matter of risk--if the critical technologies in
question are supremely important and have no space-based heritage, then
it is warranted to test the technologies in space before proceeding
through CDR. For TSAT, some critical technologies have a heritage of
being tested or operated in space, and they are all slated to be at TRL
6 at the time of CDR--an approach that GAO did not fault.
Question: The Transformational Communications Satellite program, though
still very early in the process, appears to have begun to adopt some of
the recommendations of the GAO as well as the Young Panel and is
focusing on technology maturity. Integration of the satellite appears
to be the next difficult step for the TSAT program. What plans are in
place to ensure successful integration?
The TSAT program is taking several steps to ensure its integration
efforts are successful. First, according to program officials, the plan
is to demonstrate critical technologies at TRL 6 when key integration
tests are conducted in fiscal year 2007. Second, the program plans to
use the results of its first round of integration tests to refine the
testing to be conducted during a second round of more comprehensive
integration testing. Third, the program is conducting a series of
independent tests to verify results of contractor testing as it
incrementally builds toward the two main integration tests facing the
program--tests of the Next Generation Processor Router and Optical
Standards Validation Suite. The program office plans to have knowledge
on how these two major subcomponents work to reduce risk by uncovering
technical problems before awarding the space segment contract for the
design and assembly of the satellites. Finally, the TSAT program also
plans to assess the results of the main integration tests before making
a decision to enter the production development phase.
Question: What actions would you recommend to the programs managers to
ensure successful integration?
According to GAO's prior work on best practices, leading firms ensure
that (1) the right validation events--tests, simulations, and other
means for demonstrating product maturity--occur at the right times, (2)
each validation event produces quality results, and (3) the knowledge
gained from an event is used to improve the product. Fully disclosing
the results of tests (from low-level brass board tests to the main
integration tests) and documenting the actions taken to address
shortcomings further validates product knowledge. It is imperative that
problems are fully addressed before rushing efforts to begin the next
round of testing. It is also important that program managers use the
test and evaluation parameters originally established, and any changes
should be fully disclosed along with the reasons for doing so. Finally,
the program manager needs assurance that all testing that has been done
is reflective of the capabilities that the program is trying to
deliver. Rigorous and sophisticated testing early and often will
uncover problems when they are relatively easy and inexpensive to fix.
Waiting too long to fully stress and test components will put the
program in a risky position.
In preparing answers to your questions, we relied on our prior work on
DOD's space acquisition policy, best practices in weapon system
acquisitions, and our reviews of specific space acquisitions as well as
DOD studies. In addition, for specific space systems development and
cost growth, we relied on our annual assessment of selected major
weapon programs. Because we relied on previously issued work, we did
not obtain comments from DOD on a draft of this letter. We conducted
our work from April 2006 through May 2006 in accordance with generally
accepted government auditing standards.
We are sending copies of this letter to the Secretaries of Defense and
the Air Force and interested congressional committees. We will also
make copies available to others upon request. In addition, the report
will be available at no charge on the GAO web site at [Hyperlink,
http://www.gao.gov].
If you or your staff have any questions concerning these comments,
please contact me at (202) 512-4841.
Sincerely yours,
Signed by:
Cristina Chaplain:
Acting Director:
Acquisition and Sourcing Management:
cc:
(120561):
FOOTNOTES
[1] GAO, Best Practices: Better Support of Weapon System Program
Managers Needed to Improve Outcomes, GAO-06-110 (Washington, D.C.: Nov.
30, 2005).
[2] GAO, Defense Acquisitions: Incentives and Pressures That Drive
Problems Affecting Satellite and Related Acquisitions, GAO-05-570R
(Washington, D.C.: June 23, 2005).
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: