Defense Acquisitions
Charting a Course for Improved Missile Defense Testing
Gao ID: GAO-09-403T February 25, 2009
The Missile Defense Agency (MDA) has spent about $56 billion and will spend about $50 billion more through 2013 to develop a Ballistic Missile Defense System (BMDS). This testimony is based on two reviews GAO was directed to conduct in 2008. In addition to our annual review assessing the annual cost, testing, schedule, and performance progress MDA made in developing BMDS, we have also reported on MDA's targets program. In this testimony we discuss (1) the productivity of MDA's recent test program, (2) the consequences of the testing shortfalls, and (3) key factors that should be considered as MDA revises its approach to testing. GAO assessed contractor cost, schedule, and performance; tests completed; and the assets fielded during 2008. GAO also reviewed pertinent sections of the U.S. Code, acquisition policy, and the activities of a new missile defense board.
The scale, complexity, cost and safety associated with testing the missile defense system constitute a unique challenge for MDA, test agencies and other oversight organizations. This challenge is heightened by the fact that missile defense assets are developed, produced, and fielded concurrently. Overall, during fiscal year 2008, testing has been less productive than planned. While MDA completed several key tests that demonstrated enhanced performance of BMDS, all elements of the system had test delays and shortfalls, in part due to problems with the availability and performance of target missiles. GMD in particular was unable to conduct either of its two planned intercept attempts in fiscal year 2008. While it did subsequently conduct one in December 2008, it was not able to achieve all primary objectives because the target failed to release its countermeasures. As a result, aspects of the fielded ground-launched kill vehicles may not be demonstrated since no more flight tests have been approved. Target missiles continue as a persistent problem in fiscal year 2008 as poor target performance caused several tests to either fail in part or in whole. Testing shortfalls have had several consequences. First, they have delayed the validation of models and simulations, which are needed to assess the system's overall performance. As a result, the performance of the fielded BMDS as a whole cannot yet be determined. Second, the production and fielding of assets has continued and in some cases has gotten ahead of testing. For example, enhanced Exoatmospheric Kill Vehicles will now be produced and delivered before they are flight tested. Third, MDA has relied on a reduced basis--fewer test, model, and simulation results--to declare capabilities as operational in the field. MDA has undertaken a three-phase review of the entire BMDS test program that involves identifying critical variables that have not been proven to date, determining what test scenarios are needed to collect the data, and developing an affordable, prioritized schedule of flight and ground tests. This review, as long as it continues to involve test and evaluation organizations, appears to offer a sound approach for closing the gaps that exist between testing, modeling, and simulation. Critical to being able to implement the approach will be addressing the factors that have limited the productivity of the current test approach, such as the availability and performance of targets. An additional consideration in a new testing approach must be to ensure that assets are sufficiently tested before they are produced and fielded. An important consideration in this regard is for modeling, simulation, and testing events to be re-synchronized so that they properly inform decisions on producing, fielding, and declaring assets operational. Contingency plans could then be formed for adjusting the pace of these decisions should shortfalls occur in modeling, simulation, or testing. Because MDA has indicated implementation will take time, managing the transition may need to include reassessing the ambitious fiscal year 2009 test plan. In the mean time, MDA will have to be prudent in making decisions to produce and field assets.
GAO-09-403T, Defense Acquisitions: Charting a Course for Improved Missile Defense Testing
This is the accessible text file for GAO report number GAO-09-403T
entitled 'Defense Acquisitions: Charting a Course for Improved Missile
Defense Testing' which was released on February 25, 2009.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Testimony:
Before the Subcommittee on Strategic Forces, Committee on Armed
Services, House of Representatives:
United States Government Accountability Office:
GAO:
For Release on Delivery:
Expected at 1:00 p.m. EDT:
Wednesday, February 25, 2009:
Defense Acquisitions:
Charting a Course for Improved Missile Defense Testing:
Statement of Paul Francis:
Director, Acquisition and Sourcing Management:
GAO-09-403T:
GAO Highlights:
Highlights of GAO-09-403T, a testimony before the Subcommittee on
Strategic Forces, Committee on Armed Services, House of
Representatives.
Why GAO Did This Study:
The Missile Defense Agency (MDA) has spent about $56 billion and will
spend about $50 billion more through 2013 to develop a Ballistic
Missile Defense System (BMDS). This testimony is based on two reviews
GAO was directed to conduct in 2008. In addition to our annual review
assessing the annual cost, testing, schedule, and performance progress
MDA made in developing BMDS, we have also reported on MDA‘s targets
program. In this testimony we discuss (1) the productivity of MDA‘s
recent test program, (2) the consequences of the testing shortfalls,
and (3) key factors that should be considered as MDA revises its
approach to testing.
GAO assessed contractor cost, schedule, and performance; tests
completed; and the assets fielded during 2008. GAO also reviewed
pertinent sections of the U.S. Code, acquisition policy, and the
activities of a new missile defense board.
What GAO Found:
The scale, complexity, cost and safety associated with testing the
missile defense system constitute a unique challenge for MDA, test
agencies and other oversight organizations. This challenge is
heightened by the fact that missile defense assets are developed,
produced, and fielded concurrently. Overall, during fiscal year 2008,
testing has been less productive than planned. While MDA completed
several key tests that demonstrated enhanced performance of BMDS, all
elements of the system had test delays and shortfalls, in part due to
problems with the availability and performance of target missiles. GMD
in particular was unable to conduct either of its two planned intercept
attempts in fiscal year 2008. While it did subsequently conduct one in
December 2008, it was not able to achieve all primary objectives
because the target failed to release its countermeasures. As a result,
aspects of the fielded ground-launched kill vehicles may not be
demonstrated since no more flight tests have been approved. Target
missiles continue as a persistent problem in fiscal year 2008 as poor
target performance caused several tests to either fail in part or in
whole.
Testing shortfalls have had several consequences. First, they have
delayed the validation of models and simulations, which are needed to
assess the system‘s overall performance. As a result, the performance
of the fielded BMDS as a whole cannot yet be determined. Second, the
production and fielding of assets has continued and in some cases has
gotten ahead of testing. For example, enhanced Exoatmospheric Kill
Vehicles will now be produced and delivered before they are flight
tested. Third, MDA has relied on a reduced basis”fewer test, model, and
simulation results”to declare capabilities as operational in the field.
MDA has undertaken a three-phase review of the entire BMDS test program
that involves identifying critical variables that have not been proven
to date, determining what test scenarios are needed to collect the
data, and developing an affordable, prioritized schedule of flight and
ground tests. This review, as long as it continues to involve test and
evaluation organizations, appears to offer a sound approach for closing
the gaps that exist between testing, modeling, and simulation. Critical
to being able to implement the approach will be addressing the factors
that have limited the productivity of the current test approach, such
as the availability and performance of targets. An additional
consideration in a new testing approach must be to ensure that assets
are sufficiently tested before they are produced and fielded. An
important consideration in this regard is for modeling, simulation, and
testing events to be re-synchronized so that they properly inform
decisions on producing, fielding, and declaring assets operational.
Contingency plans could then be formed for adjusting the pace of these
decisions should shortfalls occur in modeling, simulation, or testing.
Because MDA has indicated implementation will take time, managing the
transition may need to include reassessing the ambitious fiscal year
2009 test plan. In the mean time, MDA will have to be prudent in making
decisions to produce and field assets.
What GAO Recommends:
We have previously made recommendations to improve the MDA‘s testing
and targets programs that include establishing a revised business case
for providing targets for a robust flight test program as well as
adding sufficient scope to tests to enable an assessment of the BMDS‘
suitability and effectiveness, but MDA only partially agreed. We also
have a draft report that is currently with DOD for comment that
includes additional recommendations regarding testing.
View [hyperlink, http://www.gao.gov/products/GAO-09-403T] or key
components. For more information, contact Paul Francis, 202-512-4841,
Francisp@gao.gov.
[End of section]
Madame Chairman and Members of the Subcommittee:
I am pleased to be here today to discuss the future of the Missile
Defense Agency's (MDA's) testing program.
MDA has been charged with developing and fielding the Ballistic Missile
Defense System (BMDS), a system expected to be capable of defending the
United States, deployed troops, friends, and allies against ballistic
missiles of all ranges in all phases of flight. In fulfilling this
charge, MDA placed an initial set of missile defense components in the
field in December 2005.
The National Defense Authorization Acts for fiscal years 2002, 2007 and
2008 mandated that we prepare annual assessments of MDA's ongoing cost,
schedule, testing, and performance progress. In March 2009, we plan to
issue our report covering MDA's progress toward achieving its goals
during fiscal year 2008 as well as its efforts to improve transparency,
accountability, and oversight. Additionally, in September 2008, we
issued a report on MDA's Target Program. My statement today will focus
on the testing-related issues covered in both reports. We conducted
these performance audits from February 2008 to February 2009 in
accordance with generally accepted government auditing standards. Those
standards require that we plan and perform the audit to obtain
sufficient, appropriate evidence to provide a reasonable basis for our
findings and conclusions based on our audit objectives. We believe that
the evidence obtained provides a reasonable basis for our findings and
conclusions based on our audit objectives.
Background:
The Missile Defense Agency's mission is to develop an integrated and
layered BMDS to defend the United States, its deployed forces, allies,
and friends. In order to meet this mission, MDA is developing a highly
complex system of systems--land, sea and space based sensors,
interceptors and battle management. Since its initiation in 2002, MDA
has been given a significant amount of flexibility in executing the
development and fielding of the BMDS. To enable MDA to field and
enhance a missile defense system quickly, the Secretary of Defense in
2002 delayed the entry of the BMDS program into the Department of
Defense's traditional acquisition process until a mature capability was
ready to be handed over to a military service for production and
operation. Therefore, the program concurrently develops, tests and
fields assets. This approach helped MDA rapidly deploy an initial
capability. On the other hand, because MDA can field assets before all
testing is completed, it has fielded some assets whose capability is
uncertain.
Because MDA develops and fields assets continuously, it combines
developmental testing with operational testing. In general,
developmental testing is aimed at determining whether the system design
will satisfy the desired capabilities; operational testing determines
whether the system is effective, survivable, and suitable in the hands
of the user. MDA conducts testing both on the ground and in flight. The
most complex of these is an end-to-end flight test that involves a test
of all phases of an engagement including detecting, tracking and
destroying a target with an interceptor missile. An end-to-end
intercept involves more than one MDA element. For example, a recent
intercept test involved a target flown out of Kodiak, Alaska, tracked
by the AN/TPY-2 radar located in Alaska, and the Beale upgraded early
warning radar located in California, the Sea-based X-band radar and an
Aegis radar located at different points in the Pacific. All of the
radars communicated with fire control centers in Alaska to guide an
interceptor launched from California to hit the target over the Pacific
Ocean.
Due to the complexity, scale, safety constraints, and cost involved,
MDA is unable to conduct a sufficient number of flight tests to fully
understand the performance of the system. Therefore, MDA utilizes
models and simulations, anchored by flight tests, to understand both
the developmental and operational performance of the system. To ensure
confidence in the accuracy of modeling and simulation the program goes
through a process called accreditation. The models are validated
individually using flight and other test data and accredited for their
intended use. Models and simulations are used prior to a flight test to
predict performance, the flight test is then run to gather data and
verify the models, and then data is analyzed after the flight and
reconstructed using the models and simulations to confirm their
accuracy.
MDA intends to group these models into system-level representations
according to user needs. One such grouping is the annual performance
assessment, a system-level end-to-end simulation that assesses the
performance of the BMDS configuration as it exists in the field. The
performance assessment integrates element-specific models into a
coherent representation of the BMDS. Fundamentally, performance
assessments anchored by flight tests are a comprehensive means to fully
understand the performance capabilities and limitations of the BMDS.
In addition to testing, modeling and simulation, and performance
assessments, MDA also has a formal process for determining when a newly
fielded asset or group of assets can be declared operational--that is,
cleared for use by the warfighter in operational situations. MDA uses a
variety of information as a basis to assess a new capability for
declaration. For example, MDA will define in advance tests, models, and
simulations it will use to base a specific decision on whether an asset
or capability can be declared ready for fielding. Each capability
designation so designated represents upgraded capacity to support the
overall function of BMDS in its mission as well as the level of MDA
confidence in the system's performance.
To assess testing related progress in fiscal year 2008, we examined the
accomplishments of ten BMDS elements that MDA is developing and
fielding. Our work included examining documents such as Program
Execution Reviews, test plans and reports, and production plans. We
also interviewed officials within each element program office and
within MDA functional directorates. In addition, we discussed each
element's test program and its results with DOD's Office of the
Director, Operational Test and Evaluation. We also interviewed
officials from the Office of the Under Secretary of Defense for
Acquisition, Technology, and Logistics.
Test, Targets and Performance Challenges Continue During Fiscal Year
2008:
MDA continues to experience difficulties achieving its goals for
testing. During fiscal year 2008, while several tests showed progress
in individual elements and some system level capabilities, all BMDS
elements experienced test delays or shortfalls. Most were unable to
accomplish all objectives and performance challenges continued for
many. Table 1 summarizes test results and target performance for the
BMDS elements during the year.
Table 1: Fiscal Year 2008 Test and Targets Issues:
Element: Airborne Laser;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: Yes;
Target Issues: N/A.
Element: Aegis Ballistic Missile Defense (BMD);
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: Target availability delayed key test from 2008 until at
least third quarter fiscal year 2009.
Element: Command, Control, Battle Management and Communications;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: N/A.
Element: Ground-based Midcourse Defense (GMD);
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: Target failed to release countermeasures during December
2008 flight test--FTG-05.[A]
Element: Kinetic Energy Interceptor;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: N/A.
Element: Multiple Kill Vehicle (MKV);
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No[B];
Target Issues: N/A.
Element: Sensors;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: Target failed to release countermeasures during July
2008 testing (FTX-03).
Element: Space Tracking and Surveillance System;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: N/A.
Element: Targets and Countermeasures;
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: Flexible Target Family delivery delayed and experienced
cost growth.
Element: Terminal High Altitude Area Defense (THAAD);
Tests/Activities Conducted as Scheduled: No;
All Objectives Achieved: No;
Target Issues: Target experienced anomaly during a September flight
test resulting in a no-test.
Source: GAO (presentation); MDA (data).
[A] This flight test was originally scheduled for fiscal year 2008, but
was later executed in fiscal year 2009.
[B] The MKV program was able to achieve this test objective in the
first quarter of fiscal year 2009.
[End of table]
Because of delays in flight test and a key ground test, MDA was unable
to achieve any of the six knowledge points the MDA Director had
scheduled for fiscal year 2008. In May 2007, the MDA Director
established key system-level and element-level knowledge points, each
based on an event that was to provide critical information--or
knowledge--for a decision requiring his approval. For example, two
knowledge points that MDA had to defer because of testing problems were
confirmation of a new target's performance and assessment of the SM-3
Block 1A missile's ability to engage and intercept a long range target.
GMD in particular continues to experience testing problems and delays.
Based on its September 2006 plan, MDA had expected to conduct 7 GMD
interceptor flight tests from the start of fiscal year 2007 through the
first quarter of fiscal year 2009. MDA however was only able to conduct
two, as shown in figure 1.
Figure 1: GMD Reduction in Flight Test from January 2006 to March 2010:
[Refer to PDF for image]
This figure is a chart depicting the following data:
GMD Reduction in Flight Test from January 2006 to March 2010:
As of September 2006:
Integrated flight tests planned:
Date: FY 2006, Q1;
Flight: FT-1, CE-I EKV (achieved);
Date: FY 2006, Q4;
Flight: FTG-2, CE-I EKV (achieved);
Date: FY 2007, Q1;
Flight: FTG-3, CE-I EKV;
Date: FY 2007, Q3;
Flight: FTG-4, CE-I EKV;
Date: FY 2007, Q4;
Flight: FTG-5, CE-I EKV;
Date: FY 2008, Q1;
Flight: FTG-6, CE-I EKV;
Date: FY 2008, Q2;
Flight: FTG-7, CE-I EKV;
Date: FY 2008, Q4;
Flight: FTG-8, CE-I EKV;
Date: FY 2009, Q1;
Flight: FTG-9, CE-I EKV;
Date: FY 2009, Q1;
Flight: FTG-9, CE-II EKV New processor.
As of January 2009:
Date: FY 2006, Q1;
Flight: FT-1, CE-I EKV (achieved);
Date: FY 2006, Q4;
Flight: FTG-2, CE-I EKV (achieved);
Date: FY 2007, Q4;
Flight: FTG-3a, CE-I EKV (achieved);
Date: FY 2009, Q2;
Flight: FTG-5, CE-I EKV (achieved);
Date: FY 2009, Q4;
Flight: FTG-6, CE-II EKV New processor.
Source: GAO analysis of GMD's flight test and interceptor fielding
schedule as of 2/25/07 and updated as of 1/31/09.
[End of figure]
GMD was unable to conduct either of its planned intercept attempts
during fiscal year 2008 - FTG-04 and FTG-05. MDA first delayed and then
later canceled the FTG-04 test in May 2008 due to a problem with a
telemetry component in the interceptor's Exoatmospheric Kill Vehicle.
The cancellation of FTG-04 removed an important opportunity to obtain
end-game performance data needed to develop GMD models and to verify
the capability of the fielded Capability Enhancement I (CE-I) EKV.
Moreover, MDA planned to test the CE-I EKV against a dynamic target
scene with countermeasures in both the FTG-04 and FTG-05 flight tests.
However, since FTG-04 was canceled and the target failed to release the
countermeasure in FTG-05, the fielded CE-I's ability against
countermeasures still has not been verified. According to MDA no more
CE-I EKV flight tests have been approved.
The test delays led MDA to restructure its flight test plan for fiscal
year 2009, increasing the number of tests, compressing the amount of
time to analyze and prepare for subsequent tests, and increasing the
scope of individual tests. For example, MDA plans to conduct 14 of 18
flight tests in the third and fourth quarter of fiscal year 2009. Past
testing performance raises questions about whether this is realistic.
In fiscal year 2008, MDA had planned to conduct 18 flight tests, but it
only accomplished 10, and delayed several flight tests into 2009. In
the next GMD end-to-end flight test--FTG-06 in fourth quarter fiscal
year 2009 to first quarter fiscal year 2010 --MDA is accepting a higher
level of risk than it previously expected in conducting this first test
of an enhanced configuration of the Kill Vehicle called the Capability
Enhancement II (CE-II)[Footnote 1] because it will include several
objectives that had planned to be previously tested, but have not been.
For example, the FTG-06 flight test will be the first GMD test
assessing both a CE-II EKV and a complex target scene. Adding to the
risk, it will be only the second test using a newly developed FTF LV-2
target. Moreover, MDA in January 2008 had merged FTG-06 and FTG-07,
thereby eliminating an additional opportunity to gather important
information from an intercept. FTG-07 will instead be an intercept test
of the two-stage interceptor intended for the European site.
Poor Target Missile Performance Continues to Hamper BMDS Testing:
Problems with the reliability and availability of targets (which are
themselves ballistic missiles) have increasingly affected BMDS
development and testing since 2006. As MDA recently acknowledged,
target availability became, in some cases, a pacing item for the
overall test program. As was noted in Table 1, problems with targets
have reduced testing of GMD, Sensors, and THAAD during 2008.
Repeated target problems and test cancellations have particularly
reduced opportunities to demonstrate the ability of sensors to
discriminate the real target from countermeasures. In the mid-course of
flight, a more sophisticated threat missile could use countermeasures
in an attempt to deceive BMDS radars and interceptor sensors as to
which is the actual reentry vehicle. In order to improve the
effectiveness of the BMDS against evolving threats, MDA elements are
developing advanced discrimination software in their component's
sensors to distinguish the threat reentry vehicle from countermeasures
and debris. The cancellation of FTG-04 and subsequent target problems
during FTX-03 and FTG-05 prevented opportunities to gather data to test
how well discrimination software performs in an operational
environment. The current fielded configuration of the GMD kill vehicle
has not been tested against countermeasures.
To address the growing need for more sophisticated and reliable targets
for the future BMDS test program, MDA has been developing a new set of
targets called the Flexible Target Family (FTF), which was intended to
provide new short, medium, and long-range targets with ground, air, and
sea launch capabilities. It was viewed as a family in the sense that
the different target sizes and the variants within those sizes would
use common components. MDA embarked on this major development without
estimating the cost to develop the family of target missiles. MDA
proceeded to develop and even to produce some FTF targets without a
sound business case and, consequently, their acquisition has not gone
as planned. The funds required for the FTF were spent sooner than
expected and were insufficient for the development.
Development of all FTF sizes and variants has been discontinued except
for the 72-inch diameter ground-launched target, referred to as the
LV-2. With guidance from the Missile Defense Executive Board, MDA is
currently conducting a comprehensive review of the targets program to
determine the best acquisition strategy for future BMDS targets. It is
expected to be completed in mid-2009. Whether or not MDA decides to
restart the acquisition of the 52-inch diameter targets, or other FTF
variants, depends on the results of this review.
The process of qualifying FTF target components for the LV-2 was more
difficult than expected. While many of the LV-2's components are found
on existing systems, their form, fit, function, and the environment
they must fly in are different. Consequently, many critical components
initially failed shock and vibration testing and other qualification
tests and had to be redesigned. MDA has acknowledged that the component
qualification effort ran in parallel with design completion and initial
manufacturing. So far, the resultant delays in the LV-2 target have had
two consequences. First, a planned test flight of the LV-2 itself for
the Space Tracking and Space Surveillance program was delayed and
instead its first flight will be as an actual target for an Aegis BMD
intercept. Second, because the LV-2 was not ready, that Aegis intercept
test was deferred from fiscal year 2008 to third quarter fiscal year
2009.
Other Consequences of Less Productive Testing:
In addition to delaying progress on individual elements, testing
problems have had other consequences for BMDS. Specifically, the
reduced productivity of testing has delayed understanding the overall
performance of BMDS, production and fielding have in some cases gotten
ahead of testing, and declarations of capabilities ready for fielding
have been made based on fewer tests and less modeling and simulation
than planned.
Overall Performance of BMDS Can Not Yet Be Assessed:
The overall performance of the BMDS cannot yet be assessed because MDA
lacks a fully accredited end-to-end model and simulation capability
and, according to the BMDS Operational Test Agency, it will not have
that capability until 2011 at the earliest. The lack of sufficient
flight test data has inhibited the validation of the models and
simulations needed for the ground tests and the simulation. MDA's
modeling and simulation program enables it to assess the capabilities
and limitations of how BMDS performs under a wider variety of
conditions than can be accomplished through the limited number of
flight tests conducted. Flight tests alone are insufficient because
they only demonstrate a single collection data point of element and
system performance. Flight tests are, however, an essential tool used
to both validate performance of the BMDS and to anchor the models and
simulations to ensure they accurately reflect real performance.
Computer models of individual elements replicate how those elements
function. These models are then aggregated into various combinations
that simulate the BMDS engagement of enemy ballistic missiles.
Developing an end-to-end system-level model and simulation has been
difficult. MDA's first effort to bring together different element
models and simulations to produce a fully accredited, end-to-end model
and simulation was for the first annual performance assessment of the
fielded BMDS configuration in 2007. Performance Assessment 2007 was
unsuccessful primarily because of inadequate data, particularly flight
test data, for verification and validation to support accreditation.
Instead, Performance Assessment 2007 used several models and
simulations that represented different aspects of the BMD system and
were not fully integrated. Consequently, acting on a joint
recommendation between MDA and the Operational Test Agency, MDA
officials canceled the 2008 performance assessment in April 2008
because of developmental risks associated with modeling and
simulations, focusing instead on testing and models for Performance
Assessment 2009.
According to the BMDS Operational Test Agency's January 2009 Modeling
and Simulation accreditation report, confidence in MDA's Modeling and
Simulation efforts remains low although progress was made during the
year. Out of 40 models, the BMDS Operational Test Agency recommended in
January 2009 full accreditation for only 6 models, partial
accreditation for 9 models, and no accreditation for 25 models. MDA is
now exercising stronger central leadership to provide guidance and
resources as they coordinate the development of verified and validated
models and simulations.
MDA intends to verify and validate models and simulations by December
2009 for Performance Assessment 2009. However, BMDS Operational Test
Agency officials stated that there is a high risk that the performance
assessment 2009 analysis will be delayed because of remaining
challenges and MDA's delayed progress in accreditation. MDA does not
expect to have a single end-to-end simulation for use in performance
assessments until 2010.
Production and Fielding Proceed Despite Delays in Testing and
Assessments:
Testing problems have contributed to a concurrent development,
manufacturing and fielding strategy in which assets are produced and
fielded before they are fully demonstrated through testing and
modeling. For example, although a test of the ability of the SM-3 Block
1A missile to engage and intercept a long range ballistic target was
delayed until the third quarter of fiscal year 2009, MDA purchased 20
of the missiles in fiscal year 2008 ahead of schedule.
While the GMD program has only been able to conduct two intercepts
since 2006 for assessing the fielded configuration, the production of
interceptors has continued. From the beginning of fiscal year 2007
through the first quarter of fiscal year 2009, MDA planned to conduct 7
flight tests and field 16 new ground-based interceptors. The plan
included a test that would utilize two ground-based interceptors
against a single target, known as a salvo test. By January 2009, GMD
had conducted only 2 flight tests and dropped the salvo test; yet it
fielded 13 ground-based interceptors.
Moreover, the GMD program had planned to conduct an intercept test to
assess the enhanced version of the EKV called the Capability
Enhancement II (CE-II) in the first quarter of fiscal year 2008, months
before emplacing any interceptors with this configuration. However,
developmental problems with the new configuration's inertial
measurement unit and the target delayed the first flight test with the
CE-II configuration--FTG-06--until at least fourth quarter fiscal year
2009. Despite these delays, emplacements will proceed; MDA expects to
have emplaced five CE-II interceptors before this flight test. More
importantly, GMD projects that the contractor will have manufactured
and delivered 10 CE-II EKVs before that first flight test demonstrates
the CE-II capability. This amounts to over half of the CE-II EKV
deliveries that are currently under contract.
Declaration of Capabilities Proceed with Reduced Levels of Information:
When MDA determines that a capability can be considered for operational
use it does so through a formal declaration. MDA bases its declarations
on, among other things, a combination of models and simulations--such
as end-to-end performance assessments (from missile launch to attempted
intercept)--and ground tests all anchored to flight test data.
In fiscal year 2008, MDA declared it had fielded 7 of 17 BMDS
capabilities planned for 2008 (postponing 10). In doing so MDA largely
reduced the basis for the declarations due in part to test problems and
delays. Specifically, MDA had intended to use a GMD flight test that
was canceled, a key ground test that was delayed and a performance
assessment that was canceled. MDA had to shift the basis of the 7
declarations to previous flight and ground tests.
Review of BMDS Modeling and Testing Holds Promise, but Must Anticipate
Contingences:
MDA has undertaken a three-phase review of the entire BMDS modeling,
simulation, and test program. According to MDA, the three phases
involve identifying critical variables that have not been proven to
date, determining what test scenarios are needed to collect the data,
and developing an affordable and prioritized schedule of flight and
ground tests. MDA intends to complete all three phases of the review by
May 2009. At this point, our knowledge of the review is limited, as we
have only had an introductory briefing on it. Nonetheless, the review
appears to offer a sound approach for closing the gaps that exist
between testing, modeling, and simulation. Further, the involvement of
test and evaluation organizations is encouraging.
While sound, the success of this approach hinges on providing
sufficient resources, ensuring robustness, and anticipating
contingencies. In addition to linking the critical modeling and
simulation variables with test events, the review will have to address
the factors that have limited the productivity of the current test
approach, such as the availability and performance of targets. MDA's
current approach to testing could be characterized as a just-in-time
approach to having the test assets, such as targets, ready. This left
little margin to solve issues that arise leading up to the tests.
Accordingly, the third phase of MDA's new approach--properly resourcing
the tests with sufficient time, funding and reliable targets--will be
key. MDA has indicated that its revision will result in a more robust
test plan, providing more margin to conduct the tests through, for
example, having spare interceptors and targets available.
Other contingencies that a new approach to modeling, simulation, and
testing should anticipate include unexpected or incomplete test
results, and problems in accrediting the models that are needed for
aggregated simulations, such as performance assessments. An important
consideration in this regard is for modeling, simulation, and testing
events to be re-synchronized so that they properly inform decisions on
producing, fielding, and declaring assets operational. Contingency
plans could then be formed for adjusting the pace of these decisions
should shortfalls occur in modeling, simulation, or testing.
MDA has indicated that this new approach to testing will take time to
implement, with partial implementation in fiscal year 2010 and full
implementation not occurring until fiscal year 2011. Therefore, MDA
must manage the transition to the new testing approach. In particular,
the ambitious fiscal year 2009 flight test plan may need to be
reassessed with the goal of establishing a robust series of tests that
can withstand some delays without causing wholesale changes to the test
plan during the transition. In the mean time, MDA will have to be
prudent in making decisions to produce and field additional assets.
Our annual report on missile defense is in draft and with DOD for
comment. It will be issued in final by March 13, 2009. In that report,
we are recommending additional steps to further improve the
transparency, accountability, and oversight of the missile defense
program. Our recommendations include actions to improve cost reporting
as well as testing and evaluation. DOD is in the process of preparing a
formal response to the report and its recommendations.
Madame Chairman, this concludes my statement. I would be pleased to
respond to any questions you or members of the subcommittee may have.
Contacts and Staff Acknowledgments:
For questions about this statement, please contact me at (202) 512-4841
or Francisp@gao.gov. Individuals making key contributions to this
statement include David B. Best, Assistant Director; Steven B. Stern;
LaTonya D. Miller; Thomas Mahalek; Ivy Hübler; Meredith Allen Kimmett;
Kenneth E. Patton; and Alyssa Weir.
[End of section]
Footnote:
[1] The CE-II was intended to replace obsolescent parts, but it has
demonstrated improved performance.
[End of section]
GAO's Mission:
The Government Accountability Office, the audit, evaluation and
investigative arm of Congress, exists to support Congress in meeting
its constitutional responsibilities and to help improve the performance
and accountability of the federal government for the American people.
GAO examines the use of public funds; evaluates federal programs and
policies; and provides analyses, recommendations, and other assistance
to help Congress make informed oversight, policy, and funding
decisions. GAO's commitment to good government is reflected in its core
values of accountability, integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each
weekday, GAO posts newly released reports, testimony, and
correspondence on its Web site. To have GAO e-mail you a list of newly
posted products every afternoon, go to [hyperlink, http://www.gao.gov]
and select "E-mail Updates."
Order by Phone:
The price of each GAO publication reflects GAO‘s actual cost of
production and distribution and depends on the number of pages in the
publication and whether the publication is printed in color or black and
white. Pricing and ordering information is posted on GAO‘s Web site,
[hyperlink, http://www.gao.gov/ordering.htm].
Place orders by calling (202) 512-6000, toll free (866) 801-7077, or
TDD (202) 512-2537.
Orders may be paid for using American Express, Discover Card,
MasterCard, Visa, check, or money order. Call for additional
information.
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]:
E-mail: fraudnet@gao.gov:
Automated answering system: (800) 424-5454 or (202) 512-7470:
Congressional Relations:
Ralph Dawn, Managing Director, dawnr@gao.gov:
(202) 512-4400:
U.S. Government Accountability Office:
441 G Street NW, Room 7125:
Washington, D.C. 20548:
Public Affairs:
Chuck Young, Managing Director, youngc1@gao.gov:
(202) 512-4800:
U.S. Government Accountability Office:
441 G Street NW, Room 7149:
Washington, D.C. 20548: