Defense Acquisitions

Production and Fielding of Missile Defense Components Continue with Less Testing and Validation Than Planned Gao ID: GAO-09-338 March 13, 2009

The Missile Defense Agency (MDA) has spent about $56 billion and will spend about $50 billion more through 2013 to develop a Ballistic Missile Defense System (BMDS). GAO was directed to assess the annual progress MDA made in developing the BMDS as well as improvements in accountability and transparency in agency operations, management processes, and the new block strategy. To accomplish this, GAO reviewed contractor cost, schedule, and performance; tests completed; and the assets fielded during 2008. GAO also reviewed pertinent sections of the U.S. Code, acquisition policy, and the activities of the new Missile Defense Executive Board (MDEB). An appendix on the effect the cancellation of a Ground-based Midcourse Defense flight test had on BMDS development is also included.

Cost: MDA has not yet established baselines for total costs or unit costs, both fundamental markers most programs use to measure progress. Consequently, for the sixth year, GAO has not been able to assess MDA's actual costs against a baseline of either total costs or unit costs. MDA planned to establish such baselines in 2008 in response to past GAO recommendations, but has delayed this until 2009. GAO was able to assess the cost performance on individual contracts, and project an overrun at completion of between $2 billion and $3 billion. However, because in some cases the budgeted costs at completion--the basis for our projection--has changed significantly over time as adjustments were made, this projection does not capture as cost growth the difference between the original and current budgeted costs at completion. In one case, these costs increased by approximately five times its original value. Performance and Testing: While MDA completed several key tests that demonstrated enhanced performance of the BMDS, all elements of the system had test delays and shortfalls. Overall, testing achieved less than planned. For example, none of the six Director's test knowledge points established by MDA for 2008 were achieved. Poor performing target missiles have been a persistent problem. Testing shortfalls have slowed the validation of models and simulations, which are needed to assess the system's overall performance. Consequently, the performance of the BMDS as a whole can not yet be determined. Schedule: Although fewer tests have been conducted than planned, the production and fielding of assets has proceeded closer to schedule. Except for no ground-based interceptors being delivered, all other radars, standard missiles, and software were delivered as planned. However, some deliveries, such as enhanced Exoatmospheric Kill Vehicles, will now precede test results. In most cases, MDA has also reduced the bases it planned to use to declare when capabilities are operational in the field. Thus, fielding decisions are being made with a reduced understanding of system effectiveness. Transparency, Accountability, and Oversight: Improvement in this area has been limited. The Missile Defense Executive Board (MDEB) has acted with increased authority in providing oversight of MDA and the BMDS. However, transparency and accountability into MDA's work is limited by the management fluidity afforded through the lack of cost baselines, an unstable test baseline, continued use of development funds to produce assets for fielding, and renewed potential for transferring work from one predefined block to another. A better balance must still be struck between the information Congress and the Department of Defense need to conduct oversight of the BMDS and the flexibility MDA needs to manage across the portfolio of assets that collectively constitute the system's capability. At this point, the balance does not provide sufficient information for effective oversight.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-09-338, Defense Acquisitions: Production and Fielding of Missile Defense Components Continue with Less Testing and Validation Than Planned This is the accessible text file for GAO report number GAO-09-338 entitled 'Defense Acquisitions: Production and Fielding of Missile Defense Components Continue with Less Testing and Validation Than Planned' which was released on March 16, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Committees: United States Government Accountability Office: GAO: March 2009: Defense Acquisitions: Production and Fielding of Missile Defense Components Continue with Less Testing and Validation Than Planned: GAO-09-338: GAO Highlights: Highlights of GAO-09-338, a report to Congressional Committees. Why GAO Did This Study: The Missile Defense Agency (MDA) has spent about $56 billion and will spend about $50 billion more through 2013 to develop a Ballistic Missile Defense System (BMDS). GAO was directed to assess the annual progress MDA made in developing the BMDS as well as improvements in accountability and transparency in agency operations, management processes, and the new block strategy. To accomplish this, GAO reviewed contractor cost, schedule, and performance; tests completed; and the assets fielded during 2008. GAO also reviewed pertinent sections of the U.S. Code, acquisition policy, and the activities of the new Missile Defense Executive Board (MDEB). An appendix on the effect the cancellation of a Ground-based Midcourse Defense flight test (FTG-04) had on BMDS development is also included. What GAO Found: Cost: MDA has not yet established baselines for total costs or unit costs, both fundamental markers most programs use to measure progress. Consequently, for the sixth year, GAO has not been able to assess MDA‘s actual costs against a baseline of either total costs or unit costs. MDA planned to establish such baselines in 2008 in response to past GAO recommendations, but has delayed this until 2009. GAO was able to assess the cost performance on individual contracts, and project an overrun at completion of between $2 billion and $3 billion. However, because in some cases the budgeted costs at completion”the basis for our projection”has changed significantly over time as adjustments were made, this projection does not capture as cost growth the difference between the original and current budgeted costs at completion. In one case, these costs increased by approximately five times its original value. Performance and Testing: While MDA completed several key tests that demonstrated enhanced performance of the BMDS, all elements of the system had test delays and shortfalls. Overall, testing achieved less than planned. For example, none of the six Director‘s test knowledge points established by MDA for 2008 were achieved. Poor performing target missiles have been a persistent problem. Testing shortfalls have slowed the validation of models and simulations, which are needed to assess the system‘s overall performance. Consequently, the performance of the BMDS as a whole can not yet be determined. Schedule: Although fewer tests have been conducted than planned, the production and fielding of assets has proceeded closer to schedule. Except for no ground-based interceptors being delivered, all other radars, standard missiles, and software were delivered as planned. However, some deliveries, such as enhanced Exoatmospheric Kill Vehicles, will now precede test results. In most cases, MDA has also reduced the bases it planned to use to declare when capabilities are operational in the field. Thus, fielding decisions are being made with a reduced understanding of system effectiveness. Transparency, Accountability, and Oversight: Improvement in this area has been limited. The Missile Defense Executive Board (MDEB) has acted with increased authority in providing oversight of MDA and the BMDS. However, transparency and accountability into MDA‘s work is limited by the management fluidity afforded through the lack of cost baselines, an unstable test baseline, continued use of development funds to produce assets for fielding, and renewed potential for transferring work from one predefined block to another. A better balance must still be struck between the information Congress and the Department of Defense need to conduct oversight of the BMDS and the flexibility MDA needs to manage across the portfolio of assets that collectively constitute the system‘s capability. At this point, the balance does not provide sufficient information for effective oversight. What GAO Recommends: GAO recommends that the MDEB assess how the transparency and accountability of MDA acquisitions can be strengthened without losing the benefits of MDA‘s existing flexibilities. Meanwhile, MDA should improve its cost and test baselines; tie modeling and simulation needs into test objectives; provide more time to analyze tests; better coordinate with independent testers; synchronize development, manufacturing, and fielding with testing and validation; complete a key developmental test; and strengthen the basis for capability declarations. DOD agreed with 10 of the 11 recommendations and partially agreed with one. To view the full product, including the scope and methodology, click on [hyperlink, http://www.gao.gov/products/GAO-09-338]. For more information, contact Paul Francis at (202) 512-4841 or francisp@gao.gov. [End of section] Contents: Letter: Background: Cost Tracking Deficiencies Hinder Assessment of Cost Performance: While Some Tests Succeeded, Others Were Deferred; Overall System Performance Cannot Yet Be Assessed: Production, Fielding, and Declaration of Capabilities Proceed despite Delays in Testing and Assessments: Production and Fielding of BMDS Systems Getting Ahead of Testing: Limited Progress Made in Improving Transparency and Accountability: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Comments from the Department of Defense: Appendix II: BMDS Prime Contractors Exceed Budgeted Cost and Schedule Performance during Fiscal Year 2008: Aegis BMD Contractors Experienced Mixed Performance during the Fiscal Year: ABL Contractor Overran Budgeted Fiscal Year Cost: C2BMC Program Incurred Negative Cumulative and Fiscal Year Variances: GMD Contractor Maintained Negative Cumulative Cost and Schedule Variances throughout the Fiscal Year: KEI Cost and Schedule Performance Continued to Decline after Replan: Limited Contractor Data Prevented Analysis of All MKV Task Orders: Sensors' Radar Experienced Fiscal Year Cost and Schedule Growth: Technical Issues Drove STSS Cost Growth during the Fiscal Year: Targets and Countermeasures Program's Rebaseline Positively Affected Fiscal Year Schedule Variances: THAAD Contractor Spent More Money and Time Than Budgeted: Appendix III: FTG-04 Flight Test Cancellation: Faulty Telemetry Component Caused Delay and Subsequent Cancellation of FTG-04: Most FTG-04 Test Objectives Will Be Allocated to Follow-on Tests: Cancellation Eliminates One of Few Opportunities to Demonstrate GMD Capabilities: Conclusions: Appendix IV: Reduced Basis for Capability Declarations: Appendix V: Scope and Methodology: Appendix VI: GAO Contact and Staff Acknowledgments: Tables: Table 1: MDA BMDS Elements: Table 2: MDA Block Construct: Table 3: Fiscal Year 2008 Capability Goals for Blocks 1.0, 2.0, and 3.1/3.2: Table 4: Analysis of Contractor Realignments from Contract Start through Fiscal Year 2008: Table 5: Prime Contractor Fiscal Year 2008 and Cumulative Cost and Schedule Performance: Table 6: Test and Targets Issues: Table 7: Status of Fiscal Year 2008 Director's Test Knowledge Points: Table 8: BMDS Deliveries and Total Fielded Assets as of September 30, 2008: Table 9: MDA BMDS Test Baseline Revisions: Table 10: Timeline of Events: Table 11: Engagement Sequence Groups with Revised Basis for Fiscal Year 2008 Capability Declarations: Table 12: Block 1.0 Engagement Sequence Groups with Revised Basis for Completion at End of Fiscal Year 2009: Figures: Figure 1: Estimated Percentage of Total BMDS Block and Capability Development Funds through Fiscal Year 2013 Expected to Be Baselined in 2009: Figure 2: Difference in Traditional Unit Cost Reporting and MDA's Unit Cost Reporting: Figure 3: GMD Reduction in Flight Tests from January 2006 to March 2010: Figure 4: GMD Flight Test and Fielding Plan for Interceptors Comparison--September 2006 versus January 2009: Figure 5: Timeline Showing Declaration of Capabilities in Fiscal Year 2008: Figure 6: Timeline Showing Deferred Declaration of Capabilities from Fiscal Year 2008 to 2009: Figure 7: Aegis BMD Weapon System Fiscal Year 2008 Cost and Schedule Performance: Figure 8: Aegis BMD SM-3 CLIN 1 Fiscal Year 2008 Cost and Schedule Performance: Figure 9: ABL Fiscal Year 2008 Cost and Schedule Performance: Figure 10: C2BMC Fiscal Year 2008 Cost and Schedule Performance: Figure 11: GMD Fiscal Year 2008 Cost and Schedule Performance: Figure 12: KEI Fiscal Year 2008 Cost and Schedule Performance: Figure 13: MKV Task Order 6 Fiscal Year 2008 Cost and Schedule Performance: Figure 14: MKV Task Order 7 Fiscal Year 2008 Cost and Schedule Performance: Figure 15: MKV Task Order 8 Fiscal Year 2008 Cost and Schedule Performance: Figure 16: Sensors Fiscal Year 2008 Cost and Schedule Performance: Figure 17: STSS Fiscal Year 2008 Cost and Schedule Performance: Figure 18: Targets and Countermeasures Fiscal Year 2008 Cost and Schedule Performance: Figure 19: THAAD Fiscal Year 2008 Cost and Schedule Performance: Abbreviations: ABL: Airborne Laser: Aegis BMD: Aegis Ballistic Missile Defense: BMDS: Ballistic Missile Defense System: C2BMC: Command, Control, Battle Management, and Communications: CE: Capability Enhancement: CLIN: Contract Line Item Number: DOD: Department of Defense: DOT&E: Director, Operational Test and Evaluation: EKV: Exoatmospheric Kill Vehicle: FTF: Flexible Target Family: GBI: Ground-based Interceptor: GMD: Ground-based Midcourse Defense: KEI: Kinetic Energy Interceptor: MDA: Missile Defense Agency: MDEB: Missile Defense Executive Board: MKV: Multiple Kill Vehicle: PCME: Pulse Code Modulation Encoder: SAR: Selected Acquisition Report: STSS: Space Tracking and Surveillance System: THAAD: Terminal High Altitude Area Defense: [End of section] United States Government Accountability Office: Washington, DC 20548: March 13, 2009: Congressional Committees: The Missile Defense Agency (MDA) has spent almost $56 billion since its initiation in 2002 on developing and fielding a Ballistic Missile Defense System (BMDS) and is on course to spend about $50 billion more over the next 5 years. In 2002, the President directed the Department of Defense (DOD) to "deploy a set of initial missile defense capabilities beginning in 2004".[Footnote 1] MDA began delivering an initial capability in late 2004, as directed, and deployed an initial capability in 2005 by concurrently developing and fielding BMDS assets.[Footnote 2] Though this approach facilitated the rapid deployment of an initial BMDS capability, as MDA has proceeded beyond that initial capability, it has been less successful in fostering adequate knowledge of system capabilities prior to manufacturing and fielding BMDS assets. In its fiscal year 2002, 2007, and 2008 National Defense Authorization Acts, Congress directed GAO to assess the cost, schedule, testing, and performance progress that MDA is making in developing the BMDS. [Footnote 3] We have delivered assessments covering fiscal years 2003 through 2007.[Footnote 4] This report assesses the progress made during fiscal year 2008 toward BMDS goals as well as the progress MDA made in improving accountability and transparency through its agency operations, management, processes, and new block strategy. This report also includes an appendix that addresses the Senate Armed Services Committee's request that we review the reasons behind and the effects on BMDS development of the cancellation of a Ground-based Midcourse Defense flight test designated FTG-04. To assess progress during fiscal year 2008, we examined the accomplishments of 10 BMDS elements that MDA is developing and fielding: the Aegis Ballistic Missile Defense (Aegis BMD); Airborne Laser (ABL); BMDS Sensors; Command, Control, Battle Management, and Communications (C2BMC); Ground-based Midcourse Defense (GMD); Kinetic Energy Interceptors (KEI); Multiple Kill Vehicles (MKV); Space Tracking and Surveillance System (STSS); Targets and Countermeasures; and Terminal High Altitude Area Defense (THAAD).[Footnote 5] These elements collectively account for about 80 percent of MDA's research and development budget. We also examined MDA's Fiscal Year 2008 Statement of Goals, Program Execution Reviews, test plans and reports, production plans, and Contract Performance Reports. We interviewed officials within program offices and within MDA functional directorates, such as the Directorate for Cost Estimating. In addition, we discussed each element's test program and its results with the BMDS Operational Test Agency and DOD's Office of the Director, Operational Test and Evaluation (DOT&E). In assessing progress made toward improving accountability and transparency, we held discussions with officials in MDA's Directorate of Business Operations to determine whether its new block structure improved accountability and transparency of the BMDS. In addition, we reviewed pertinent sections of the U.S. Code to compare MDA's current level of accountability with federal acquisition laws. We also interviewed officials from the Office of the Under Secretary of Defense for Acquisition, Technology and Logistics to discuss the oversight role of the Missile Defense Executive Board. Additionally, we reviewed the board's charter to determine its oversight responsibility. Our scope and methodology is discussed in more detail in appendix V. We conducted this performance audit from May 2008 to March 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Background: MDA's mission is to develop an integrated and layered system to defend the United States and its deployed forces, friends, and allies against ballistic missile attacks. The BMDS aims to engage all ranges of ballistic missiles during all phases of flight. This challenging expectation requires complex coordination within an integrated system of defensive components--space-based sensors, surveillance and tracking radars, advanced interceptors, and a battle management, command, control, and communication component. A typical engagement scenario to defend against an intercontinental ballistic missile would occur as follows: * Infrared sensors aboard early-warning satellites detect the hot plume of a missile launch and alert the command authority of a possible attack. * Upon receiving the alert, land-or sea-based radars are directed to track the various objects released from the missile and, if so designed, to identify the warhead from among spent rocket motors, decoys, and debris. * When the trajectory of the missile's warhead has been adequately established, an interceptor--consisting of a kill vehicle mounted atop a booster--is launched to engage the threat. The interceptor boosts itself toward a predicted intercept point and releases the kill vehicle. * The kill vehicle uses its onboard sensors and divert thrusters to detect, identify, and steer itself into the warhead. With a combined closing speed of up to 10 kilometers per second (22,000 miles per hour), the warhead is destroyed above the atmosphere through a "hit to kill" collision with the kill vehicle. * Some interceptors use sensors to steer themselves into the inbound ballistic missile. Inside the atmosphere, these systems kill the ballistic missile using a range of mechanisms such as direct collision between the missile and the inbound ballistic missile or killing it with the combined effects of a blast fragmentation warhead (heat, pressure, and grains/shrapnel) in cases where a direct hit does not occur. Table 1 provides a brief description of 10 BMDS elements currently under development by MDA. Table 1: MDA BMDS Elements: BMDS element: Aegis Ballistic Missile Defense; Missile defense role: Aegis BMD is a ship-based missile defense system designed to destroy short-to intermediate-range ballistic missiles during the midcourse phase of their flight; its capability is being expanded to include the terminal phase of flight. Aegis BMD's mission is twofold--an engagement capability against regional ballistic missile threats providing the BMDS with its first mobile, global, deployable and proven capability that can destroy ballistic missiles both above and within the atmosphere, as well as a forward-deployed combatant to search, detect, and track ballistic missiles of all ranges and transmit track data to the BMDS, performing a strategic role in homeland defense. To date, 18 ships have been upgraded for the Aegis BMD mission. MDA is planning to procure 147 Aegis BMD missiles--the Standard Missile-3 (SM-3)--from calendar years 2004 through 2013. BMDS element: Airborne Laser; Missile defense role: ABL is an air-based missile defense system designed to destroy all classes of ballistic missiles during the boost phase of their flight. ABL employs a high-energy chemical laser to rupture a missile's motor casing, causing the missile to lose thrust or flight control. MDA plans to demonstrate proof of concept in a system demonstration in 2009. BMDS element: BMDS Sensors; Missile defense role: MDA is developing radars for fielding as part of the BMDS. The BMDS uses these sensors to identify and track ballistic missiles. The ultimate goal is to provide continuous tracking of ballistic missiles in all phases of flight and increase the probability for successful intercept. BMDS element: Command, Control, Battle Management and Communications; Missile defense role: C2BMC is the integrating element of the BMDS. Its role is to provide deliberate planning, situational awareness, sensor management, and battle management for the integrated BMDS. BMDS element: Ground-based Midcourse Defense; Missile defense role: GMD is a ground-based missile defense system designed to destroy intercontinental ballistic missiles during the midcourse phase of their flight. Its mission is to protect the U.S. homeland against ballistic missile attacks from North Korea and the Middle East. Currently, GMD has fielded 24 interceptors with the original configuration, known as Capability Enhancement-I (CE-I). GMD has recently begun emplacing a new configuration of the kill vehicle known as the Capability Enhancement-II (CE-II). This configuration was designed to replace obsolete parts. MDA is planning on fielding 44 interceptors at Fort Greely, Alaska, and Vandenberg Air Force Base, California, by fiscal year 2011. MDA also plans to field 10 interceptors in Europe. BMDS element: Kinetic Energy Interceptors; Missile defense role: KEI is a mobile land-based missile defense system designed to destroy medium, intermediate, and intercontinental ballistic missiles during the boost and midcourse phases of their flight. The agency plans to conduct the first booster flight test in 2009. The KEI capability could be expanded to sea basing in subsequent blocks. BMDS element: Multiple Kill Vehicle; Missile defense role: The MKV is being designed as a spiral improvement to counter advancements in the threat for midcourse interceptors. This approach mitigates the need to pinpoint a single lethal object in a threat cluster by using numerous kill vehicles to engage all objects that might be lethal. The system under development consists of a carrier vehicle housing a number of smaller kill vehicles, which would primarily benefit the Ground-based and Kinetic Energy interceptors as well as the Aegis BMD SM-3. To mitigate risk, MDA has initiated a parallel acquisition with a second contractor. Because MKV is in the technology development stage, it does not project an initial capability date, but the program expects that the capability could be available by 2017. BMDS element: Space Tracking and Surveillance System; Missile defense role: STSS is designed to be a low-orbit constellation of space-based sensors that is able to observe targets in all phases of trajectory. MDA intends to launch two demonstration satellites in 2009. If the demonstration satellites perform successfully, MDA plans to have an operational capability of next-generation satellites. BMDS element: Targets and Countermeasures; Missile defense role: MDA maintains a series of targets used in BMDS flight tests to present authentic threat scenarios. The targets are designed to encompass the full spectrum of threat missile ranges and capabilities. In 2005, MDA began developing a new family of targets, the Flexible Target Family (FTF), which was to represent evolving threats of all ranges. However, in 2008, MDA narrowed the FTF focus to developing one long-range 72-inch target, the LV-2. The first launch of this target is scheduled for 2009. BMDS element: Terminal High Altitude Area Defense; Missile defense role: THAAD is a ground-based missile defense system designed to destroy short-and medium-range ballistic missiles during the terminal phase of flight, both inside and outside of the atmosphere. Its mission is to defend deployed U.S. forces and population centers. MDA plans to field a fire unit, which includes 24 missiles, in 2010 and a second unit in 2011. MDA also plans to field two additional fire units, which includes 24 missiles each, in 2012 and 2013, respectively. Source: MDA data. [End of table] To manage BMDS development, MDA uses an acquisition strategy defined by a block structure. From its inception in 2002 through 2007, MDA developed BMDS capability in biennial increments, ultimately delivering two blocks--Block 2004 and Block 2006. These 2-year blocks each built on preceding blocks and enhanced the development and capability of the BMDS. However, in response to recommendations from GAO, in December 2007 MDA announced a new block structure that was intended to improve the program's transparency, accountability, and oversight. The new blocks are not based on biennial time periods, but instead focus on fielding capabilities that address particular threats. Because the new block structure is not aligned to regular time periods, multiple blocks are underway concurrently. Table 2 details the current blocks and categories included in the BMDS block structure. Table 2: MDA Block Construct: Block: Block 1.0: Defend U.S. from Limited North Korean Long-Range Threats; Description: Provides an initial capability to protect the United States from a limited North Korean attack. It is the most mature capability and will be the first block delivered to the warfighter. Block: Block 2.0: Defend Allies and Deployed Forces from Short-to Medium-Range Threats in One Region/Theater; Description: Includes capabilities needed to defend allies and deployed forces from short-to medium-range threats in one region/theater. Block: Block 3.0: Expand Defense of the U.S. to Include Limited Iranian Long-Range Threats[A]; Description: Builds on the foundation established in Block 1.0 and includes capabilities needed to expand the defense of the United States against limited Iranian long-range threats. Block: Block 4.0: Defend Allies and Deployed Forces in Europe from Limited Iranian Long-Range Threats; Description: Builds on the foundation established by Blocks 1.0 and 3.0 capabilities needed to defend allies and deployed forces in Europe from limited Iranian long-range threats and to expand protection of the U.S. homeland. Block: Block 5.0: Expand Defense of Allies and Deployed Forces from Short-to Intermediate-Range Threats in Two Regions/Theaters; Description: Builds on the foundation established by Block 2.0 and includes capabilities needed to expand defense of allies and deployed forces from short-to intermediate-range threats in two regions/ theaters. Categories: Capability Development; Description: Includes BMDS elements and other development elements that are not baselined in the existing agency block structure, such as ABL, KEI, and MKV. These programs have knowledge points tailored to critical risks. Categories: Sustainment; Description: Funding for Contractor Logistics Support and other operation and support activities. Categories: Mission Area Investment; Description: Investments that cut across several blocks and cannot be reasonably allocated to a specific block. Examples include modeling and simulation and intelligence and security. Categories: MDA Operations; Description: Contains operations support functions such as MDA headquarters management. Source: MDA data. [A] Block 3.0 is subdivided into three sections: 3.1, 3.2, and 3.3. Block 3.0 will focus on more sophisticated sensors and algorithms, and therefore includes upgrades to the Ground-based Interceptors, sensors, and the C2BMC system to allow discrimination of the threat missile. MDA is pursuing two parallel and complementary approaches to counter complex countermeasures. The full implementation of this approach will be conducted in phases, with the first phase referred to as Near Term Discrimination (Block 3.1/3.2) and the second phase as Improved Discrimination and System Track. (Block 3.3). [End of table] MDA uses a Statement of Baselines and Goals to report modifications to established block baselines.[Footnote 6] For those blocks that are currently or will soon be underway, block baselines are created to make a firm commitment to Congress. The Statement of Goals also includes the following: * BMDS Baseline Capabilities - Assets and engagement sequence groups that will be made available for fielding for a particular block. [Footnote 7] During 2008, cost baselines were under development for Block 2.0 and Block increments 3.1/3.2. MDA established schedule and performance baselines for Blocks 1.0, 2.0, 3.1, and 3.2 in 2008. * BMDS Capability Goals - Assets and engagement sequence groups expected to be made available for future blocks. * Adversary Benchmarks - Adversary missile systems used for block performance estimates. * BMDS Budget Breakdowns - Detailed fielding, development, and integration budgets for each block and BMDS Capability Development activity. MDA also uses an incremental declaration process to designate BMDS capability for its blocks. Three capability designations are applied to all BMDS elements, their hardware and software components, and engagement sequence groups. This allows these BMDS features to play a limited role in system operations before they have attained their expected level of capability. Each capability designation in the delivery schedule represents upgraded capacity to support the overall function of BMDS in its mission as well as the level of MDA confidence in the system's performance. The designations are defined as follows: * Early Capability Delivery signifies readiness for contingency use. At this point, MDA has determined that the capability can be utilized by the BMDS. When integrated, the capability must be adequately demonstrated to build sufficient confidence that it will safely perform as intended without degrading the existing capabilities of the BMDS. * Partial Capability Delivery is an intermediate state of maturity indicating that a capability has been shown through testing to perform as intended in certain scenarios. At this point, MDA is sufficiently confident that the capability can support the warfighter's partially mission-capable objectives and logistics support is adequate to achieve defensive operations. * Full Capability Delivery is the point at which a capability satisfies the BMDS block objectives and is considered to be completely mature and ready for full operational use. MDA's capability goals for fiscal year 2008 for Blocks 1.0, 2.0, 3.1, and 3.2 are shown in table 3. Table 3: Fiscal Year 2008 Capability Goals for Blocks 1.0, 2.0, and 3.1/3.2: Block 1.0--Initial Defense of U.S. from North Korea Expected Completion: Fiscal Year 2009: Engagement Sequence Group: Ground-Based Interceptor (GBI) Launch on COBRA DANE/Upgraded Early Warning Radar; 2008 Planned Capability Deliveries: Early. Engagement Sequence Group: GBI Engage on COBRA DANE/Upgraded Early Warning Radar; 2008 Planned Capability Deliveries: Full. Engagement Sequence Group: GBI Engage on forward-based AN/TPY-2 radar; 2008 Planned Capability Deliveries: Full. Engagement Sequence Group: GBI Engage on Sea-based X-band radar; 2008 Planned Capability Deliveries: Early. Engagement Sequence Group: GBI Launch on Sea-based X-band radar; 2008 Planned Capability Deliveries: Early. Block 2.0--Initial Defense of Allied Forces Expected Completion: Fiscal Year 2011: Engagement Sequence Group: SM-2 Engage on shipboard Aegis radar (AN/ SPY-1); 2008 Planned Capability Deliveries: Early. Engagement Sequence Group: SM-3 Engage on shipboard Aegis radar (AN/ SPY-1); 2008 Planned Capability Deliveries: Full. Engagement Sequence Group: SM-3 Launch on remote on shipboard Aegis radar (AN/SPY-1); 2008 Planned Capability Deliveries: Early. Engagement Sequence Group: THAAD Interceptor Engage on AN/TPY-2 radar in the terminal mode; 2008 Planned Capability Deliveries: Early. Engagement Sequence Group: Block 3.1/3.2--Initial Defense of U.S. from Iran Expected Completion: Fiscal Year 2013; 2008 Planned Capability Deliveries: [Empty]. Engagement Sequence Group: GBI Engage on COBRA DANE/Upgraded Early Warning Radar Mod 1 (Fylingdales, UK; Forward-Based mobile radar (AN/ TPY-2)); 2008 Planned Capability Deliveries: Partial. Engagement Sequence Group: GBI Launch on shipboard Aegis radar Mod 1 (Fylingdales, UK; Sea-based X-band radar); 2008 Planned Capability Deliveries: Early. Source: GAO analysis of MDA data. Note: In addition to the engagement sequence groups listed above, as of October 2007 MDA had planned to declare the capability of several more engagement sequence groups in fiscal year 2008. However these were excluded from the February 2008 Statement of Goals. MDA continues to work toward declaring these additional engagement sequence groups. [End of table] Cost Tracking Deficiencies Hinder Assessment of Cost Performance: MDA has not yet established baselines for total costs or unit costs, both fundamental markers that most programs use to measure progress. MDA had planned to establish total cost baselines at the element and block levels in 2008, but the initial set of total cost baselines will not be available until the spring of 2009. Similarly, MDA has not established unit costs for selected assets, such as GBIs. Consequently, for the sixth year, we have been unable to assess MDA's overall progress on total or unit cost. While MDA plans to establish some total cost baselines in 2009, most efforts will not be captured in a baseline. MDA also plans to establish unit costs, another improvement, but is considering a narrower definition of unit cost than is used by other weapon system programs. MDA's definition will report a subset of procurement costs called flyaway costs, which only includes the major piece of equipment and excludes all research and development as well as some procurement costs--those for support equipment and spares. Moreover, these unit costs will only be tracked within those blocks that are baselined, which will represent a minority of those assets being produced and fielded. Without total cost baselines in place, the BMDS concept continually evolves, as indicated by the number of realignments to the program of work at the individual contract level. While the changing nature of the BMDS and the lack of total cost baselines precludes analysis of total cost progress, we were able to analyze contractor fiscal year performance on the current work under the contract. We were also able to project overruns or underruns at completion for BMDS contracts using the contracts' current budgeted costs at completion as a basis for our projections. However, in some cases, the current budgeted cost at completion changed significantly over time. In one case, the budgeted cost at completion increased by approximately five times its original value. Our analysis of fiscal year 2008 progress shows that several prime contractors exceeded budgeted costs. Absence of Cost Baselines Prevents Assessment of System-Level Costs: To provide accountability, major defense acquisition programs are required by statute to document program goals in an acquisition program baseline.[Footnote 8] MDA is not yet required to establish an acquisition program baseline because of the acquisition flexibilities it has been granted. However, Congress has enacted legislation requiring MDA to establish some baselines.[Footnote 9] Baselines serve an important discipline both by ensuring that the full cost commitment is considered before embarking on major development efforts and by identifying cost growth as a program proceeds. Since we began annual reporting on missile defense in 2004, we have been unable to assess overall progress on cost--that is, comparing BMDS baselined costs with actual costs. For example, under the prior block structure, we reported that BMDS costs grew by at least $1 billion, but the total cost growth could not be determined because MDA did not account for all costs for a given block. In response to recommendations we made in March 2008, MDA agreed to develop cost estimates and to provide independent verification for blocks outlined under its new approach. Upon conclusion of those estimates, MDA will develop cost baselines. In addition, on April 1, 2008, the Director, MDA, testified before a Senate Armed Services Subcommittee that cost baselines for the new block structure would be available by the end of 2008. As of January 2009, the agency had not yet developed full cost baselines for any blocks. MDA plans to have these cost baselines for Blocks 2.0, 3.1 and 3.2 completed, independently reviewed by DOD's Cost Analysis Improvement Group, and released by the spring of 2009. The only information that was available for this report was limited to budget projections for BMDS blocks and capability development for fiscal years 2008 through 2013, totaling approximately $42 billion. Even with the release of some block cost estimates in 2009, all block costs will not be baselined and no date has been established for when the remaining block costs will be baselined. MDA does not plan to baseline Block 1.0 costs, but will provide the actual costs of the block since it is near completion. Full cost baselines for Block 2.0 are anticipated to be available in the spring of 2009, but only portions of Block 3.0 and none of Block 4.0 and 5.0 costs will be baselined at this time. As figure 1 shows, if MDA does complete baselines as planned, they will only cover about 26 percent of its block and capability development costs. Figure 1: Estimated Percentage of Total BMDS Block and Capability Development Funds through Fiscal Year 2013 Expected to Be Baselined in 2009: [Refer to PDF for image: vertical bar graph] Fiscal year: 2010; Percentage of funding baselined (Blocks 2.0, 3.1, and 3.2): 26.26%. Fiscal year: 2011; Percentage of funding baselined (Blocks 2.0, 3.1, and 3.2): 5.65%. Fiscal year: 2012; Percentage of funding baselined (Blocks 2.0, 3.1, and 3.2): 3.36%. Fiscal year: 2013; Percentage of funding baselined (Blocks 2.0, 3.1, and 3.2): 2.26%. Source: GAO analysis of MDA‘s Fiscal Year 2009 Budget Estimate Submission and January 2008 Statement of Goals. Note: Analysis is based on MDA's fiscal year 2009 projected funding through fiscal year 2013 from the February 2008 request. Funding includes defense-wide resources projected for MDA. [End of figure] At this point, MDA plans to baseline between 2 and 26 percent of BMDS block and capability development costs from fiscal years 2010 to 2013 as depicted above. MDA has not determined when other blocks will be baselined. If other blocks were to be baselined before 2013, the percentage of funding baselined would be increased. The rapid decline in percentage of baselined funds also shows that initial baselines are being set late in a block's duration. For example, Block 2.0 will be completed within 2 years of its baseline being set. If baselines are to facilitate management and oversight, they will have to be set sooner for Block 3.3 and beyond. Additionally, agency officials stated that although cost estimates will be developed for individual capability development efforts--such as ABL, KEI, and MKV--the agency does not plan to baseline their costs until these elements are matured and moved into a defined block. MDA may eventually baseline these elements as part of a block once a firm commitment can be made to Congress. The budgets for capability development elements account for approximately $22 billon--or more than half of MDA's fiscal year 2008 6-year Future Years Defense Plan for BMDS blocks and capability development. Planned Unit Cost Reporting Will Not Be Comprehensive: Major defense acquisition programs are required by statute to report certain unit costs to Congress,[Footnote 10] track unit cost growth against their original and current baseline estimates, and perform an additional assessment of the program if certain cost growth thresholds are reached.[Footnote 11] This cost monitoring mechanism helps ensure that programs are being held accountable. MDA is not yet required to report these unit costs because of the acquisition flexibilities it has been granted by DOD, but Congress has enacted legislation requiring MDA to provide unit cost reporting data for certain BMDS elements, and MDA does plan to develop and report unit costs for some of its assets in the spring of 2009.[Footnote 12] The agency has also established thresholds for reporting cost growth. However, the approach MDA is taking, while an improvement, provides a much less comprehensive assessment of unit cost compared to the traditional acquisition costs that are typically reported for major defense acquisition programs. Normally, unit costs are reported in two ways: (1) program acquisition unit cost, which is the total cost for the development and procurement of, and system-specific military construction for, the acquisition program divided by the number of fully configured end items to be produced, or (2) average procurement unit cost, which is the total of all funds programmed to be available for obligation for procurement divided by the number of end items to be procured.[Footnote 13] MDA's development of the BMDS outside of DOD's normal acquisition process makes it difficult to compare the actual unit cost of a delivered asset with its planned unit cost. For example, MDA plans to only report recurring unit flyaway costs for the blocks that are baselined.[Footnote 14] Figure 2 reveals the significant reduction in standard areas of costs covered by MDA's approach compared to that normally reported for major defense acquisition programs. Figure 2: Difference in Traditional Unit Cost Reporting and MDA's Unit Cost Reporting: [Refer to PDF for image: illustration] Most major acquisitions programs report: Program acquisition cost: Development cost: Research, Development, Test, and Evaluation: * Development costs of Prime Mission Equipment and support items; * Systems engineering * Program management * Test and evaluation Military Construction: * Facilities Procurement cost: Procurement: * Prime equipment; * Support items; * Initial spares. MDA plans to report: Flyaway cost: Procurement: * Prime equipment. Source: GAO analysis. [End of figure] MDA's decision to report only flyaway unit costs will not capture research and development costs associated with BMDS assets--which account for more than 97 percent of the nearly $56 billion MDA costs to date. In addition, the procurement costs for initial spares and support equipment are not included. Thus, while the flyaway cost baseline will provide visibility into changes in recurring manufacturing costs, it will not provide a basis for comparison with the costs of other DOD programs. If a cost increase occurs in research and development or nonrecurring procurement, it will not be reported in MDA's unit cost. Agency officials told us that the reason for using flyaway unit costs was the new MDA block structure. They further explained that within the block structure, there are many cases where MDA procures and delivers a weapon system for more than one block and, in some cases, the same configuration in more than one block. For example, THAAD deliveries are included in Blocks 2.0 and 5.0. Agency officials cite this as a key difference between MDA and other defense programs. For MDA, most of the development costs are assigned to the first block where a capability is delivered and very little of the development cost is assigned to subsequent blocks. MDA officials further stated that if the agency were to use the standard unit cost methodology, it would show very dissimilar unit costs between the first and subsequent block deliveries and the difference could not be explained by learning curves and manufacturing efficiencies. MDA officials also told us that they chose unit flyaway cost for unit cost reporting because flyaway cost provides a better measure of what the individual system components cost to procure. However, MDA is not precluded from also determining and reporting unit costs by taking the entire cost of the asset being developed without regard to the capability or block for which it is originally developed. Frequent Realignments Indicate That Full Scope Is Not Yet Determined: Without a firm cost commitment or baseline in place for the full scope of the BMDS, in some cases the work currently under contract changes frequently. These changes manifest themselves in realignments that often add scope, cost and time to the value of the work under the contract. Since contracts began on the BMDS, MDA has performed 31 realignments where, in some cases, the period of stability between these major realignments average only between 1 to 2 years. These frequent changes indicate that the total BMDS effort has not been fully determined and is likely to grow. While we have been able to make an assessment of contractor costs, that assessment is limited to the current approved program. Until total and unit cost baselines are established, the only tool for us to use in assessing BMDS costs is the costs reported on individual contracts under BMDS's Earned Value Management System.[Footnote 15] All BMDS contracts that we assessed have a cost and schedule baseline against which progress, measured by cost and schedule performance, can be measured. It is appropriate for a program to realign its current status with the remaining contractual effort when officials conclude that the baseline no longer provides valid performance assessment information.[Footnote 16] A program can realign its current status with the remaining contractual effort through rebaselines, replans, and restructures. * A rebaseline is a more general term to describe a major realignment of the performance measurement baseline used to better correlate the work plan with the baseline budget, scope, and schedule and can refer to replans and restructures as well. * A replan is a reallocation of schedule or budget for the remaining effort within the existing constraints of the contract. * A restructure includes adding funds to the performance management budget that exceed the value of the negotiated contract and result in a contract modification. For purposes of this report, we refer to each of these types of program changes as realignments. Since work under the BMDS contracts first began, there have been a total of 31 realignments to the program which have added nearly $14 billion dollars to the value of the work under the contracts. Table 4 shows the BMDS elements' realignments since contract start. Table 4: Analysis of Contractor Realignments from Contract Start through Fiscal Year 2008: ABL; Contract start: Nov-96; Date of the last realignments[A]: June-07; Total number realignments: 6; Total contract value increase due to realignments: $2,590,881,491; Total period of performance increase due to realignments (in months): 80; Average time between realignments: 2 years; Average contract value increase per year due to realignments: $218,947,732; Average period of performance increase per year due to realignments (in months): 7. Aegis BMD; Contract start: Oct-03; Date of the last realignments[A]: Mar-06; Total number realignments: 1; Total contract value increase due to realignments: 0; Total period of performance increase due to realignments (in months): 0; Average time between realignments: 4 years 11 months; Average contract value increase per year due to realignments: 0; Average period of performance increase per year due to realignments (in months): 0. C2BMC; Contract start: Feb-02; Date of the last realignments[A]: Nov-06; Total number realignments: 1; Total contract value increase due to realignments: $36,514,950; Total period of performance increase due to realignments (in months): 0; Average time between realignments: 6 years 7 months; Average contract value increase per year due to realignments: $5,546,575; Average period of performance increase per year due to realignments (in months): 0. GMD[B]; Contract start: Jan-01; Date of the last realignments[A]: Ongoing as-of Sep-08; Total number realignments: 8; Total contract value increase due to realignments: $8,086,607,706; Total period of performance increase due to realignments (in months): 27; Average time between realignments: 1 year; Average contract value increase per year due to realignments: $1,054,774,918; Average period of performance increase per year due to realignments (in months): 4. KEI; Contract start: Dec-03; Date of the last realignments[A]: Apr-08; Total number realignments: 4; Total contract value increase due to realignments: $1,639,800,000; Total period of performance increase due to realignments (in months): 20; Average time between realignments: 1 year 2 months; Average contract value increase per year due to realignments: $345,000,000[C]; Average period of performance increase per year due to realignments (in months): 4. MKV; Contract start: Jan-04; Date of the last realignments[A]: July-07; Total number realignments: 4; Total contract value increase due to realignments: $51,213,935; Total period of performance increase due to realignments (in months): 14; Average time between realignments: 1 year 2 months; Average contract value increase per year due to realignments: $10,974,415; Average period of performance increase per year due to realignments (in months): 3. Sensors; Contract start: Apr-03; Date of the last realignments[A]: N/A; Total number realignments: 0; Total contract value increase due to realignments: 0; Total period of performance increase due to realignments (in months): 0; Average time between realignments: N/A; Average contract value increase per year due to realignments: 0; Average period of performance increase per year due to realignments (in months): 0. STSS; Contract start: Apr-02; Date of the last realignments[A]: Oct-07; Total number realignments: 1; Total contract value increase due to realignments: $232,293,329; Total period of performance increase due to realignments (in months): 13; Average time between realignments: 6 years 5 months; Average contract value increase per year due to realignments: $36,201,558; Average period of performance increase per year due to realignments (in months): 2. THAAD; Contract start: Aug-00; Date of the last realignments[A]: May-08; Total number realignments: 5; Total contract value increase due to realignments: $1,179,000,000; Total period of performance increase due to realignments (in months): 15; Average time between realignments: 1 year 7 months; Average contract value increase per year due to realignments: $146,000,000[C]; Average period of performance increase per year due to realignments (in months): 2. Targets and Countermeasures; Contract start: Dec-03; Date of the last realignments[A]: June-08; Total number realignments: 1; Total contract value increase due to realignments: $41,300,000; Total period of performance increase due to realignments (in months): 0; Average time between realignments: 4 years 9 months; Average contract value increase per year due to realignments: $8,694,737; Average period of performance increase per year due to realignments (in months): 0. Total; Total number realignments: 31; Total contract value increase due to realignments: $13,857,611,411. Source: GAO analysis of MDA data. [A] Dates for some elements reflect when realignments were completed and not necessarily when realignments were incorporated into cost and schedule performance reporting. [B] The GMD program began a restructure during the fiscal year that includes a proposal to add between $350 million and $580 million to the contract value as well as 39 months to the period of performance. Since the replan is still ongoing and has not yet been placed on contract, this information is not totaled in the table above. [C] The realignment data provided for the KEI and THAAD programs included rounding for contract value increases; therefore we have rounded the average contract value increase per year since contract start to reflect this. [End of table] Some programs realigned more often than others. For example, GMD realigned work under its contract every year on average since its contract start in 2001, adding nearly 4 months and close to $1.1 billion to the time and value of the work under the contract with each realignment. KEI realigned its contract about every 14 months on average, adding more than $345 million and 4 months every year. Since contract start in 1996, ABL also added more than $218 million to the value of the work under its contract every year on average. Additionally, ABL adds approximately 7 months to its period of performance every year on average--more than any other element. During fiscal year 2008, 5 of 10 BMDS elements performed a realignment- -KEI, Targets and Countermeasures, GMD, THAAD, and STSS. The KEI replan in April 2008 reflected an 8-month delay to the booster flight test date because of technical issues experienced by the program over the past 2 years. Since the replan, the booster flight test has been further delayed to the fourth quarter of fiscal year 2009. However, during the replan, the program did not extend the period of performance or add value to the work under the contract. In June 2008, a delivery order under the Targets and Countermeasures element that is developing a new family of targets--the FTF--performed a rebaseline adding more than $41 million to the value of the work under the contract but not extending the period of performance. The program changed major milestone delivery dates as a result of manufacturing delays for some systems, caused principally by qualification program failures, subsequent redesigns, and requalification efforts. GMD, THAAD, and STSS added time and money to the value of the work under their contracts during the fiscal year.[Footnote 17] GMD's ongoing restructure includes a proposal to add between $350 million and $580 million to the value of the work under contract and more than 3 years to the period of performance. This ongoing restructure rephases and rescopes ongoing efforts to refine European capability requirements and to adjust program content as well as perform weapon system integration, perform flight test planning, and work to develop the two- stage booster among other tasks. During its realignment in May 2008, THAAD added approximately $80 million and 3 months citing cost effects from insufficient target availability. In October 2007, STSS replanned work citing funding constraints and the addition of STSS software upgrades. This resulted in the program changing its launch date from December 2007 to July 2008 and adding approximately $232 million to the value of the work under contract and 13 months to its period of performance. Since the replan, the program has further delayed launch of its demonstrator satellite to the third quarter of fiscal year 2009. MDA Contractors Overran Fiscal Year Cost and Schedule: Our analysis of contractor costs indicates that during fiscal year 2008, MDA contractors collectively overran budgeted costs by $152.4 million.[Footnote 18] These overruns occurred in 11 of 14 MDA contracts we reviewed, with the STSS contract accounting for more than 50 percent of the total.[Footnote 19] Based on cost performance during the fiscal year and using formulas accepted within the cost community, we estimate that at completion the cumulative overrun in the contractors' budgeted costs could be from about $2.0 billion to $3.0 billion. Our projections are based on the current budgeted costs at completion for each contract we assessed, which represents the total current planned value of the contract.[Footnote 20] However, the budgeted costs at completion, in some cases, have grown significantly over time. For example, the ABL contractor reported budgeted costs at completion totaling about $724 million in 1997, but as depicted in table 5, that cost has since grown to about $3.6 billion. Our assessment only reveals the overrun or underrun since the latest adjustment to the budget at completion. It does not capture, as cost growth, the difference between the original and current budgeted costs at completion. As a result, comparing the underruns or overruns for MDA programs in table 5 with cost growth on major defense acquisition programs is not appropriate because those major defense acquisition programs have established their full scope of work as well as developed total cost baselines, while these have not been developed for MDA programs. Our analysis is presented in table 5. Appendix II provides further details on the cost and schedule performance outlined in the table. Table 5: Prime Contractor Fiscal Year 2008 and Cumulative Cost and Schedule Performance: (Dollars in millions): ABL; Fiscal year 2008 cost performance[A]: (10.6); Fiscal year 2008 schedule performance[A]: 2.2; Cumulative cost performance: (84.8); Cumulative schedule performance: (23.6); Percentage of contract completed: 91.1; Estimated contract overrun/underrun at completion: Overrun of $89.7 to $95.4; Budget at completion: $3,626.7; Period of performance: Nov. 1996 - Feb. 2010. Aegis BMD Weapon System; Fiscal year 2008 cost performance[A]: (7.0); Fiscal year 2008 schedule performance[A]: (5.1); Cumulative cost performance: 0.0; Cumulative schedule performance: (8.4); Percentage of contract completed: 81.1; Estimated contract overrun/underrun at completion: Overrun of $1.9 to $12.2; Budget at completion: 1,247.0; Period of performance: Oct. 2003 - Sept. 2010. Aegis BMD SM-3 CLIN 9 (20 Block 1A missiles)[B]; Fiscal year 2008 cost performance[A]: (3.9); Fiscal year 2008 schedule performance[A]: 3.9; Cumulative cost performance: 2.3; Cumulative schedule performance: (0.1); Percentage of contract completed: 94.2; Estimated contract overrun/underrun at completion: Underrun of $7.5; Budget at completion: 179.0; Period of performance: Aug. 2006 - Aug. 2008. Aegis BMD SM-3 CLIN 1 (27 Block 1A missiles)[C]; Fiscal year 2008 cost performance[A]: 3.0; Fiscal year 2008 schedule performance[A]: (7.6); Cumulative cost performance: 3.3; Cumulative schedule performance: (7.0); Percentage of contract completed: 46.3; Estimated contract overrun/underrun at completion: Underrun of $6.6 to overrun of $0.7; Budget at completion: 237.5; Period of performance: May 2007 - Apr. 2010. C2BMC; (Dollars in millions): [Empty]; Fiscal year 2008 cost performance[A]: (9.8); Fiscal year 2008 schedule performance[A]: (3.6); Cumulative cost performance: (24.3); Cumulative schedule performance: (7.1); Percentage of contract completed: 71.1; Estimated contract overrun/underrun at completion: Overrun of $37.1 to $76.8; Budget at completion: 1,040.0; Period of performance: Jan. 2002 - Dec. 2009. GMD; (Dollars in millions): [Empty]; Fiscal year 2008 cost performance[A]: 53.9; Fiscal year 2008 schedule performance[A]: (77.4); Cumulative cost performance: (1,027.9); Cumulative schedule performance: (130.3); Percentage of contract completed: 84.0; Estimated contract overrun/underrun at completion: Overrun of $950.2 to $1,251.3; Budget at completion: 14,934.9; Period of performance: Jan. 2001 - Dec 2011. KEI[D]; (Dollars in millions): [Empty]; Fiscal year 2008 cost performance[A]: (8.3); Fiscal year 2008 schedule performance[A]: (8.5); Cumulative cost performance: (2.6); Cumulative schedule performance: (21.3); Percentage of contract completed: 13.9; Estimated contract overrun/underrun at completion: N/A; Budget at completion: 6,068.3; Period of performance: Dec. 2003 - Oct 2014. MKV Task Order 6 (Prototype Carrier Vehicle Seeker)[E]; Fiscal year 2008 cost performance[A]: (1.4); Fiscal year 2008 schedule performance[A]: (1.5); Cumulative cost performance: (1.1); Cumulative schedule performance: (0.6); Percentage of contract completed: 78.3; Estimated contract overrun/underrun at completion: Overrun of $1.6 to $2.5; Budget at completion: 19.3; Period of performance: Nov. 2006 - May 2009. MKV Task Order 7 (Engagement Management Algorithms)[E]; Fiscal year 2008 cost performance[A]: 1.4; Fiscal year 2008 schedule performance[A]: 0.0; Cumulative cost performance: 1.7; Cumulative schedule performance: 0.1; Percentage of contract completed: 52.8; Estimated contract overrun/underrun at completion: Underrun of $3.9 to $3.2; Budget at completion: 43.9; Period of performance: Dec. 2006 - May 2010. MKV Task Order 8 (Hover Test Bed)[E]; Fiscal year 2008 cost performance[A]: (10.7); Fiscal year 2008 schedule performance[A]: (0.0); Cumulative cost performance: (10.3); Cumulative schedule performance: 0.3; Percentage of contract completed: 81.4; Estimated contract overrun/underrun at completion: Overrun of $5.7 to $13.8; Budget at completion: 48.0; Period of performance: Dec. 2006 - Jan 2009. Sensors; Fiscal year 2008 cost performance[A]: (2.2); Fiscal year 2008 schedule performance[A]: (27.4); Cumulative cost performance: 22.0; Cumulative schedule performance: (9.6); Percentage of contract completed: 80.7; Estimated contract overrun/underrun at completion: Underrun of $25.0 to overrun of $9.1; Budget at completion: 1,125.2; Period of performance: Mar. 2003 - Dec 2010. STSS[F]; Fiscal year 2008 cost performance[A]: (87.9); Fiscal year 2008 schedule performance[A]: 1.9; Cumulative cost performance: (319.3); Cumulative schedule performance: (17.8); Percentage of contract completed: 53.2; Estimated contract overrun/underrun at completion: Overrun of $621.7 to $1,157.9; Budget at completion: 1,603.0; Period of performance: Apr 2002 - Sept. 2011. Targets and Countermeasures; Fiscal year 2008 cost performance[A]: (35.7); Fiscal year 2008 schedule performance[A]: 23.2; Cumulative cost performance: (52.8); Cumulative schedule performance: (6.4); Percentage of contract completed: 84.5; Estimated contract overrun/underrun at completion: Overrun of $63.7 to $75.9; Budget at completion: 1,056.4; Period of performance: Dec. 2003 - Dec 2009. THAAD[G]; Fiscal year 2008 cost performance[A]: (33.5); Fiscal year 2008 schedule performance[A]: (7.4); Cumulative cost performance: (228.7); Cumulative schedule performance: (16.5); Percentage of contract completed: 91.4; Estimated contract overrun/underrun at completion: Overrun of $252.0 to $274.0; Budget at completion: 4,649.4; Period of performance: Aug. 2000 - Sept. 2009. Total; Fiscal year 2008 cost performance[A]: (152.4); Fiscal year 2008 schedule performance[A]: (107.4); Cumulative cost performance: (1,722.5); Cumulative schedule performance: (248.3); Percentage of contract completed: [Empty]; Estimated contract overrun/underrun at completion: Overrun of $1,980.8 to $2,959.0. Source: Contract Performance Reports (data); GAO (analysis). Note: Comparing the percentage of total overrun to total budget at completion for MDA contracts with percentage of total cost growth for major acquisition defense programs that are past milestone B is not appropriate because the major defense acquisition programs have established their full scope of work as well as developed total cost baselines, while these have not been developed for MDA programs. [A] Cost performance here is defined as the difference between the budget for the work performed and the actual cost of work performed; while the schedule performance is the difference between the budgeted cost of planned work and the budgeted cost of work performed. Negative cost performance (budget overruns) and negative schedule performance (less work performed than planned) are shown with parentheses around the dollar amounts. [B] The Aegis BMD SM-3 contractor began work on contract line item number (CLIN) 9 in February 2007 that concluded in August 2008 for the acquisition of an additional 20 SM-3 Block 1A missiles. All corresponding analysis is based on data through August 2008. [C] The Aegis BMD SM-3 contractor began reporting performance on CLIN 1 in August 2007. This CLIN is for the production of a fourth lot of 27 Block 1A missiles. [D] We could not estimate the likely outcome of the KEI contract at completion because a trend cannot be predicted until 15 percent of the planned work is complete. [E] Out of the five task orders open during fiscal year 2008, there was sufficient cost performance data to report on the three listed above. [F] The STSS contract includes line items for work that do not necessarily apply to the program being launched in the third quarter of fiscal year 2009. Removing these line items from our analysis, the program's contract would be considered 78% complete. [G] Earned Value data for the THAAD contract is reported under two CLINs, 1 and 10. We report only the contractor's cost and schedule performance for CLIN 1 because it represents the majority of the total work performed under the contract. CLIN 10 provides for Patriot Common Launcher initiatives funded by the Army's Lower Tier Program Office. [End of table] Technical difficulties caused most elements to overrun their fiscal year 2008 budgeted costs. For example, STSS attributed most of its overrun of approximately $87.9 million to hardware problems on the program's second space vehicle including the flight communication box and the main spacecraft computer. The box overheated during testing which required a thorough test of the unit. Upon successful completion of this testing, it was determined it did not require a replacement. In addition, the program had a failure in the main spacecraft computer for which the program office initially recommended the removal of the entire computer from the spacecraft. However, after extensive research and testing, the program manager determined that the event with the spacecraft was an unverifiable failure with a low probability of occurrence and low mission impact and decided not to remove the computer from the spacecraft to resolve the issue. However, as a result of these issues, the launch was delayed from April 2008 to at least the third quarter of fiscal year 2009. The MKV Task Order 6 and ABL contractors also experienced technical difficulties. The MKV contractor for Task Order 6 reported cost overruns during the fiscal year of $1.4 million due mostly to software development issues and late delivery of government-furnished components. ABL's fiscal year cost overruns of $10.6 million were mainly related to late deliveries of key laser system components and the acquisition or refurbishment of the Beam Control/Fire Control system components. For example, a major component of the laser system required redesign and fabrication, delaying planned delivery and installation onto the aircraft. Also, multiple Beam Control/Fire Control hardware components either were not refurbished to specification or failed initial testing, delaying delivery and integration testing. The overall effect was an approximate 1-month slip that the contractor believes will be made up in time to make the current lethality demonstration planned for the end of fiscal year 2009. Three elements' contracts--Aegis BMD's contract for 27 SM-3 Block 1A missiles, MKV Task Order 7, and GMD--performed below their fiscal year budgeted costs by nearly $58.3 million with the GMD element accounting for approximately $53.9 million of that. The GMD element's underruns occurred partially because the contractors delayed or eliminated some planned work. For example, the GMD program did not emplace the three GBIs it expected to in fiscal year 2008 or conduct either of its two planned flight tests as scheduled during the fiscal year. As a result, it employed less labor than originally intended. Drivers for the MKV Task Order 7 contract's fiscal year cost underruns include a restructuring of the effort and decisions to use one rather than several approaches for coordinated attack, and using less manpower than originally planned in its procurement and software efforts. Lastly, the Aegis BMD contract for 27 SM-3 Block 1A missiles underran its fiscal year 2008 budget by approximately $3 million due in part to spending less than planned for engineering efforts with the missile's third stage component as well as adjustments made in program management, labor efficiencies, and material transfers in the missile's fourth stage component. While Some Tests Succeeded, Others Were Deferred; Overall System Performance Cannot Yet Be Assessed: Although several tests showed progress in individual elements and some system-level capabilities, all BMDS elements experienced test delays and shortfalls, in part due to problems with the availability and performance of target missiles. Most significantly, GMD was unable to conduct either of its planned intercept attempts during fiscal year 2008, however it was able to conduct one delayed intercept test in December 2008. As a result, key performance capabilities of the current configuration of the GMD kill vehicles may not be demonstrated and the new configuration is being fielded prior to flight testing. As a consequence of testing problems, none of the six MDA Director's test knowledge points for 2008 were achieved. Poor performance of targets continues to be a problem that caused several tests to either fail in part or in whole. Shortfalls in testing have delayed validating the models and simulations that are used to assess the overall performance of the BMDS as a whole. Consequently, comprehensive assessments of the capabilities and limitations of the BMDS are not currently possible and therefore MDA still does not have the capability to model or simulate BMDS capability from enemy missile launch to its engagement. Test, Targets, and Performance Challenges Continued during Fiscal Year 2008 for Several Elements: During fiscal year 2008, all BMDS elements experienced delays in conducting tests, most were unable to accomplish all objectives, and performance challenges continued for many. Moreover, the inability of MDA to conduct its full fiscal year 2008 flight test campaign as planned precluded the agency from collecting key information specified by the Director, MDA--known as Director's test knowledge points--to make certain decisions at critical points in some BMDS programs. Table 6 below summarizes test results and target performance for BMDS elements during the fiscal year. Table 6: Test and Targets Issues: Element: ABL; Tests/activities conducted as scheduled: No; All objectives achieved: Yes; Target issues: N/A. Element: Aegis BMD; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: Target availability delayed key test from 2008 until at least the third quarter fiscal year 2009. Element: C2BMC; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: N/A. Element: GMD; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: Target failed to release countermeasures during December 2008 flight test--FTG-05.[A] Element: KEI; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: N/A. Element: MKV; Tests/activities conducted as scheduled: No; All objectives achieved: No[B]; Target issues: N/A. Element: Sensors; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: Target failed to release countermeasures during July 2008 testing (FTX-03). Element: STSS; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: N/A. Element: Targets and Countermeasures; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: FTF delivery delayed and experienced cost growth. Element: THAAD; Tests/activities conducted as scheduled: No; All objectives achieved: No; Target issues: Target experienced anomaly during a September flight test resulting in a no-test. Sources: GAO (presentation); MDA (data). [A] This flight test was originally scheduled for fiscal year 2008, but was later executed in fiscal year 2009. [B] The MKV program was able to achieve its objective in the first quarter of fiscal year 2009. [End of table] As a result of test delays, MDA restructured its flight test plan for fiscal year 2009, increasing the number of tests and compressing the amount of time to analyze and prepare for subsequent tests. For example, MDA plans to conduct 14 of 18 flight tests in the third and fourth quarter of fiscal year 2009. MDA's past performance raises questions about whether this is realistic. In fiscal year 2008, MDA had planned to conduct 18 flight tests, but it only accomplished 10, plus it had several flight tests delayed into 2009 from 2008. An MDA official acknowledged that the 2009 plan is aggressive, but stated that it can be achieved. Specifics of each element's testing experience during fiscal year 2008 follow. According to Aegis BMD officials, budgetary constraints prompted the Aegis BMD element to delay some tests, reducing the number of tests planned for 2008. However, the program was able to successfully complete its first test involving two non-separating targets, conduct a short-range ballistic missile intercept, and participate in a THAAD intercept during the fiscal year. The program also planned to participate in a BMDS-level ground test during the year, but the test was delayed until at least the second quarter of fiscal year 2009 because of real-world events. Finally, Aegis BMD standard missile flight tests showed that interoperability issues persist between THAAD and Aegis BMDS with respect to correlation and object reporting. ABL experienced delays during fiscal year 2008, but achieved all of its primary test objectives. The program planned to complete the installation of its high energy laser on the aircraft by June 2008 in preparation for testing. However, it was not completed until August 2008 because of problems with activating some of the laser's subsystems. The program delayed the final testing of the laser until the problems could be resolved. Once the problems were resolved, the ABL program was able to complete functionality testing of the laser in September 2008. C2BMC experienced delays in conducting tests, but achieved several test objectives. For example, software upgrade verification testing slipped from fiscal year 2008 to 2009 but the program was able to participate in many other system-level ground and flight tests during the year that enabled the program to demonstrate multiple capabilities, including situational awareness and sensor management.[Footnote 21] The C2BMC element extended development for its next software release, 6.4, by more than a year because of delays in system-level BMDS testing and challenges in developing the network server for version 6.2, as well as unplanned work to incorporate effects from earth rotation in the 6.4 C2BMC planning architecture. C2BMC added earth rotation effects to address a requirement that Spiral 6.2 and later releases have the ability to model the true extent of ranges for long-range threats. Finally, C2BMC is still developing its capability to generate a single track from multiple sensors through a new resource management function, the Global Engagement Manager. For example, the development team for this function had to modify the design of this function's new track processing that experienced an unacceptable level of delays when processing data. In fiscal year 2008, the GMD program was unable to conduct either of its two planned intercept attempts--FTG-04 and FTG-05. MDA first delayed and then later canceled the FTG-04 test in May 2008 due to a problem with a telemetry component in the interceptor's Exoatmospheric Kill Vehicle (EKV) needed to gather test data. MDA also delayed FTG-05 from fiscal year 2008 and conducted it in December 2008. Over the past two years MDA had expected to conduct seven GMD interceptor flight tests by the end of the first quarter of fiscal year 2009. However, MDA was only able to conduct two, as shown in figure 3. Figure 3: GMD Reduction in Flight Tests from January 2006 to March 2010: [Refer to PDF for image: illustration] As of September 2005: Integrated flight tests planned: FY06, Q1: FT-1 (achieved); Type: CE I EKV; FY06, Q4: FTG-2 (achieved); Type: CE I EKV; FY07, Q1: FTG-3; Type: CE I EKV; FY07, Q3: FTG-4; Type: CE I EKV; FY07, Q4: FTG-5; Type: CE I EKV; FY08, Q1: FTG-6; Type: CE II EKV New processor; FY08, Q2: FTG-7; Type: CE I EKV; FY08, Q4: FTG-8; Type: CE I EKV; FY09, Q1: FTG-9. Type: CE I EKV and CE II EKV New processor. As of January 2009: Integrated flight tests planned: FY06, Q1: FT-1 (achieved); Type: CE I EKV; FY06, Q4: FTG-2 (achieved); Type: CE I EKV; FY07, Q4: FTG-3a (achieved); Type: CE I EKV; FY09, Q1: FTG-5 (achieved); Type: CE I EKV; FY09, Q4: FTG-6. Type: CE II EKV New processor. Source: GAO analysis of MDA data. [End of figure] The cancellation of FTG-04 raised concerns within the test community and members of Congress. FTG-04 was at first delayed and then canceled. MDA replaced it with a test to assess sensor capability--FTX-03. The sensor test allowed GMD to verify fire control software and integration with multiple sensors. The DOT&E was not consulted on the decision to cancel FTG-04 and expressed concern that the elimination of any intercept test reduced the opportunity to gather data that might have increased confidence in the models and simulations. In the conference report accompanying the National Defense Authorization Act for Fiscal Year 2008, conferees expressed concern about the loss of the FTG-04 flight test and requested that we review the circumstances and the effects on the BMDS. Details of our review of the FTG-04 flight test cancellation appear in appendix III. Because GMD conducted FTG-05 in December 2008, there are only two full sets of GMD intercept data to date available for analysis which limits the ability to verify and validate the models and simulations.[Footnote 22] Additionally, FTG-04 and the subsequent test--FTG-05--were planned to present different stresses to the kill vehicle which would provide critical data needed to further verify the fielded configuration of the kill vehicle. The cancellation and subsequent restructuring of the first test caused a delay in FTG-05 from the third quarter of fiscal year 2008 until December 2008. In the FTG-05 test, the interceptor hit its intended target. However, MDA judged the target as a failure because it failed to release its countermeasures as planned. Consequently, all primary test objectives were not achieved. Looking forward to the next GMD intercept flight test--FTG-06 in at least the fourth quarter of fiscal year 2009--MDA is accepting a higher level of risk than it previously expected in conducting this first test of the CE-II EKV because it will contain several objectives that had planned to be previously tested, but have not been. MDA had set up an approach to test one new major component change at a time. For example, MDA had planned to test the CE-I EKV first against simple targets, then the CE-I against a complex target, and once that had been proven MDA planned to test the CE-II EKV against a complex target. However, MDA was not able to test the CE-I EKV against a complex target due to a target failure. Due to testing problems, GMD has only been able to assess the CE-I EKV with a target without countermeasures. As a result, the FTG-06 flight test will be the first GMD test assessing both a CE- II EKV and a complex target scene. Adding to the risk, this will be only the second test using a newly developed FTF LV-2 target. During fiscal year 2008, the KEI program experienced problems during testing that required it to rework components which, in turn, caused a delay to subsequent testing. More importantly, due to technical issues experienced by the program over the past two years, the first booster flight test--a key decision point for the program--has been delayed by nearly a year and is not scheduled to occur until at least the fourth quarter of fiscal year 2009. Technical difficulties delayed the MKV program's fiscal year 2008 hover test until fiscal year 2009. This hover test will allow the program to integrate and test key components of the system in a repeatable ground-based free flight environment as their technologies reach maturity. Although originally planned for the fourth quarter of fiscal year 2008, the test was successfully conducted in December 2008. The STSS program encountered problems during testing that forced the program to delay the launch of its demonstration satellites from April 2008 to at least the third quarter of fiscal year 2009. The program continued to experience technical difficulties with its space vehicles. For example, during testing, the program experienced problems with its main spacecraft computer as well as an overheating flight communications box. After extensive testing, the program determined that these components were acceptable for flight. Similarly, in fiscal year 2008, the Sensors element also experienced flight test delays as well as difficulties in achieving planned objectives due to target performance, but met some primary objectives. The element successfully participated in other tests during the fiscal year which demonstrated the ability for the sensors to acquire and track a target. One test event, FTX-03, provided the first opportunity for four key sensors--Sea-based X-band radar, AN/TPY-2, Upgraded Early Warning Radar, and an Aegis BMD radar--to operate in a more operationally realistic test scenario.[Footnote 23] This test demonstrated the capability for the sensors to correlate target information in order to conduct an intercept test. However, the target failed to release its countermeasures as planned. This failure precluded sensors from assessing capability against a dynamic lethal target scene with countermeasures. As a result, the sensors could not collect all of the expected data, which delayed the element's ability to develop algorithms needed for the discrimination capability. These objectives will need to be addressed in future testing. The BMDS Operational Test Agency has had ongoing concerns regarding the formatting, tracking, and accounting of messages from GMD sensors. [Footnote 24] The timely reception of messages from sensors to weapon systems is key to support decisions and achieve effective intercepts. Since 2000 the BMDS Operational Test Agency has reported these concerns to MDA about poor data collection and management practices involving sensors affecting its assessment of tests. These data management problems prevented the analysis of message data, according to BMDS Operational Test Agency officials. In response, the contractor proposed a message monitoring system among communications nodes. Consequently, MDA recommended that this issue be closed out, but the BMDS Operational Test Agency still considers the matter to be open because GMD has not funded the monitoring system. THAAD planned to conduct three intercept attempts, but due to a target failure, it was only able to conduct two. The program could not complete its final flight test of the fiscal year because the target experienced an anomaly during flight. The test was planned to be a BMDS- level event and was designated as a developmental test/operational test mission utilizing multiple BMDS elements and operationally realistic criteria. The program also expected to demonstrate that it could launch more than one THAAD interceptor during the engagement. Program officials rescheduled this test for the second quarter of fiscal year 2009. In addition, THAAD's radar data collection test, RDC-2, was planned for 2008 but was deleted due to target availability and funding. As a result, program officials told us that these test objectives will be covered in the future with hardware-in-the-loop simulations and other radar events. The program successfully completed its first two planned tests for the fiscal year. In October 2007, THAAD successfully demonstrated an intercept of a target outside the earth's atmosphere. This was the first time THAAD had successfully conducted an intercept outside of the atmosphere since 1999. Additionally, in June 2008, THAAD completed a successful intercept of a separating target. This intercept utilized warfighter procedures developed by the U.S. Army Air Defense School. Key MDA Test Knowledge Points Not Achieved: As a consequence of flight test delays as well as a delay in a key ground test, MDA was unable to achieve any of the Director's test knowledge points scheduled for fiscal year 2008 as shown in table 7. Table 7: Status of Fiscal Year 2008 Director's Test Knowledge Points: Knowledge point: Assess Capability to Deliver Real-Time Engagement Tracks; Knowledge gained: Verification of initial Global Engagement Manager capability to support BMDS-level sensor/weapon system pairing; Flight and ground test: GTD-03[A]; Original completion: 4th Quarter 2008; Current projection: 2nd Quarter 2009. Knowledge point: Verify 72-inch Flexible Target Family; Knowledge gained: Confirmation of 72" performance. Viability of FTF concept to efficiently configure and transport target to launch facility. Confidence to discontinue use of STARS; Flight and ground test: FTM-15; Original completion: 4th Quarter 2008; Current projection: 3rd Quarter 2009. Knowledge point: Demonstrate High-Acceleration Booster; Knowledge gained: Confirmation of Boost Phase Capability alternative to ABL and High Acceleration Booster for Midcourse Defense (mobile and fixed sites); Flight and ground test: FTK-01; Original completion: 4th Quarter 2008; Current projection: 4th Quarter 2009. Knowledge point: Confirm Constellation Affordability; Knowledge gained: Space sensor performance against operationally realistic targets confirmed with existing Block 06 technology (anchors performance-cost baseline for future STSS); Flight and ground test: FTS-01; Original completion: 4th Quarter 2008; Current projection: 4th Quarter 2009. Knowledge point: Verify Capability to Conduct Launch on Tactical Digital Information Link BM Engagement; Knowledge gained: Assessment of Aegis BMD 3.6 and SM-3 Block IA performance and ability to successfully engage and intercept a long- range ballistic missile target and to use an off-board sensor's track data via Link-16 to initiate that engagement; Flight and ground test: FTM-15; Original completion: 4th Quarter 2008; Current projection: 3rd Quarter 2009. Knowledge point: Confirm Constellation Performance; Knowledge gained: Space sensor performance against operationally realistic targets confirmed with existing Block 06 technology (anchors performance-cost baseline for future STSS); Flight and ground test: FTS-03; Original completion: 4th Quarter 2008; Current projection: To Be Determined. Source: GAO analysis of MDA data. [A] GTD-03 was delayed to accommodate a real-world contingency as requested by the warfighter. [End of table] In May 2007, the Director, MDA, established key system-level and element-level knowledge points to provide critical information for making key decisions regarding the BMDS. According to MDA, these knowledge points are unique management approaches chosen to manage MDA's critical program risks.[Footnote 25] Each knowledge point is based on an event that provides critical information--or knowledge--for a key MDA decision requiring the Director's approval. In fiscal year 2008, among the Director's test knowledge points delayed, MDA had to defer the confirmation of the 72" target performance due to delays in qualifying components. Additionally, MDA had to delay the confirmation of the booster for the KEI program as problems were encountered during testing of the nozzle. Poor Target Missile Performance Continues to Hamper BMDS Testing: While targets have caused problems in fiscal year 2008 testing, poor performance of targets is not new. Targets' reliability and availability problems have significantly affected BMDS development and testing since 2006, and issues have grown even more problematic in recent years. Although target anomalies and failures have affected many of the missile defense elements, THAAD and GMD have been most affected. In 2006, the THAAD program was unable to achieve its first intercept attempt (FTT-04) because the target did not function properly. In 2007, two THAAD radar characterization tests (RDC-1c and RDC-1d) were unsuccessful due to target anomalies. These tests flew targets with characteristics needed for radar observation in support of advanced discrimination algorithm development. However, target problems prevented an opportunity for the radar to exercise all of the planned algorithms, causing a loss of expected data. In addition to target failure issues, the THAAD program deferred some flight tests because targets were not available, which cost the program about $201 million. GMD also experienced similar long-term effects on its flight test schedule when it was unable to achieve primary test objectives in a 2007 intercept attempt (FTG-03) due to a target failure. MDA's existing targets are becoming less capable of meeting requirements for near-term flight tests. These targets are aging and likely to grow even less reliable with time; some components, such as the rocket motors, are more than 40 years old. Among other things, MDA's Targets and Countermeasures program office has also had problems incorporating requirements into contracts and has experienced problems obtaining supplies as vendors left the market due to the lack of business. To address the growing need for more sophisticated and reliable targets for the future BMDS test program, MDA was developing a new family of targets called the FTF, which was originally intended to be a family of new short, medium, and long-range targets with ground-, air-, and sea- launch capabilities. MDA embarked on this major development without estimating the cost to develop the family of target missiles. MDA proceeded to develop and even to produce some FTF targets without a sound business case and, consequently, their acquisition has not gone as planned.[Footnote 26] The funds required for the FTF were spent sooner than expected and were insufficient for the development. Getting the FTF target's components through the qualification process, however, was more difficult and costly than the program expected. For example, MDA originally planned to launch the first FTF target--a 72-inch LV-2- -in a 2008 STSS flight test, but the test was rescheduled due to delays in satellite integration and target affordability and availability. While many of the target missile's components are found on existing systems, their form, fit, function, and the environment they fly in have been changed for the 72-inch LV-2 target. Consequently, many critical components initially failed shock and vibration testing and other qualification tests and had to be redesigned. The process was recently scheduled to be complete in early October 2008 but, after several delays, was not finished until December 2008. Despite this, MDA expects the target to be complete and ready for its first launch in a third quarter fiscal year 2009 Aegis BMD flight test (FTM-15). We recently reported that the FTF has been delayed, costs have increased and exceeded $1 billion, and the program's scope has been reduced.[Footnote 27] Work on all but one of the FTF target variants, the 72-inch LV-2, was canceled in June 2008, including plans for development and production of the second type of FTF target, the 52- inch, originally scheduled to launch in 2009. With guidance from the Missile Defense Executive Board, MDA is currently conducting a comprehensive review of the targets program to determine the best acquisition strategy for future BMDS targets. It is expected to be completed in mid-2009. Whether or not MDA decides to restart the acquisition of the 52-inch targets, or other FTF variants, and the nature of those targets depends on the results of this review. Currently, MDA has narrowed its FTF development efforts, focusing on a single vehicle, the 72-inch LV-2 ground-launched target. The first launch was supposed to determine the viability of the FTF concept and the feasibility of discontinuing the use of existing targets. However, rather than first conducting a separate developmental test to confirm the target's capability, MDA has chosen a much riskier approach. The first launch of the new LV-2 target will be in an Aegis BMD intercept test. Aegis BMD originally planned to use this new target in a fiscal year 2008 flight test; however, because the target was not ready, the test is delayed until at least the third quarter of fiscal year 2009. Repeated target problems and test cancellations have also affected the development of capabilities needed to discriminate the real target from countermeasures. Without opportunities to test the functionality of the software, there now is a system-level shortfall in BMDS progress toward developing a target discrimination capability against more sophisticated countermeasures in the midcourse phase of flight. In order to improve the effectiveness of the BMDS against evolving threats, MDA elements are developing advanced discrimination software in their component's sensors. The advanced discrimination software is critical to distinguish the threat re-entry vehicle from associated countermeasures and debris. Target failures during tests prevented opportunities to gather data to assess how well discrimination software performs in an operational environment. Overall Performance of BMDS Can Not Yet Be Assessed: MDA's modeling and simulation program enables MDA to assess the capabilities and limitations of how BMDS performs under a wider variety of conditions than can be accomplished through the limited number of flight tests conducted. Flight tests alone are insufficient because they only demonstrate a single collection data point of element and system performance. Flight tests are, however, an essential tool used to both validate performance of the BMDS and to anchor the models and simulations to ensure that they accurately reflect real performance. Computer models of individual elements replicate how those elements function. These models are then combined into various configurations that simulate the BMDS engagement of enemy ballistic missiles. To ensure confidence in the accuracy of modeling and simulation in representing BMDS capability, the program goes through a process called accreditation.[Footnote 28] Element models are validated individually using flight and other test data and accredited for their intended use. MDA intends to group these models into system-level representations according to user needs. One such grouping is the annual performance assessment, a system-level end-to-end simulation that assesses the performance of the current BMDS configuration.[Footnote 29] The performance assessment integrates element-specific models into a coherent representation of the BMDS. Performance assessments are used to: * assess objectives from MDA's Deputy of Engineering and the BMDS Operational Test Agency,[Footnote 30] * support MDA decisions about engagement sequence group capability deliveries, and: * support MDA decisions about BMDS fielding and declaring capabilities. Fundamentally, performance assessments anchored by flight and ground tests are a comprehensive means to fully understand the performance capabilities and limitations of the BMDS. Developing an end-to-end system-level model and simulation has been difficult. BMDS Operational Test Agency officials told us that they do not anticipate a fully accredited, system-level model and simulation capability to be available until 2011. MDA's first effort to bring together different element models and simulations to produce a fully accredited, end-to-end model and simulation for Performance Assessment 2007 was unsuccessful primarily because of inadequate data for verification and validation to support accreditation and a lack of common threat and environment input data among element models. Though Performance Assessment 2007 was a success in establishing a capability for integrated modeling and simulation in a short time frame, it was unsuitable to assess system-level performance due to low confidence from a lack of accreditation. Consequently, acting on a joint recommendation between MDA and the Operational Test Agency, MDA officials canceled their 2008 performance assessment efforts in April 2008 because of developmental risks associated with modeling and simulations, focusing instead on testing and models for Performance Assessment 2009. MDA officials believe that the refocused efforts will increase the chances for success during Performance Assessment 2009. According to the BMDS Operational Test Agency's January 2009 Modeling and Simulation Accreditation Report, confidence in MDA's modeling and simulation efforts remains low although progress was made during the year. MDA is now exercising stronger central leadership to provide guidance and resources as it coordinates the development of verified and validated models and simulations, as recommended by a 2004 Defense Science Board study. MDA and element officials are now working more closely with the BMDS Operational Test Agency. For example, MDA and the BMDS Operational Test Agency have agreed on performance parameters and criteria used to validate element models and simulations. Nonetheless, BMDS Operational Test Agency officials stated that there are several weaknesses in the BMDS testing program such as: * Insufficient consideration of modeling and simulation requirements in MDA flight test plans, though they emphasized that MDA is finalizing a list of such parameters for future flight test plans, * Use of artificialities in flight tests which limit the realism of scenarios for anchoring models and simulations,[Footnote 31] and: * Inadequate test planning for comprehensive modeling of weather conditions.[Footnote 32] MDA intends to verify and validate models and simulations by December 2009 for Performance Assessment 2009. However, BMDS Operational Test Agency officials stated that there is a high risk that the Performance Assessment 2009 analysis will be delayed because of remaining challenges and MDA's slow progress in accreditation, as follows: * The compressed schedule of ground and flight tests leaves little time for data analysis that is essential to anchor models to those tests, particularly for a complete analysis supporting MDA's Performance Assessment 2009. * Out of 40 models, the BMDS Operational Test Agency recommended in January 2009 full accreditation for only 6 models, partial accreditation for 9 models, and no accreditation for 25 models. * Because MDA canceled the follow-on Performance Assessment 2008, the BMDS Operational Test Agency did not receive verification and validation data that would have been included in the modeling and simulation portion of its 2008 operational assessment. BMDS Operational Test Agency officials told us that MDA also does not adequately plan for the collection of flight test data and post-flight reconstruction to support anchoring MDA models and simulations, even though post-flight reconstruction is needed to validate that models and simulations are adequate representations of the real world for their intended purpose.[Footnote 33] MDA guidance emphasizes that one of the primary objectives of the MDA ground and flight test program is to anchor BMDS models and simulations. Additionally, this guidance requires MDA's testing program to work with the MDA engineers to define a test program that anchors these models and simulations across the operating spectrum. According to BMDS Operational Test Agency officials, the first full post-flight reconstruction was conducted in December 2008. Despite the guidance delineating responsibilities for test data, MDA test plans currently do not include enough detail to allocate and synchronize resources in order to anchor models and simulations. MDA recently initiated a three-phase review of the entire BMDS test program. According to MDA, this three-phase review will emphasize the need for basing BMDS test planning and test design on critical factors that have not been proven to date and will drive target selection requirements. One outcome of the review will be to create integrated campaigns of ground and flight tests to efficiently collect data needed to validate the models and simulations. MDA intends to complete all three phases of the review by May 2009, after which MDA intends to have a date when all MDA models and simulations will be verified and validated. However, the current lack of flight test data for MDA's and BMDS Operational Test Agency analysis prevents the timely validation of models and simulations that are used to build the 2009 end-to-end performance assessment. Production, Fielding, and Declaration of Capabilities Proceed despite Delays in Testing and Assessments: In fiscal year 2008, MDA met most of its delivery goals. However, it continued to pursue a concurrent development, manufacturing and fielding strategy in which assets are produced and fielded before they are fully demonstrated through testing and modeling. Although flight tests and modeling and simulation produced less validation of performance than planned, MDA continued manufacturing untested components and declaring capabilities ready for fielding. For example, 10 of the new configuration kill vehicles for the GBI will have been manufactured and delivered before being flight-tested. MDA also declared that it had fielded 9 of 22 BMDS capabilities planned for 2008 (postponing 13), but due to test cancellations and performance assessment delays, it had to change the basis of these declarations, often relying on previous, less realistic testing. MDA Met Most 2008 Asset Delivery Goals: MDA achieved four of the five delivery goals it set for fiscal year 2008 as shown in the table 8. Table 8: BMDS Deliveries and Total Fielded Assets as of September 30, 2008: BMDS element: GMD; Fiscal year 2008 delivery goals: 3 interceptors; Assets delivered in fiscal year 2008: 0 interceptors; Total assets available (cumulative total of assets since 2005): 24 interceptors[A]. BMDS element: Sensors; Fiscal year 2008 delivery goals: 1 AN/TPY-2 radar; Sea-based X-band radar; Assets delivered in fiscal year 2008: 1 AN/TPY-2 radar; Sea-based X- band radar[B]; Total assets available (cumulative total of assets since 2005): 4 AN/TPY-2 radars[C]; Sea-based X-band radar. BMDS element: Aegis BMD; Fiscal year 2008 delivery goals: 20 SM-3 missiles; Assets delivered in fiscal year 2008: 20 SM-3 missiles; Total assets available (cumulative total of assets since 2005): 34 SM-3 missiles; 15 destroyers; 3 cruisers. BMDS element: C2BMC; Fiscal year 2008 delivery goals: 1 fielding and activation site; Assets delivered in fiscal year 2008: 1 fielding and activation site; Total assets available (cumulative total of assets since 2005): 6 suites; 31 Web browsers; 1 fielding and activation site; 46 enterprise workstations. Source: MDA (data); GAO (presentation). [A] The GMD program did not deliver any interceptors as planned in fiscal year 2008, but was able to deliver two interceptors--one in October 2008 and one in November 2008. Therefore, the cumulative total for GBIs as of December 2008 is 26. [B] Partial capability for the Sea-based X-band radar will be based on satisfying planned objectives for two tests scheduled for fiscal year 2009. [C] AN/TPY-2 radars were formerly known as Forward-Based X-Band- Transportable radars. According to MDA, an additional AN/TPY-2 radar has been provided and is undergoing Government ground testing. [End of table] The agency planned to deliver the Sea-based X-band radar and three additional GBIs for Block 1.0, 20 additional SM-3 missiles for its Block 2.0 capability, a C2BMC site for fielding and activation for its Blocks 3.1/3.2 and 5.0 capabilities, and an additional AN/TPY-2 radar.[Footnote 34] Although partial capability for the Sea-based X- band radar will not be declared until at least fiscal year 2009, it was approved for Early Capability Delivery in fiscal year 2008. The agency delivered the Aegis BMD SM-3 missiles, the AN/TPY-2 radar, and the C2BMC site in fiscal year 2008 as planned, but was unable to deliver the GBIs because the GMD element encountered development challenges with components for the CE-II EKV. In addition, the Navy Commander, Operational Test and Evaluation Force declared the Aegis BMD 3.6 system as operationally suitable and effective in October 2008. This decision signifies that 18 Aegis BMD-equipped ships and 90 SM-3 missiles are ready for transition to the Navy. Production and Fielding of BMDS Systems Getting Ahead of Testing: Despite developmental problems, test delays and MDA's inability to complete all fiscal year 2008 Director's test knowledge points, manufacturing, production, and fielding have proceeded close to schedule. In some cases fielding has gotten ahead of testing. For example, Aegis BMD expected to assess the ability of the SM-3 Block 1A missile to engage and intercept a long-range ballistic target to satisfy a Director's test knowledge point. Even though that test has been delayed until the third quarter of fiscal year 2009, MDA purchased 20 SM-3 Block 1As in fiscal year 2008. Furthermore, MDA intended to assess, through flight tests, the CE-I EKV's capability against scenarios that included complex target scenes with countermeasures. However, due to the frequent restructuring of its test plan and a target failure during its most recent flight test, the fielded configuration for GMD has not completed a test against countermeasures. According to MDA, no more CE-I flight tests have been approved, although the agency is considering additional flight testing of the CE-I EKV in the future. Moreover, earlier ground and flight testing, along with manufacturing discoveries prompted the GMD program to initiate a refurbishment program for the kill vehicles and the boosters. Refurbishment consists of: (1) reliability improvements to address high priority risks and to support the development and understanding of GBI reliability and (2) surveillance of aging through the examination of removed components. Consequently, the capability of the CE-I, including improvements designed to mitigate risk, as well as understand its capabilities and limitations against targets employing countermeasures may not be flight-tested, yet all 24 interceptors with this configuration are already emplaced and declared operational. More importantly, the GMD program continues to experience test delays, causing fielding to outpace flight tests as shown in figure 4. Figure 4: GMD Flight Test and Fielding Plan for Interceptors Comparison--September 2006 versus January 2009: [Refer to PDF for image: illustration] As of September 2005: Integrated flight tests planned: FY06, Q1: FT-1 (achieved); Type: CE I EKV; FY06, Q4: FTG-2 (achieved); Type: CE I EKV; FY07, Q1: FTG-3; Type: CE I EKV; FY07, Q3: FTG-4; Type: CE I EKV; FY07, Q4: FTG-5; Type: CE I EKV; FY08, Q1: FTG-6; Type: CE II EKV New processor; FY08, Q2: FTG-7; Type: CE I EKV; FY08, Q4: FTG-8; Type: CE I EKV; FY09, Q1: FTG-9. Type: CE I EKV and CE II EKV New processor. Fieldings: FY06, Q1: CE I EKV (achieved); FY06, Q2: CE I EKV (achieved); FY06, Q4: CE I EKV (2) (achieved); FY07, Q1: CE I EKV (3); FY07, Q2: CE I EKV (2); FY07, Q4: CE I EKV (2); FY08, Q1: CE I EKV (2); FY08, Q2: CE II EKV New processor; FY08, Q3: CE II EKV New processor (2); FY08, Q4: CE II EKV New processor (3); FY09, Q1: CE I EKV; FY09, Q2: CE II EKV New processor (3); FY09, Q3: CE II EKV New processor (3); FY09, Q4: CE II EKV New processor (3); FY10, Q1: CE II EKV New processor (3); FY10, Q2: CE II EKV New processor (2). As of January 2009: Integrated flight tests planned: FY06, Q1: FT-1 (achieved); Type: CE I EKV; FY06, Q4: FTG-2 (achieved); Type: CE I EKV; FY07, Q4: FTG-3a (achieved); Type: CE I EKV; FY09, Q1: FTG-5 (achieved); Type: CE I EKV; FY09, Q4: FTG-6. Type: CE II EKV New processor. Fieldings: FY06, Q1: CE I EKV (achieved); FY06, Q2: CE I EKV (achieved); FY06, Q4: CE I EKV (2) (achieved); FY07, Q1: CE I EKV (achieved); FY07, Q2: CE I EKV (2) (achieved); FY07, Q3: CE I EKV (3) (achieved); FY07, Q4: CE I EKV (5) (achieved); FY08, Q1: CE I EKV (2) (achieved); FY09, Q1: CE II EKV New processor (2) (achieved); FY09, Q2: CE II EKV New processor; FY09, Q3: CE II EKV New processor (3). Source: GAO analysis of MDA data. [End of figure] For example, the program has only been able to conduct two intercepts since 2006 for verifying the fielded configuration yet the production of interceptors continues. According to GMD's September 2006 flight test plan, for fiscal years 2007 and 2008, and the first quarter of fiscal year 2009 it was going to conduct seven flight tests, including a test that would utilize 2 GBIs against a single target--known as a salvo test[Footnote 35] --and field 16 new GBIs. By January 2009 GMD had changed its plan, removing the salvo test and conducting two flight tests, yet it fielded 13 GBIs. Similarly, GMD had planned to conduct an intercept test to assess the enhanced version of the EKV called the Capability Enhancement II in the first quarter of fiscal year 2008, months before emplacing any interceptors with this configuration. However, developmental problems with the new configuration's inertial measurement unit and the target delayed the first flight test with the CE-II configuration--FTG-06-- until at least the fourth quarter of fiscal year 2009. Despite these delays, MDA expects to have emplaced five CE-II interceptors before this flight test. MDA indicated that these five interceptors will not be declared operational until the satisfactory completion of the test and the Program Change Board declares their status. However, MDA projects that 10 CE-II EKVs will have been manufactured and delivered before that first flight test demonstrates the CE-II capability. This amounts to over half of the CE-II EKV deliveries that are currently on contract. MDA did not emplace the three GBIs it needed to meet its fiscal year 2008 fielding goals. MDA will have to emplace twice as many GBIs than planned in fiscal year 2009 before Block 1.0 can be declared complete. As of January 2009, the agency had emplaced two and must emplace four more in order to complete Block 1.0 as planned. Major defense and acquisition programs must complete operational test and evaluation before entering full-rate production.[Footnote 36] Because MDA considers the assets it has fielded to be developmental, it has not advanced BMDS elements to DOD's acquisition cycle or begun full- rate production. Therefore, MDA has not yet triggered the requirement for an operational test and evaluation prior to fielding. However, MDA's concurrent approach to developing and fielding assets has led to testing problems and concerns about the performance of some fielded assets. After two flight test failures in 2005, MDA undertook a Mission Readiness Task Force to establish confidence in GMD's ability to reliably hit its target, establish credibility in setting and meeting test event dates, build increasing levels of operationally realistic test procedures and scenarios, raise confidence in successful outcomes of flight missions, and conduct the next flight test as soon as practical within acceptable risk bounds. However, GMD accelerated the objectives for its test program after the first Mission Readiness Task Force flight test and the program continues to experience developmental problems with key interceptor components. MDA also separately established a refurbishment program designed to replace questionable interceptor parts and increase reliability of GBIs. Since 2006, we have reported that the performance of some fielded GBIs was uncertain. Despite MDA's previous efforts to build confidence in its test program, MDA continues to pursue a risky approach for fielding BMDS assets under its new block structure. In March 2008, we reported that MDA's new block structure did not address whether it would continue its practice of concurrently developing and fielding BMDS elements and components. However, in 2008 the agency continued to field assets without adequate knowledge. MDA emplaced GBIs during the year although its refurbishment program was barely underway, meaning that the risks of rework continue. To date, 26 GBIs have been emplaced--many of which may contain unreliable parts--and only a few have been refurbished since the initiation of the refurbishment program in 2007. According to program officials, some improvements have already been introduced into the manufacturing flow and demonstrated during flight testing. While it is always a concern when tests are eliminated or the complexity of a planned test is reduced, the concern is heightened for a system of systems such as the BMDS because of the complex interaction of components within an element, and between that element and the other elements within the BMDS. Consequently, the need to synchronize the development and testing of different capabilities is crucial before fielding begins. For example, for certain engagement scenarios, the ground-based interceptor will launch based on information provided by an entirely separate element such as an Aegis cruiser or destroyer. If a problem is discovered during these flight tests, post-flight reconstruction using models needs to be conducted, the root-cause must be determined, a solution or mitigation must be developed and implemented, and a new test to confirm the effectiveness of the solution or mitigation must be performed. Reduced Testing Has Delayed Some Capability Declarations and Weakened the Basis for Others: When MDA determines that a capability can be considered for operational use, it does so through a formal declaration. MDA uses an incremental declaration process to designate BMDS capability for its blocks in three levels--early, partial and full. The first two levels allow these BMDS features to play a limited role in system operations before they have attained their full level of capability. Each capability designation in the delivery schedule represents upgraded capacity to support the overall function of BMDS in its mission as well as the level of MDA confidence in the system's performance. Capability declarations are important because MDA uses them to assess progress toward block completion. MDA guidance calls for an orderly sequence of events that lead to declaring that a fielded capability has been achieved and is ready for consideration for operational use. MDA bases its declarations on, among other things, a combination of models and simulations--such as end-to-end performance assessments-- and ground tests all anchored to flight test data. Because performance assessments analyze the BMDS as an entire system in a variety of ways, they provide more comprehensive information than flight tests alone. These events and assessments build on each other every year as MDA adds capabilities by improving hardware and software. MDA decision makers would then declare the achievement of capability goals for engagement sequence groups based on performance assessments. While in some instances, declarations of capability have been deferred, in other instances MDA has declared capabilities despite shortfalls in testing, modeling and simulation, and performance assessments. The agency declared the delivery of nine capabilities during fiscal year 2008 as shown in figure 5 below. It declared three early capabilities for Block 1.0 engagement sequence groups, three early as well as one full capability for Block 2.0 engagement sequence groups, and one early capability and one partial capability for Block 3.1/3.2 engagement sequence groups. Figure 5: Timeline Showing Declaration of Capabilities in Fiscal Year 2008: [Refer to PDF for image: illustration] Block 1.0G: BI Launch-on COBRA DANE radar (Beale): FY08, Q1: Capability declaration, early; FY08, Q4: Milestone achieved, early. GBI Engage-on sea-based X-band radar: FY08, Q1: Capability declaration, early; FY08, Q4: Milestone achieved, early. GBI Launch-on sea-based X-band radar: FY08, Q1: Capability declaration, early; FY08, Q4: Milestone achieved, early. Block 2.0: SM-3 Engage-on shipboard Aegis radar: FY08, Q2: Full milestone scheduled; FY08, Q4: Full milestone achieved. SM-3 Launch on remote shipboard Aegis radar: FY08, Q2: Milestone scheduled, early; FY08, Q4: Milestone achieved, early. SM-2 Engage-on shipboard Aegis radar: FY08, Q4: Milestone scheduled, early; FY08, Q4: Milestone achieved, early. THAAD Engage-on AN/TPY-2 radar (terminal mode): FY08, Q3: Milestone scheduled, early; FY08, Q4: Milestone achieved, early. Block: 3.1/3.2: GBI Engage-on sea-based X-band radar: FY08, Q2: Milestone achieved, partially. GBI Launch-on sea-based X-band radar: FY08, Q3: Milestone scheduled, early; FY08, Q4: Milestone achieved, early. Source: GAO analysis of MDA. Note: Our analysis above is based on the MDA Master Execution Fielding Schedule dated October 2007 as well as the Master Fielding Plan dated February 2008. Commensurate with its new block structure, MDA reported a subset of these as part of its fiscal year 2008 Statement of Goals dated January 2008. However, MDA continues to work toward declaring all of the October 2007 engagement sequence groups. [End of figure] MDA had intended to use the results of a flight test (FTG-04), that was later canceled;[Footnote 37] a distributed ground test (GTD-03), that was delayed into fiscal year 2009; and the results of Performance Assessments 2007 and 2008 to determine if capabilities were ready for declaration in fiscal year 2008. Consequently, these shortfalls in knowledge led MDA to reduce the basis for declaring capability goals. Performance Assessment 2007--identified by MDA as a key source to assess capabilities during fiscal year 2008--achieved only limited accreditation.[Footnote 38] This less-than-full accreditation indicated that MDA could not rely on the assessment's results to gauge end-to-end BMDS performance. Subsequently, MDA officials decided to cancel Performance Assessment 2008 because they needed time to address problems and prepare for Performance Assessment 2009. While MDA officials declared these capabilities during fiscal year 2008, they did so after mostly reducing the basis for the declarations. They reverted back in several cases to older ground and flight tests, though MDA in a few cases added some newer flight and ground tests as well. For example, MDA declared early Block 1.0 capability for three engagement sequence groups in fiscal year 2008 without the planned results from Performance Assessment 2007. In all cases, though MDA had intended to use the final results from comprehensive performance assessments, after revising the basis for declaring capability goals it eliminated them entirely. Specifically, MDA dropped some sources of data it expected to use, such as the canceled Performance Assessment 2008, and shifted from flight and ground tests planned to occur in fiscal year 2008 to older flight and ground tests. For example, in Block 2.0 MDA declared full capability during fiscal year 2008 for one engagement sequence group, Aegis BMDS engage on its shipboard radar, even though Performance Assessment 2008 had been canceled. MDA instead based its decision on integrated and distributed ground tests (GTI-02 and GTD-02) conducted in calendar year 2007 as well as prior flight tests during fiscal years 2006 through 2008. However, the BMDS Operational Testing Agency raised concerns about the comprehensiveness of the GTI-02 scenarios,[Footnote 39] specifically, the incorrect configuration of U.S. satellites and threat data. MDA also deferred 13 capability goals scheduled to occur in fiscal year 2008 to the end of fiscal year 2009, as shown in figure 6 below. Figure 6: Timeline Showing Deferred Declaration of Capabilities from Fiscal Year 2008 to 2009: [Refer to PDF for image: illustration] Block 1.0: GBI Engage-on COBRA DANE radar (Beale): FY08, Q2: Full milestone scheduled; FY09, Q3: Full milestone achieved. GBI Launch-on COBRA DANE radar (Beale): FY08, Q1: Partial milestone scheduled; FY09, Q4: Partial milestone achieved. GBI Engage-on shipboard Aegis radar: FY09, Q2: Full milestone scheduled; FY09, Q4: Full milestone achieved. GBI Launch-on shipboard Aegis radar: FY08, Q2: Full milestone scheduled; FY09, Q4: Full milestone achieved. GBI Engage-on forward-based AN/TPY-2 radar: FY08, Q2: Full milestone scheduled; FY09, Q4: Full milestone achieved. GBI Launch-on forward-based AN/TPY-2 radar: FY08, Q2: Full milestone scheduled; FY09, Q4: Full milestone achieved. GBI Engage-on sea-based X-band radar: FY08, Q1: Partial milestone scheduled; FY09, Q4: Partial milestone achieved. GBI Launch-on sea-based X-band radar: FY08, Q1: Partial milestone scheduled; FY09, Q4: Partial milestone achieved; FY09, Q2: Full milestone scheduled; FY09, Q4: Full milestone achieved. Block 2.0: SM-3 Launch-on shipboard Aegis radar: FY08, Q1: Partial milestone scheduled; FY09, Q4: Partial milestone still delayed; FY08, Q4: Full milestone scheduled; FY09, Q4: Full milestone still delayed. Block 3.1/3.2: GBI Engage on COBRADANE radar Mod 1 (Fylingdales, UK; forward-based: AN/TPY-2 radar): FY08, Q2: Full milestone scheduled; FY09, Q4: Full milestone still delayed. GBI Launch on forward-based AN/TPY-2 radar, Mod 1a (Hercules 1): FY08, Q3: Early milestone scheduled; FY09, Q4: Early milestone still delayed. GBI Engage on forward-based AN/TPY-2 radar, Mod 1a (Hercules 1): FY08, Q3: Early milestone scheduled; FY09, Q4: Early milestone still delayed. Source: GAO analysis of MDA data. Note: Our analysis above is based on the MDA Master Execution Fielding Schedule dated October 2007 as well as the Master Fielding Plan dated February 2008. Commensurate with its new block structure, MDA reported a subset of these as part of its fiscal year 2008 Statement of Goals dated January 2008. However, MDA continues to work toward declaring all of the October 2007 engagement sequence groups. Several engagement sequence groups are not shown here because they are Block 3.3, which MDA has not yet baselined. [End of figure] MDA intended to declare all Block 1.0 engagement sequence groups as fully capable by the middle of fiscal year 2009.[Footnote 40] However, as MDA encountered test delays and technical challenges, it had to defer full capability declaration for these engagement sequence groups until the end of fiscal year 2009. For Block 2.0, MDA also deferred declaring full capability for one of the two planned full capability declarations for fiscal year 2008. This declaration is contingent upon the review of a ground test that has been rescheduled to the second quarter of fiscal year 2009 and a flight test rescheduled to the third quarter of fiscal year 2009. MDA also deferred one full and two early capability declarations for Block 3.1/3.2 beyond the end of fiscal year 2009. In response to the limitations of Performance Assessment 2007, the cancellation of Performance Assessment 2008 and FTG-04, and the delayed GTD-03 and FTG-05 flight tests, MDA is planning to rely on older ground and flight tests; a sensor flight test, FTX-03, instead of intercept flight tests; and the initial quick look review of Performance Assessment 2009 instead of the previously planned full analysis. Appendix IV provides a detailed layout for the reduced basis of capability declarations for fiscal years 2008 and 2009. Since MDA was only able to declare a few of the capabilities it planned for fiscal year 2008, the schedule for fiscal year 2009 and subsequent years will be compressed if the agency plans to maintain the schedule it has set for its blocks. For example, MDA may need to declare three times as many capabilities than originally planned for fiscal year 2009 in order to meet the 2009 capability declaration schedule. In addition, if the schedule cannot be maintained, MDA will likely have to make further adjustments to mitigate additional delays in BMDS capabilities. Increased reliance on integrated ground testing will provide less knowledge than a complete analysis of capabilities from a performance assessment. Integrated ground testing involves less robust conditions than distributed ground testing, which involves operational systems in the field. The MDA master fielding plan indicates that the agency originally intended to take a more comprehensive approach upon which to base capability declarations. Reliance on an upcoming Performance Assessment 2009 quick look for Block 1.0 completion is a particular concern because the knowledge it provides may be limited with respect to testing, according to BMDS Operational Test Agency officials. For example, officials told us that a quick look may indicate anomalies from a test but will not analyze their causes. In contrast, MDA originally planned to have a complete analysis from the Performance Assessment 2009 models, simulations, and tests. Limited Progress Made in Improving Transparency and Accountability: In March 2008, we reported that efforts were underway to improve BMDS management, transparency, accountability, and oversight including a new executive board outside of MDA and a new block structure along with other improvements within MDA.[Footnote 41] Since that time, the executive board that was established in 2007 has acted with increased oversight. MDA's efforts, however, have not made the expected progress. In particular, MDA has decided to retain the option of deferring work from one block to another; cost baselines have not been established; test baselines remain relatively unstable; and requesting procurement funds for some assets, as directed by Congress, will not occur until fiscal year 2010.[Footnote 42] To accomplish its mission, in 2002 the Secretary of Defense gave MDA requirements, acquisition, and budget flexibilities and relief from some oversight mechanisms and reporting responsibilities. The flexibility granted to MDA has allowed concurrent development, testing, manufacturing, and fielding. MDA used this flexibility to quickly develop and field the first increment of capability in 2005. In August 2008, in response to Congressional direction to assess the current and future missions, roles, and structure of MDA, an independent study group agreed that there is a need to move MDA toward more normal acquisition processes. However, the group noted that the continuous evolution of the BMDS requires that the approach to setting requirements for, developing, and fielding increments of capability should remain as special authorities with oversight of the Missile Defense Executive Board (MDEB). Further, in regards to budget flexibilities, the independent group concluded that while these flexibilities may have been deemed necessary at the time, it should not have been expected that all the special authorities granted to MDA would continue or would have a need to continue in full force beyond achieving the President's goal of deploying a set of initial capabilities. Missile Defense Executive Board's Oversight Role More Active in 2008: During 2008, the MDEB appeared to act with an increased level of authority in providing oversight of MDA and the BMDS. For example, the board took on a major role in making key decisions regarding the transition of elements to military services. We previously reported that MDA and the military services had been negotiating the transition of responsibilities for the sustainment of fielded BMDS elements, but this process had been proven to be arduous and time consuming. However, in 2008, with the influence of the MDEB, a lead military service designation was appointed for one BMDS asset--the Sea-based X-band radar.[Footnote 43] In March 2008, we reported that the MDEB could play a key role in the Joint Requirements Oversight Council's proposal to return the BMDS to the Joint Capabilities Integration and Development System requirements process--a formal DOD procedure followed by most DOD programs that defines acquisition requirements and evaluation criteria for future defense programs. In responding to the proposal, the Acting Under Secretary of Defense for Acquisition, Technology, and Logistics recommended that the Deputy Secretary of Defense delay his approval of the Joint Staff's proposal until the MDEB could review the proposal and provide a recommendation. According to Acquisition, Technology and Logistics officials, no decision has been made regarding returning the BMDS to the requirements process. However, the Deputy Secretary of Defense, in September 2008, appeared to strengthen the oversight role of the MDEB, clarifying the roles of the MDEB as well as MDA, the Office of the Secretary of Defense, Combatant Commands, and Military Departments. With respect to the role of the MDEB, he established a life cycle management process for the BMDS stating that the MDEB will recommend and oversee implementation of strategic policies and plans, program priorities, and investment options to protect our Nation and our allies from missile attack. One of the MDEB functions is to provide the Under Secretary of Defense for Acquisition, Technology, and Logistics--or Deputy Secretary of Defense, as necessary--a recommended strategic program plan and feasible funding strategy for approval. The Deputy Secretary further noted that, through the use of the BMDS Life Cycle Management Process outlined in the memo, the MDEB will oversee the annual preparation of the BMDS portfolio, including BMDS-required capabilities and a program plan to meet the requirements with Research, Development Test & Evaluation, procurement, operations and maintenance, and military construction resources in defense-wide accounts. To further increase BMDS transparency and oversight of the BMDS, the Under Secretary of Defense Acquisition, Technology, and Logistics plans to hold program reviews for several BMDS elements commensurate with the authority granted to the MDEB by the Deputy Secretary of Defense. According to Under Secretary of Defense for Acquisition, Technology, and Logistics officials, the MDEB conducted its first of such reviews in November 2008 of the THAAD program. This review covered production and the element's contract schedule. Under Secretary of Defense for Acquisition, Technology, and Logistics officials told us that these reviews are designed to provide the Deputy Director Acquisition, Technology, and Logistics with comprehensive information that will be used as the basis for MDEB recommendations for the BMDS business case and baseline processes--a process which, according to these officials, is similar to the traditional Defense Acquisition Board process for reviewing other major acquisition programs. However, it is unclear whether the information provided to the MDEB will be comparable to that produced for other major acquisition program reviews as most of the information appears to be derived or presented by MDA as opposed to independent sources as required for traditional major defense acquisition programs.[Footnote 44] Efforts to Improve Transparency of MDA's Work Have Not Progressed as Planned: Deferral of Work: In 2007, MDA redefined its block structure to better communicate its plans and goals to Congress. The agency's new structure is based on fielding capabilities that address particular threats instead of the biennial time periods previously used to develop and field the BMDS. Last year, we reported that MDA's new block plans included many positive changes.[Footnote 45] However, MDA, with its submission of its Fiscal Year 2008 Statement of Goals, reversed some of the positive aspects of the new block structure. For example, we previously reported that the new block structure would improve the transparency of each block's actual cost by disallowing the deferral of work from one block to another. Under its prior block structure, MDA deferred work from one block to another; but it did not track the cost of the deferred work so that it could be attributed to the block that it benefited. Because MDA did not track the cost of the deferred work, the agency was unable to adjust the cost of its blocks to accurately capture the cost of each. This weakened the link between budget funds and the work performed. Last year, MDA officials told us that under its new block approach, MDA would no longer transfer work under any circumstances to a different block. However, MDA officials recently said that they are retaining the option to move work from one block to another as long as it is accompanied by a rebaseline. This change allows the agency to continue the practice of moving work from one block to another, which thereby reduces the transparency of the new block structure and undermines any baselines that are established. Use of Procurement Funds: In March 2007, we reported that the majority of MDA's funding comes from the Research, Development, Test, and Evaluation appropriation account, another flexibility provided by law.[Footnote 46] In past years, Congress authorized MDA to pay for assets incrementally using research and development funds. This allowed MDA to fund the purchase of assets over multiple years. Congress recently restricted MDA's authority and required MDA to purchase certain assets with procurement funds and directed that for any year after fiscal year 2009, MDA's budget materials must delineate between funds needed for research, development, test, and evaluation, procurement, operations and maintenance, and military construction. Requiring MDA to request funding in these appropriation categories will mean that it will be required to follow the funding policies for each category. For example, using procurement funds will mean that MDA will be required to ensure that assets are fully funded in the year of their purchase, rather than incrementally funded over several years. Congress directed in the 2008 National Defense Authorization Act, for any year after fiscal year 2009, that MDA's budget materials delineate between funds needed for research, development, test, and evaluation, procurement, operations and maintenance, and military construction. We have previously reported that using procurement funds will mean that MDA generally will be required to adhere to congressional policy that assets be fully funded in the year of their purchase, rather than incrementally funded over several years. The Congressional Research Service reported in 2006 that "incremental funding fell out of favor because opponents believed it could make the total procurement costs of weapons and equipment more difficult for Congress to understand and track, create a potential for DOD to start procurement of an item without necessarily stating its total cost to Congress, permit one Congress to 'tie the hands' of future Congresses, and increase weapon procurement costs by exposing weapons under construction to uneconomic start-up and stop costs."[Footnote 47] In the 2008 National Defense Authorization Act, Congress also provided MDA with the authority to use procurement funds for fiscal years 2009 and 2010 to field its BMDS capabilities on an incremental funding basis, without any requirement for full funding. Congress has granted similar authority to other DOD programs. In the conference report accompanying the Act, Conferees cautioned DOD that additional authority will be considered on a limited case-by-case basis and noted that they expect that future missile defense programs will be funded in a manner more consistent with other DOD acquisition programs. MDA did not request any procurement funds for fiscal year 2009. During our audit, the agency had not yet released the 2010 budget request to include such funding categories. However, MDA officials told us that the agency plans to request procurement funds for some of its assets in its fiscal year 2010 request, but could not elaborate on its plans to do so. Given that data was unavailable, it is unclear for which assets procurement funds will be requested or the extent to which the request will meet the direction given by Congress. According to MDA officials, information regarding its plans to request procurement funding will not be released until spring 2009. Baselines: Baselines represent starting points against which actual progress can be measured. They are thus used to provide indications of when a program is diverting from a plan. Baselines can be established to gauge progress in different areas, including cost, schedule and testing. Overall, the BMDS does not have baselines that are useful for oversight. With regard to cost, we have already discussed the lack of total and unit cost baselines for missile defense as well as the frequency of changes in contract baselines. MDA made some progress with developing a schedule baseline for its blocks and their associated capabilities. The agency's annual Statement of Goals identifies its schedule baseline as the fiscal year dates for early, partial, and full capability deliveries of hardware and functionality for a block. Thus, while MDA has changed its schedule for making declarations, the effect of the change can be determined by comparison with the original schedule. MDA does not have test baselines for its blocks. The agency does however, have baselines for its test program, but revises them frequently. They are therefore not effective for oversight. The agency identified its Integrated Master Test Plan as the test baseline for the BMDS. However, as depicted in table 9, the agency has made a number of changes to the content of the baseline. Table 9: MDA BMDS Test Baseline Revisions: Version: 5.6.2; Revisions/change date: February 20; Rationale for change: Interim Update: Changed for signature; Version approved: [Check]. Version: 8.01; Revisions/change date: August 15; Rationale for change: Incorporated MDA new block construct; migrated from calendar year format to fiscal year format; Version approved: [Empty]. Version: 8.04; Revisions/change date: October 12; Rationale for change: Updated funding status. Incorporated Operational Test Agency input; Version approved: [Empty]. Version: 8.06; Revisions/change date: November 6; Rationale for change: Preparation for internal MDA coordination; Version approved: [Empty]. Version: 8.07; Revisions/change date: December 11; Rationale for change: Program Change Board changes incorporated. Includes some fiscal year 2008 and 2009 budget decisions; Version approved: [Empty]. Version: 8.1; Revisions/change date: February 5; Rationale for change: Updated with Program Change Board changes; Version approved: [Check]. Version: 8.4; Revisions/change date: July 19; Rationale for change: Quarterly update based on Program Change Board and working group decisions; Version approved: [Empty]. Version: 9.1; Revisions/change date: September 26; Rationale for change: Quarterly update limited to schedules; Version approved: [Empty]. Source: GAO analysis of MDA documents. [End of table] The official approved test baseline changes every year and there are numerous more informal changes that happen more frequently. Most of the annual revisions to the test baseline occur either because MDA has changed the substance of test, changed the timing of tests, or added tests to the baseline. The Integrated Master Test Plan establishes the executable test program for the current fiscal year and extends through the following fiscal year. According to MDA, the test plan is updated quarterly based upon decisions from MDA leadership or formal decision-making forums such as the Program Control Board, and is also coordinated annually with external agencies such as the Office of the Director of Operational Test and Evaluation. However, as shown in table 9, there are several versions for a given quarter and as many as seven versions have been developed since the fiscal year 2008 baseline was established. It is unclear which Integrated Master Test Plan version MDA manages to at any given time. For example, in November 2008, we requested the latest version of the Integrated Master Test Plan and were told that the latest approved version was 8.4, which was revised in July 2008. However, the signature page for that version is from a prior version-- version 8.1. Since there is no signature page referring to version 8.4, it appears that this version is unapproved though MDA officials told us that it was being used to manage the BMDS. Agency officials maintain that the document is used to manage tests and associated requirements for key test events. However, it is unclear how well the baseline is integrated with other key aspects of testing such as the acquisition of targets needed to execute tests. For example, in some instances, targets are placed on contract for two or more years in advance of the planned tests. Yet, the test baseline--the Integrated Master Test Plan--does not appear to include events beyond the following fiscal year that are key to the BMDS test program. As we reported in September 2008, MDA officials acknowledged that its target contracts did not capture all element testing requirements and target baselines were not always established before targets contracts were signed.[Footnote 48] Conclusions: In 2002, MDA was given unusual authorities to expedite the fielding of an initial BMDS capability. As this initial capability was fielded in 2005, it showed the benefits of these flexibilities. MDA has improved on this capability in the ensuing years, including 2008, the focus of this report. Today, the program is still operating at a fast pace, as production and fielding of assets outstrips the ability to test and validate them. A collateral effect of these flexibilities has been reduced visibility into actual versus planned progress. Some fundamental questions of an oversight nature are not yet answerable. For example, a comparison of actual versus planned costs at the system or asset level is not yet possible, nor is an assessment of the performance of the fielded system as a whole. Beginning in 2007, MDA began efforts to improve visibility into its actual performance, beginning with the new way of defining blocks, coupled with DOD's creation of the MDEB. However, progress has been slow in some areas and value for money cannot be satisfactorily assessed. Delays are especially important in a program of this size, as a year delay in establishing cost baselines means another $8 billion to $10 billion may be spent in the meantime. With the transition to a new administration, the deployment and subsequent improvement of an initial capability, a new agency Director, and a new block structure for managing the BMDS, an opportunity exists to revisit and strengthen the processes by which MDA operates. Looking to the future, decision makers in Congress and DOD face multi-billion dollar investment decisions in allocating funds both within MDA and between MDA and other DOD programs. At this point, a better balance must still be struck between the information Congress and DOD need to conduct oversight of the BMDS and the flexibility MDA needs to manage across the portfolio of elements that collectively constitute the system's capability. At this point, the balance does not provide sufficient information for effective oversight. In particular: * Total cost and unit cost baselines have not been set and contract baselines are subject to frequent changes. Even if such baselines are set as planned, they will only capture about 26 percent of MDA's work. * Less testing is conducted than planned, thus delaying the validation of the models and simulations needed to assess the overall performance of the BMDS. Moreover, test plans do not hold and are revised often, in many cases due to the poor performance of target missiles. The current test plan is at risk given its ambitious scope. * Manufacturing, production, and fielding are outpacing testing, modeling, and validation. Consequently, fielding decisions and capability declarations are being made with limited understanding of system effectiveness. Recommendations for Executive Action: We recommend that the Secretary of Defense direct the MDEB to assess how the transparency and accountability of MDA's acquisitions can be strengthened to enhance oversight, such as by adopting relevant aspects of DOD's normal requirements, acquisition and budgeting processes, without losing the beneficial features of MDA's existing flexibility. In the near term we recommend that the Secretary of Defense direct MDA to undertake the following 10 actions: In the area of cost: 1. Complete total cost baselines before requesting additional funding for Blocks 2.0 and 3.0 and commit to a date when baselines for all blocks will be established. 2. Ensure that transfers of work from one block to another are transparent and reported as cost variances. 3. Provide additional unit costs reports, beyond flyaway unit costs, that incorporate both procurement and research and development funding so that there is a more comprehensive understanding of the progress of the acquisitions. In the area of testing and performance: 4. Expand the BMDS test baseline to include tests scheduled beyond the first succeeding year of the plan to ensure its synchronization with BMDS contracts. 5. Ensure that DOT&E is consulted before making significant changes to the test baseline so that the tests planned provide DOT&E with sufficient data to assess the performance of the BMDS elements. 6. Ensure that planned test objectives include concrete data requirements anchoring models and simulations to real-world tests, synchronized with flight and ground test plans and that the effects on models and simulations of test cancellations, delays or problems are clearly identified and reported. 7. Reassess the flight tests scheduled for the end of fiscal year 2009 to ensure that they can be reasonably conducted and analyzed given targets and other constraints. In the area of knowledge-based decisions: 8. Synchronize the development, manufacturing, and fielding schedules of BMDS assets with the testing and validation schedules to ensure that items are not manufactured for fielding before their performance has been validated through testing. 9. Conduct a flight test of the CE-I EKV against a complex target scene with countermeasures to complete MDA's previous testing goal of understanding the performance capabilities of the first 24 fielded GBIs. 10. Strengthen the capability declarations by using the complete analysis from annual performance assessments as the basis for declaring engagement sequence groups as fully capable and block development as fully complete; otherwise, indicate the limitations of the capabilities and steps that MDA will take to reduce the risks. Agency Comments and Our Evaluation: DOD provided written comments on a draft of this report. These comments are reprinted in appendix I. DOD also provided technical comments, which were incorporated as appropriate. DOD fully concurred with 10 of our 11 recommendations and partially concurred with our recommendation that the Secretary of Defense direct MDA to synchronize the development, manufacturing, and fielding schedules of BMDS assets with testing and validation schedules to ensure that items are not manufactured for fielding before their performance has been validated through testing. Yet, even DOD's response to this recommendation appears to be, in substance, concurrence. DOD concurred with our recommendation that the Secretary of Defense direct MDA to ensure that transfers of work from one block to another are transparent and reported as cost variances. DOD noted in its response that MDA will report block baselines and variances annually to Congress in the BMDS Accountability Report. The Department further noted that for the purposes of unit cost reporting, MDA has defined a cost variance as a confirmed increase of 10 percent or more in block or unit costs when compared to the current cost baseline or 20 percent or more compared to the original cost baseline, stating that transfers of work creating such cost variances will be reported. The intent of our recommendation is to increase visibility into transfers of work between blocks regardless of the amount of the increase or the baseline status of the blocks. The trigger for reporting the variances selected by DOD will not necessarily provide that visibility. Given that only between 2 and 26 percent of BMDS block and capability development costs from fiscal year 2010 to 2013 will be baselined initially, visibility into transfers into blocks that are not yet baselined may not occur. Further, an increase may not be reported in the baselined block from which work is transferred because the transfer would actually yield a decrease in the cost of the baselined block. An increase would also not be reported in the receiving block if that block is not baselined or if the transfer did not increase costs above the threshold. MDA may need to consider a different approach to reporting that captures meaningful transfers of work into and out of blocks regardless of whether any of the blocks are baselined. MDA should work with Congress to determine what constitutes a meaningful or significant cost increase. DOD also concurred with our recommendation that the Secretary of Defense strengthen the capability declarations by using the complete analysis from annual performance assessments. In responding to our recommendation, DOD noted that if there is limited performance assessment data, the overall capability assessment will factor in the knowledge gained from ground tests and flight tests against the identified risks. While we recognize that MDA is not always able to complete all of its planned tests in a given time period, when MDA decides to change the planned basis for its capability declarations to a different or reduced set of data it is important for the agency to clearly report the limitations that affect the capability declaration as well as the mitigation steps it is taking. We are sending copies of this report to the Secretary of Defense and to the Director, MDA. In addition, the report will be available at no charge on the GAO Web site at [hyperlink, http://www.gao.gov]. If you, or your staff have any questions concerning this report, please contact me at (202) 512-4841. Contact Points for our offices of Congressional Relations and Public Affairs may be found on the last page of this report. The major contributors are listed in appendix VI. Signed by: Paul Francis: Director, Acquisition and Sourcing Management: List of Congressional Committees: The Honorable Carl Levin: Chairman: The Honorable John McCain: Ranking Member: Committee on Armed Services: United States Senate: The Honorable Daniel K. Inouye: Chairman: The Honorable Thad Cochran: Ranking Member: Subcommittee on Defense: Committee on Appropriations: United States Senate: The Honorable Ike Skelton: Chairman: The Honorable John M. McHugh: Ranking Member: Committee on Armed Services: House of Representatives: The Honorable John P. Murtha: Chairman: The Honorable C.W. Bill Young: Ranking Member: Subcommittee on Defense: Committee on Appropriations: House of Representatives: [End of section] Appendix I: Comments from the Department of Defense: Office Of The Under Secretary Of Defense: Acquisition Technology And Logistics: 3000 Defense Pentagon: Washington, DC 20301-3000: March 3, 2009: Mr. Paul Francis: Director, Acquisition and Sourcing Management: U.S. Government Accountability Office: 441 G Street, N.W. Washington, DC 20548: Dear Mr. Francis: This is the Department of Defense (DoD) response to the GAO Draft Report, GAO-09-338, "Defense Acquisitions: Production and Fielding of Missile Defense Components Continue With Less Testing and Validation Than Planned," dated February 11, 2009 (GAO Code 120744). Detailed comments on the report recommendations are enclosed. The DoD concurs with ten of the draft report's recommendations and partially concurs with one. The rationale for our position is included in the enclosure. I submitted separately a list of technical and factual errors for your consideration. We appreciate the opportunity to comment on the draft report. My point of contact for this effort is Mr. David Crim, (703) 697-5385, david.crim@osd.mil. Sincerely, Signed by: David G. Ahern: Director: Portfolio Systems Acquisition: Enclosure: As stated: GAO Draft Report Dated February 11, 2009: GAO-09-338 (GAO Code 120744): "Defense Acquisitions: Production And Fielding Of Missile Defense Components Continue With Less Testing And Validation Than Planned" Department Of Defense Comments To The GAO Recommendations: Recommendation 1: The GAO recommends that the Secretary of Defense direct the Missile Defense Executive Board to assess how the transparency and accountability of Missile Defense Agency (MDA's) acquisitions can be strengthen to enhance oversight, such as by adopting relevant aspects of DoD's normal requirements, acquisition and budgeting processes, without losing the beneficial features of MDA's existing flexibility. DoD Response: Concur. As GAO noted, DoD has recently enhanced the transparency, accountability, and oversight of the missile defense program. For example, the Missile Defense Executive board (MDEB) has played an increasingly important role in Ballistic Missile Defense System (BMDS) policy and programmatic decisions. Existing flexibilities and the associated integrated decision authority for requirements, acquisition, and budget have facilitated MDA's efforts to provide critical capabilities to the war fighter. However, it's an appropriate time to take a fresh look. Such a review might also identify where the flexibilities and integrated decision authority granted MDA could be applied beneficially to other DoD programs. Recommendation 2: The GAO recommends that the Secretary of Defense direct MDA to complete total cost baselines before requesting additional funding for blocks two and three and commit to a date when baselines for all blocks will be established. DoD Response: Concur. MDA intends to present its cost baselines for Blocks 2.0 and 3.1/3.2 in the BMDS Accountability Report (BAR) accompanying the President's Budget for Fiscal Year (FY) 2010. As for Blocks 3.3 and 5.0, assuming events unfold as expected, we intend to baseline their costs no later than the issuance of next year's BAR. MDA intends to baseline Block 4.0 within one fiscal year of reaching agreements with the Czech and Polish governments and obtaining needed Congressional approvals. Recommendation 3: The GAO recommends that the Secretary of Defense direct MDA to ensure that transfers of work from one block to another are transparent and reported as cost variances. DoD Response: Concur. MDA will report block baselines and variances annually to Congress in the BMDS Accountability Report. For the purposes of unit cost reporting, MDA has defined a cost variance as a confirmed increase of 10 percent or more in block or unit costs when Attachment Page 1 of 3 compared to the current cost baseline or 20 percent or more compared to the original cost baseline. Transfers of work creating such cost variances will be reported. Recommendation 4: The GAO recommends that the Secretary of Defense direct MDA to provide additional unit costs reports, beyond flyaway unit costs, that incorporate both procurement and research and development funding so that there is a more comprehensive understanding of the progress of the acquisitions. DoD Response: Concur. MDA will provide additional unit cost reports that include development and integration costs, flyway costs, initial spares, and support items (incorporating research and development or procurement funds as appropriated) for major pieces of equipment such as interceptors, sensors, fire control/command systems, and launch systems in baselined blocks. Recommendation 5: The GAO recommends that the Secretary of Defense direct MDA to expand the Ballistic Missile Defense Systems (BMDS) test baseline to include tests scheduled beyond the first succeeding year of the plan to ensure its synchronization with BMDS contract. DoD Response: Concur. The Missile Defense Agency is developing the Integrated Master Test Plan (IMTP) that spans the Future Year Defense Program (FYDP) rather than in two year increments. It will include test objectives based on specifications, modeling and simulation verification, validation and accreditation and Critical Operational Issues. Recommendation 6: The GAO recommends that the Secretary of Defense direct MDA to ensure Director, Operational Test and Evaluation (DOT&E) is consulted before making significant changes to the test baseline so that the tests planned provide the Director of DOT&E with sufficient data to assess the performance of the BMDS elements. DoD Response: Concur. Proposed updates to the test baseline and the IMTP will be coordinated through established Working Groups, which include DOT&E, Operational Test Agencies (OTAs), and the War fighter as permanent members. The updates will continue to be staffed through MDA Leadership and both DOT&E and the OTAs, and DOT&E will continue as a signator to the IMTP. Recommendation 7: The GAO recommends that the Secretary of Defense direct MDA to ensure that planned test objectives include concrete data requirements anchoring models and simulations to real-world tests, synchronized with flight and ground test plans and that the effects on models and simulations of test cancellations, delays or problems are clearly identified and reported. DoD Response: Concur. Flight tests include objectives to anchor models and simulations to real-world test data. MDA currently does a post- flight-test reconstruction of each flight test, which includes the requirement to "re-fly" flight tests in our ground Hardware in Loop Tests. This helps to ensure that models and simulations accurately represent element and system performance. Additionally, MDA has clearly identified and reported the effects on models and simulations of test cancellations, delays or problems and will continue to do so. Recommendation 8: The GAO recommends that the Secretary of Defense direct MDA to reassess the flight tests scheduled for the end of fiscal year 2009 to ensure that they can he reasonably conducted and analyzed given targets and other constraints. DoD Response: Concur. MDA concurs with reassessment of the test schedule as it applies to the conduct of flight tests through the end of fiscal year 2009. Recommendation 9: The GAO recommends that the Secretary of Defense direct MDA to synchronize the development, manufacturing, and fielding schedules of BMDS assets with testing and validation schedules to ensure that items are not manufactured for fielding before their performance has been validated through testing. DoD Response: Partially Concur. MDA is pursuing synchronization of development, manufacturing and fielding of BMDS assets with the IMTP's testing and validation requirements. That synchronization will be captured in the draft BMDS Master Plan (BMP) and its associated Integrated Master Plan (IMP) and Integrated Master Schedule (IMS). While successful ground and flight tests have provided confidence in BMDS capabilities being fielded, MDA and the war fighter recognize that additional validation through modeling and simulation is needed. Recommendation 10: The GAO recommends that the Secretary of Defense direct MDA to conduct a flight test of the Capability Enhancement (CE)- I Exoatmospheric Kill Vehicle (EKV) against a complex target scene with countermeasures to complete MDA's previous testing goal of understanding the performance capabilities of the first 24 fielded Ground-Based Interceptors (GBIs). DoD Response: Concur. MDA is currently reexamining its flight testing program and expects to include additional flight testing of the Capability Enhancement CE-I EKV. This testing will be reflected in the next IMTP update. Such testing will include the specific objective to discriminate and intercept the dynamic lethal object from a target scene with countermeasures. Recommendation 11: The GAO recommends that the Secretary of Defense direct MDA to strengthen the capability declarations by using the complete analysis from annual performance assessments as the basis for declaring engagement sequence groups as fully capable and block development as fully complete; otherwise, indicate the limitations of the capabilities and steps the MDA will take to reduce the risks. DoD Response: Concur. MDA makes a capability declaration based on complete analysis of data from all available ground test, flight test and performance assessment events. As part of the technical assessment criteria for a capability delivery (whether an engagement sequence group or BMDS Block capability). MDA identifies the capabilities and limitations and provides the MDA Director with the summary of any remaining risk to the capability. If there is limited performance assessment data, the overall capability assessment will factor in the knowledge gained from ground tests and flight tests against the identified risks. [End of section] Appendix II: BMDS Prime Contractors Exceed Budgeted Cost and Schedule Performance during Fiscal Year 2008: Based on our analysis of 14 Ballistic Missile Defense System (BMDS) elements' prime contractor earned value management performance, we determined that collectively the contractors overran budgeted cost by $152.4 million and were behind schedule by approximately $107.4 million during the fiscal year.[Footnote 49] Our insight of the dollar gained or lost for each dollar invested is based on monthly earned value reports which are required of each BMDS program office's prime contractor. These reports compare monthly progress to the cost or schedule performance baseline to reveal whether the work scheduled is being completed on time and if the work is being completed at the cost budgeted. For example, if the contractor was able to complete more work than scheduled and for less cost than budgeted, the contractor reports a positive schedule and cost variance. Alternatively, if the contractor was not able to complete the work in the scheduled time period and spent more than budgeted, the contractor reports both a negative schedule and cost variance. The results can also be mixed by, for example, completing the work under cost (a positive cost variance) but taking longer than scheduled to do so (a negative schedule variance). We also used contract performance report data to base predictions of likely overrun or underrun of each prime contractor's budgeted cost at completion. Our predictions of final contract cost are based on the assumption that the contractor will continue to perform in the future as it has in the past. In addition, since they provide the basis for our projected overruns, we also provide the total budgeted contract cost at completion for each contract we assessed in this appendix.[Footnote 50] However, the budgeted costs at completion, in some cases, have grown significantly over time. For example, in one case the budgeted cost at completion increased by approximately five times its original value. Since our assessment does not reveal, as cost growth, the difference between the original and current budgeted costs at completion it would be inappropriate to compare the underruns or overruns for MDA programs with cost growth on major defense acquisition programs since those major defense acquisition programs have established their full scope of work as well as developed total cost baselines, while these have not been developed for MDA programs. [Footnote 51] Aegis BMD Contractors Experienced Mixed Performance during the Fiscal Year: The Aegis Ballistic Missile Defense (Aegis BMD) program manages two prime contractors for work on its two main components--the Aegis BMD Weapon System and the Standard Missile-3 (SM-3). We report on work under one of the two separate Aegis BMD SM-3 contract's contract line item numbers (CLIN)on which we received sufficient performance data during fiscal year 2008. The first Aegis BMD SM-3 contract's CLIN 9 was for the production of 20 Block 1A missiles which began in February 2007 and finished deliveries in August 2008. Deliveries were completed $7.5 million under budget on the contractor's total budgeted cost of $179.0 million. The other Aegis BMD SM-3 contract's CLIN 1 is for a fourth lot of 27 Block 1A missiles and began reporting performance data in August 2007 for work that is still ongoing. The weapon system contractor experienced cost growth and schedule delays while the SM-3 contractor for the ongoing CLIN 1 for 27 Block 1A missiles had mixed performance. Neither of these CLINs experienced a realignment during fiscal year 2008. The Aegis BMD weapon system contractor experienced cumulative cost growth and schedule delays throughout the year. The Aegis BMD weapon system contractor overran budgeted cost and schedule during the fiscal year by $7 million and $5.1 million respectively. Although cumulative cost performance remains positive at $16 thousand, cumulative schedule performance continued to decline to negative $8.4 million. The negative cumulative schedule variance is driven by late engineering data, delays to qualification efforts, and the need to return components experiencing issues back to the vendor which required more time than originally planned. See figure 7 for cumulative cost and schedule performance during the fiscal year. Figure 7: Aegis BMD Weapon System Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $6.979 million; Cumulative Schedule Variance: -$3.258 million. Month: October 07; Cumulative Cost Variance: $8.085 million; Cumulative Schedule Variance: -$3.704 million. Month: November 07; Cumulative Cost Variance: $1.761 million; Cumulative Schedule Variance: -$3.61 million. Month: December 07; Cumulative Cost Variance: -$1.343 million; Cumulative Schedule Variance: -$2.384 million. Month: January 08; Cumulative Cost Variance: -$0.761 million; Cumulative Schedule Variance: -$5.098 million; Month: February 08; Cumulative Cost Variance: -$7.333 million; Cumulative Schedule Variance: -$6.143 million. Month: March 08; Cumulative Cost Variance: -$2.407 million; Cumulative Schedule Variance: -$5.35 million. Month: April 08; Cumulative Cost Variance: -$0.496 million; Cumulative Schedule Variance: -$8.977 million. Month: May 08; Cumulative Cost Variance: $0.12 million; Cumulative Schedule Variance: -$8.575 million. Month: June 08; Cumulative Cost Variance: -$1.922 million; Cumulative Schedule Variance: -$8.842 million. Month: July 08; Cumulative Cost Variance: -$2.384 million; Cumulative Schedule Variance: -$10.276 million. Month: August 08; Cumulative Cost Variance: -$2.442 million; Cumulative Schedule Variance: -$8.914 million. Months: September 08; Cumulative Cost Variance: $0.016 million; Cumulative Schedule Variance: -$8.397 million. Source: Contractor (data), GAO (presentation). [End of figure] The program attributes the fiscal year cost and schedule overruns mainly to the additional time and testing needed to ensure that the weapon system fielded capability was what was originally promised to the warfighter. To account for some of the overruns, the program performed fewer risk reduction efforts for a future weapon system capability release. If the contractor continues to perform as it has during the fiscal year, we project that at contract completion in September 2010, the contractor will overrun its budgeted cost of $1.2 billion by between $1.9 million and $12.2 million. The Aegis BMD SM-3 contractor, producing another lot of 27 Block 1A missiles under its CLIN 1, ended the fiscal year by underrunning budgeted costs by $3.0 million. The Aegis BMD SM-3 contractor for CLIN 1 work also ended the year with a negative $7.6 million schedule variance, which means that the contractor was unable to accomplish $7.6 million worth of planned work. Since reporting began in August 2007, cumulative and fiscal year variances are nearly equal with cumulative cost variances at a positive $3.3 million and cumulative schedule variances at negative $7.0 million. See figure 8 for a graphic representation of the cumulative cost and schedule variances during fiscal year 2008. Figure 8: Aegis BMD SM-3 CLIN 1 Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $0.2657 million; Cumulative Schedule Variance: $0.5567 million. Month: October 07; Cumulative Cost Variance: $0.6644 million; Cumulative Schedule Variance: -$0.2938 million. Month: November 07; Cumulative Cost Variance: -$0.1986 million; Cumulative Schedule Variance: $1.7739 million. Month: December 07; Cumulative Cost Variance: -$0.8043 million; Cumulative Schedule Variance: -$0.2961 million. Month: January 08; Cumulative Cost Variance: -$0.3651 million; Cumulative Schedule Variance: $1.4555 million; Month: February 08; Cumulative Cost Variance: $0.017 million; Cumulative Schedule Variance: $1.9683 million. Month: March 08; Cumulative Cost Variance: -$1.1359 million; Cumulative Schedule Variance: $2.3834 million. Month: April 08; Cumulative Cost Variance: -$0.0039 million; Cumulative Schedule Variance: $4.7049 million. Month: May 08; Cumulative Cost Variance: $0.3091 million; Cumulative Schedule Variance: $3.1361 million. Month: June 08; Cumulative Cost Variance: $1.236 million; Cumulative Schedule Variance: $1.4158 million. Month: July 08; Cumulative Cost Variance: $1.8045 million; Cumulative Schedule Variance: -$4.6013 million. Month: August 08; Cumulative Cost Variance: $1.5336 million; Cumulative Schedule Variance: -$5.4807 million. Months: September 08; Cumulative Cost Variance: $3.313 million; Cumulative Schedule Variance: -$7.0103 million. Source: Contractor (data), GAO (presentation). [End of figure] The contractor was able to accomplish fiscal year 2008 work for $3.0 million less than originally planned in part due to adjustments made in program management, labor efficiencies, reductions in vendor cost, and material transfers in the missile's fourth stage component. The unaccomplished work in negative $7.6 million worth of fiscal year schedule variances is largely in the first, second, and fourth stages portion of work. In the first stage booster, the contractor attributes some of the negative schedule variance to more than a year delay in testing the first stage due to rework needed to correct errors in the original drawing packages. In addition, the contractor cites second stage component delivery delays as drivers for the negative schedule variance. Vendors were unable to deliver these components due to holdups in approving waivers, achieving recertification after test equipment failures, and property damage to facilities. Lastly, the contractor experienced delays in components for the fourth stage which also contributed to the unfavorable schedule variance. If the contractor continues to perform as it did through September 2008, our analysis predicts that, at completion in April 2010, the work under the contract could cost from $6.6 million less to $0.7 million more than the budgeted cost of $237.5 million. ABL Contractor Overran Budgeted Fiscal Year Cost: For fiscal year 2008, the Airborne Laser (ABL) contractor overran fiscal year budgeted costs by $10.6 million but had a positive fiscal year schedule variance of $2.2 million. Despite some gains in its schedule variance during the fiscal year, the program still maintains negative cumulative cost and schedule variances of $84.8 million and $23.6 million respectively. The contractor mostly attributes the negative cumulative variances in cost and schedule to late beam control/fire control hardware deliveries. Despite a replan in June 2007, the ABL contractor did not perform any type of realignment during fiscal year 2008. Figure 9 shows cumulative variances at the beginning of fiscal year 2008 along with a depiction of the contractor's cost and schedule performance throughout the fiscal year. Figure 9: ABL Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$74.217 million; Cumulative Schedule Variance: -$25.767 million. Month: October 07; Cumulative Cost Variance: -$74.533 million; Cumulative Schedule Variance: -$22.015 million. Month: November 07; Cumulative Cost Variance: -$74.602 million; Cumulative Schedule Variance: -$20.48 million. Month: December 07; Cumulative Cost Variance: -$76.455 million; Cumulative Schedule Variance: -$20.444 million. Month: January 08; Cumulative Cost Variance: -$75.803 million; Cumulative Schedule Variance: -$19.487 million; Month: February 08; Cumulative Cost Variance: -$83.76 million; Cumulative Schedule Variance: -$24.927 million. Month: March 08; Cumulative Cost Variance: -$87.968 million; Cumulative Schedule Variance: -$26.377 million. Month: April 08; Cumulative Cost Variance: -$89.066 million; Cumulative Schedule Variance: -$25.155 million. Month: May 08; Cumulative Cost Variance: -$89.248 million; Cumulative Schedule Variance: -$24.72 million. Month: June 08; Cumulative Cost Variance: -$90.097 million; Cumulative Schedule Variance: -$26.772 million. Month: July 08; Cumulative Cost Variance: -$88.78 million; Cumulative Schedule Variance: -$26.546 million. Month: August 08; Cumulative Cost Variance: -$91.003 million; Cumulative Schedule Variance: -$29.548 million. Months: September 08; Cumulative Cost Variance: -$84.775 million; Cumulative Schedule Variance: -$23.606 million. Source: Contractor (data), GAO (presentation). [End of figure] Technical issues with key components of the Beam Control Fire Control system that required new hardware or refurbishment of existing components as well as late deliveries of key laser system components are the primary drivers of the unfavorable fiscal year cost variance of $10.6 million. These issues have caused delays in integration and test activities for the overall ABL weapon system. Based on the contractor's performance up through fiscal year 2008, we estimate that, at completion in February 2010, the contractor will overrun its budgeted cost of $3.6 billion by between $89.7 million and $95.4 million. C2BMC Program Incurred Negative Cumulative and Fiscal Year Variances: Our analysis of the Command, Control, Battle Management, and Communications' (C2BMC) cumulative contract performance indicates that the prime contractor's performance declined during fiscal year 2008. The contractor overran its fiscal year 2008 budget by $9.8 million and did not perform $3.6 million of work on schedule. By September 2008, this resulted in an unfavorable cumulative cost variance of $24.3 million and an unfavorable cumulative schedule variance of $7.1 million. The main drivers for the negative cumulative cost variances were costs associated with unplanned work, increased technical complexity, and reduction to cost efficiency due to losing key staff. The contractor attributes the unfavorable cumulative schedule variances to software issues related to the global engagement manager and components of test training operations. Although the C2BMC contractor performed a replan in November 2006, the contractor did not perform any type of realignment during fiscal year 2008. Trends in cost and schedule performance during the fiscal year are depicted in figure 10. Figure 10: C2BMC Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$14.476 million; Cumulative Schedule Variance: -$3.535 million. Month: October 07; Cumulative Cost Variance: -$18.458 million; Cumulative Schedule Variance: -$4.59 million. Month: November 07; Cumulative Cost Variance: -$21.255 million; Cumulative Schedule Variance: -$5.988 million. Month: December 07; Cumulative Cost Variance: -$17.271 million; Cumulative Schedule Variance: -$0.776 million. Month: January 08; Cumulative Cost Variance: -$19.3 million; Cumulative Schedule Variance: -$1.375 million; Month: February 08; Cumulative Cost Variance: -$23.593 million; Cumulative Schedule Variance: -$5.695 million. Month: March 08; Cumulative Cost Variance: -$22.957 million; Cumulative Schedule Variance: -$3.534 million. Month: April 08; Cumulative Cost Variance: -$26.384 million; Cumulative Schedule Variance: -$7.521 million. Month: May 08; Cumulative Cost Variance: -$20.172 million; Cumulative Schedule Variance: $0.309 million. Month: June 08; Cumulative Cost Variance: -$20.408 million; Cumulative Schedule Variance: $1.452 million. Month: July 08; Cumulative Cost Variance: -$21.062 million; Cumulative Schedule Variance: -$2.696 million. Month: August 08; Cumulative Cost Variance: -$23.253 million; Cumulative Schedule Variance: -$5.647 million. Months: September 08; Cumulative Cost Variance: -$24.273 million; Cumulative Schedule Variance: -$7.121 million. Source: Contractor (data), GAO (presentation). [End of figure] The negative fiscal year cost variance of $9.8 million is driven mainly by problems in the performance of work under Part 4 and Part 5 of the contract. The Part 4 effort, which began in December 2005, includes completing several spiral capabilities, upgrading spiral suites, and implementing initial global engagement capabilities at an operations center. The Part 5 effort began in December 2007 and covers operations and sustainment support for fielded C2BMC; the delivery of spiral hardware, software, and communications; as well as development, planning, and testing for other spiral capabilities. The contractor was able to use reserves to cover some of its Part 4 unfavorable fiscal year cost variances. The Part 5 fiscal year cost variance's primary drivers are unexpected complexities with the network design, unplanned work that required more resources for developing the planner, and the extension of efforts past the completion date on the global engagement management portion of work. The unfavorable fiscal year schedule variance of $3.6 million is attributable to the Part 5 portion of work and primarily caused by an unexpected reallocation of resources off of the global engagement management portion of work to other areas, delays in requesting material procurement also for global engagement management, and a lagging schedule for building out a testing lab. If the contractor continues to perform as it has in the past, we predict that the contractor will overrun its budgeted cost of $1.0 billion at completion in December 2009 by between $37.1 million and $76.8 million. GMD Contractor Maintained Negative Cumulative Cost and Schedule Variances throughout the Fiscal Year: The government and the Ground-based Midcourse Defense (GMD) contractor began a contract restructuring during the fiscal year to rephase and rescope on-going efforts to refine capability requirements and to adjust program content as well as perform weapon system integration, perform flight test planning, and work to develop the two-stage booster among other tasks. The ongoing realignment includes a proposal to add between $350 million and $580 million to the cost of the work under contract and to add 39 months to the period of performance. The GMD contractor reports a cumulative negative cost variance of more than $1.0 billion that it attributes to technical challenges with its Exoatmospheric Kill Vehicle (EKV) as well as supplier component quality problems. The contractor also carries a total unfavorable cumulative schedule variance of $130.3 million, the bulk of which the contractor attributes to the technical issues connected with the Ground-based Interceptor (GBI), particularly the EKV. For example, during the fiscal year the program experienced difficulties in manufacturing the Capability Enhancement II (CE-II) EKVs. Although the CE-II EKVs are expected to provide better performance, the contractor produced the kill vehicles before completing developmental tests, discovered problems during manufacturing, incorporated a new design, and continued manufacturing them. Although these issues contributed unfavorable fiscal year cost variances of $42.7 million, the program was able to make up for these losses in other areas. The variances, depicted in figure 11 represent the GMD contractor's cumulative cost and schedule performance over fiscal year 2008. Figure 11: GMD Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$1081.76 million; Cumulative Schedule Variance: -$52.887 million. Month: October 07; Cumulative Cost Variance: -$1077.93 million; Cumulative Schedule Variance: -$118.649 million. Month: November 07; Cumulative Cost Variance: -$1066.65 million; Cumulative Schedule Variance: -$115.545 million. Month: December 07; Cumulative Cost Variance: -$1049.57 million; Cumulative Schedule Variance: -$120.363 million. Month: January 08; Cumulative Cost Variance: -$1058.01 million; Cumulative Schedule Variance: -$111.88 million; Month: February 08; Cumulative Cost Variance: -$1060.91 million; Cumulative Schedule Variance: -$117.452 million. Month: March 08; Cumulative Cost Variance: -$1059.82 million; Cumulative Schedule Variance: -$125.855 million. Month: April 08; Cumulative Cost Variance: -$1045.19 million; Cumulative Schedule Variance: -$143.324 million. Month: May 08; Cumulative Cost Variance: -$1046.85 million; Cumulative Schedule Variance: -$123.692 million. Month: June 08; Cumulative Cost Variance: -$1037.42 million; Cumulative Schedule Variance: -$115.797 million. Month: July 08; Cumulative Cost Variance: -$1038.12 million; Cumulative Schedule Variance: -$113.994 million. Month: August 08; Cumulative Cost Variance: -$1030.16 million; Cumulative Schedule Variance: -$119.216 million. Months: September 08; Cumulative Cost Variance: -$1027.87 million; Cumulative Schedule Variance: -$130.261 million. Source: Contractor (data), GAO (presentation). [End of figure] The GMD contractor did have a favorable fiscal year cost variance of $53.9 million, which it attributed to labor efficiencies in the ground system as well as less field maintenance support required than planned, and labor efficiencies in the deployment and sustainment portion of the work under the contract. However, the GMD element's underruns occurred partially because the contractors delayed or eliminated some planned work. For example, the GMD program did not accomplish the emplacement of three GBIs, or conduct either of its two planned flight tests. As a result, it employed less labor than originally intended. The program also reports an unfavorable fiscal year schedule variance of $77.4 million which it attributes to an administrative error that occurred in September 2007. This error incorrectly adjusted the baseline to the booster effort in September which was then updated in October. However, it should also be noted that Missile Defense Agency (MDA) officials believe that ongoing adjustments to the GMD element's baseline have skewed recent variances to such a degree that they should not be used to predict future costs. We did perform analysis based on the contractor's reported performance through fiscal year 2008, and our analysis estimates that at contract end planned for December 2011, the contractor could overrun its budgeted cost of $14.9 billion by between $950.2 million and $1.25 billion. KEI Cost and Schedule Performance Continued to Decline after Replan: Despite a replan in April 2007 and again in April 2008, the Kinetic Energy Interceptors (KEI) contractor continued to experience declining cost and schedule performance during the fiscal year. Although the contractor began the year with a positive cost variance, the contractor overran fiscal year 2008 budgeted costs by $8.3 million, ending the year with an unfavorable cumulative cost variance of $2.6 million. In addition, the program was unable to accomplish $8.5 million worth of work which added to an unfavorable cumulative schedule variance of $21.3 million. Cumulative cost and schedule variances were mainly driven by costs associated with delays to booster drawing releases, delays in procurement, and unexpected costs and rework related to issues with the second stage. Figure 12 depicts the cost and schedule performance for the KEI contractor during fiscal year 2008. Figure 12: KEI Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $5.672 million; Cumulative Schedule Variance: -$12.784 million. Month: October 07; Cumulative Cost Variance: $7.276 million; Cumulative Schedule Variance: -$11.519 million. Month: November 07; Cumulative Cost Variance: $2.096 million; Cumulative Schedule Variance: -$12.699 million. Month: December 07; Cumulative Cost Variance: $7.025 million; Cumulative Schedule Variance: -$14.872 million. Month: January 08; Cumulative Cost Variance: $6.535 million; Cumulative Schedule Variance: -$17.217 million; Month: February 08; Cumulative Cost Variance: $5.144 million; Cumulative Schedule Variance: -$19.172 million. Month: March 08; Cumulative Cost Variance: $3.333 million; Cumulative Schedule Variance: -$18.34 million. Month: April 08; Cumulative Cost Variance: $2.689 million; Cumulative Schedule Variance: -$4.326 million. Month: May 08; Cumulative Cost Variance: -$1.871 million; Cumulative Schedule Variance: -$5.097 million. Month: June 08; Cumulative Cost Variance: $2.418 million; Cumulative Schedule Variance: -$7.655 million. Month: July 08; Cumulative Cost Variance: $1.23 million; Cumulative Schedule Variance: -$12.58 million. Month: August 08; Cumulative Cost Variance: $0.434 million; Cumulative Schedule Variance: -$17.127 million. Months: September 08; Cumulative Cost Variance: -$2.584 million; Cumulative Schedule Variance: -$21.277 million. Source: Contractor (data), GAO (presentation). [End of figure] The KEI contractor attributes its unfavorable fiscal year cost and schedule variances of $8.3 million and $8.5 million, respectively, to issues with its interceptor booster. Problems initially arose in fiscal year 2007 with a motor case failure during acceptance testing which led to unexpected redesigns. In October 2007, the program experienced several issues with the nozzle during a second stage ground test and also experienced a deviation in measured performance from pre-test predictions. These issues added costly redesigns and delays to its knowledge point, a booster flight test. The program performed a replan of its work in April 2008 because of these issues in order to realign the schedule with their booster flight test knowledge point which was delayed from August 2008 to April 2009. Since the replan, the booster flight test has been further delayed to the fourth quarter of fiscal year 2009. As a result of the replan, the program zeroed out some schedule variances from the baseline to reflect the program's progress toward the newly defined schedule. Despite this replan in April, our analysis shows that the replan has not improved overall performance as cumulative cost and schedule variances continue their downward trend. We were unable to estimate whether the total work under the contract is likely to be completed within budgeted cost since trends cannot be developed until at least 15 percent of the work under the contract is completed. Limited Contractor Data Prevented Analysis of All MKV Task Orders: The Multiple Kill Vehicles (MKV) program began utilizing an indefinite delivery indefinite quantity contract in January 2004. Since then, the program has initiated eight task orders, five of which were open during fiscal year 2008--Task Orders 4 through 8. Task Order 4 provided insufficient data to complete full earned value analysis for the fiscal year. In addition, Task Order 5 was completed shortly after the fiscal year began, without providing enough data to show performance trends. Therefore we performed analysis for Task Orders 6, 7, and 8 as shown below. None of the task orders were realigned during the fiscal year. MKV Task Order 6 began in November 2006 for the component development and testing of a prototype carrier vehicle seeker (a long-range sensor). According to the task order, this seeker for the carrier vehicle will assign individual kill vehicles for target destruction. This task will culminate in a demonstration planned for fiscal year 2010. As shown in figure 13 below, performance data over the course of the fiscal year illustrates declining cost and schedule performance. Although it began the fiscal year with slightly positive cumulative cost and schedule variances, the program ended the year with slightly negative cumulative cost and schedule variances of $1.1 million and $0.6 million respectively. In addition, the contractor has unfavorable fiscal year cost and schedule variances of $1.4 million and $1.5 million, respectively. The program attributes its negative cumulative cost and schedule variances to increased work necessary to resolve software development issues, unplanned efforts as a result of late hardware arrivals, and a government-directed change in vendors for hardware resulting in additional design work. Based on our analysis and the assumption that the contractor will continue to perform as it has through fiscal year 2008, we predict that at its contract completion in May 2009, the contractor on Task Order 6 will overrun its budgeted cost of $19.3 million by between $1.6 million and $2.5 million. Figure 13: MKV Task Order 6 Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $0.266 million; Cumulative Schedule Variance: $0.913 million. Month: October 07; Cumulative Cost Variance: -$0.053 million; Cumulative Schedule Variance: $0.536 million. Month: November 07; Cumulative Cost Variance: -$0.191 million; Cumulative Schedule Variance: $0.382 million. Month: December 07; Cumulative Cost Variance: -$0.436 million; Cumulative Schedule Variance: $0.354 million. Month: January 08; Cumulative Cost Variance: -$0.489 million; Cumulative Schedule Variance: $0.172 million; Month: February 08; Cumulative Cost Variance: -$0.591 million; Cumulative Schedule Variance: $0.226 million. Month: March 08; Cumulative Cost Variance: -$0.602 million; Cumulative Schedule Variance: $0.189 million. Month: April 08; Cumulative Cost Variance: -$0.344 million; Cumulative Schedule Variance: $0.218 million. Month: May 08; Cumulative Cost Variance: -$0.269 million; Cumulative Schedule Variance: $0.151 million. Month: June 08; Cumulative Cost Variance: -$0.554 million; Cumulative Schedule Variance: -$0.067 million. Month: July 08; Cumulative Cost Variance: -$0.718 million; Cumulative Schedule Variance: -$0.319 million. Month: August 08; Cumulative Cost Variance: -$0.93 million; Cumulative Schedule Variance: -$0.47 million. Months: September 08; Cumulative Cost Variance: -$1.128 million; Cumulative Schedule Variance: -$0.616 million. Source: Contractor (data), GAO (presentation). [End of figure] The MKV Task Order 7 is for the development and testing of engagement management algorithms and the test bed in which it will be demonstrated. These algorithms will be a critical capability of the carrier vehicle to manage the kill vehicle engagements relying on target information from the BMDS sensors and the carrier vehicle long- range sensor. The contractor on this task order performed positively during the fiscal year, both in terms of its cumulative and fiscal year cost and schedule variances. The program had a favorable fiscal year cost variance of $1.4 million and a positive fiscal year schedule variance of $11 thousand, adding to its favorable cumulative cost and schedule variances of $1.7 million and $0.1 million, respectively. The program attributes its cumulative cost underruns to several reasons including a programmatic decision to proceed with one approach for organizing kill vehicles in attack formation rather than funding several different approaches. In addition, the contractor experienced cost savings with greater efficiencies than expected in the kill vehicle portion of the work under the contract and less manpower than planned in other portions of the work under the contract. If the contractor continues to perform as it has in the past, we estimate that at completion in May 2010 the work under the contract could cost between $3.2 million and $3.9 million less than the expected $43.9 million budgeted for the work under the contract. See figure 14 below for an illustration of cumulative cost and schedule performance during fiscal year 2008. Figure 14: MKV Task Order 7 Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $0.25 million; Cumulative Schedule Variance: $0.068 million. Month: October 07; Cumulative Cost Variance: $0.252 million; Cumulative Schedule Variance: $0.056 million. Month: November 07; Cumulative Cost Variance: $0.347 million; Cumulative Schedule Variance: $0.029 million. Month: December 07; Cumulative Cost Variance: $0.474 million; Cumulative Schedule Variance: $0.002 million. Month: January 08; Cumulative Cost Variance: $0.588 million; Cumulative Schedule Variance: $0.064 million; Month: February 08; Cumulative Cost Variance: $0.567 million; Cumulative Schedule Variance: $0.044 million. Month: March 08; Cumulative Cost Variance: $0.24 million; Cumulative Schedule Variance: -$0.191 million. Month: April 08; Cumulative Cost Variance: $0.635 million; Cumulative Schedule Variance: -$0.07 million. Month: May 08; Cumulative Cost Variance: $0.846 million; Cumulative Schedule Variance: -$0.074 million. Month: June 08; Cumulative Cost Variance: $1.036 million; Cumulative Schedule Variance: -$0.016 million. Month: July 08; Cumulative Cost Variance: $1.264 million; Cumulative Schedule Variance: -$0.009 million. Month: August 08; Cumulative Cost Variance: $1.269 million; Cumulative Schedule Variance: $0.002 million. Months: September 08; Cumulative Cost Variance: $1.665 million; Cumulative Schedule Variance: $0.079 million. Source: Contractor (data), GAO (presentation). [End of figure] MKV Task Order 8 was awarded in January 2007 and began reporting full performance data in July 2007. The task order is for the development and testing of a hover test bed and hover test vehicle. This hover test bed will allow the program to integrate and test key components of the system in a repeatable ground-based free flight environment as their technologies reach maturity. The program experienced a continuing schedule performance decline as seen in figure 15. Figure 15: MKV Task Order 8 Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $0.379 million; Cumulative Schedule Variance: $0.267 million. Month: October 07; Cumulative Cost Variance: -$0.259 million; Cumulative Schedule Variance: $0.106 million. Month: November 07; Cumulative Cost Variance: -$1.287 million; Cumulative Schedule Variance: -$0.895 million. Month: December 07; Cumulative Cost Variance: -$1.178 million; Cumulative Schedule Variance: -$0.759 million. Month: January 08; Cumulative Cost Variance: -$1.314 million; Cumulative Schedule Variance: -$1.375 million; Month: February 08; Cumulative Cost Variance: -$1.493 million; Cumulative Schedule Variance: -$0.176 million. Month: March 08; Cumulative Cost Variance: -$2.284 million; Cumulative Schedule Variance: $2.186 million. Month: April 08; Cumulative Cost Variance: -$4.147 million; Cumulative Schedule Variance: $2.717 million. Month: May 08; Cumulative Cost Variance: -$5.747 million; Cumulative Schedule Variance: $2.154 million. Month: June 08; Cumulative Cost Variance: -$6.843 million; Cumulative Schedule Variance: $1.906 million. Month: July 08; Cumulative Cost Variance: -$7.854 million; Cumulative Schedule Variance: $1.469 million. Month: August 08; Cumulative Cost Variance: -$9.131 million; Cumulative Schedule Variance: $0.772 million. Months: September 08; Cumulative Cost Variance: -$10.303 million; Cumulative Schedule Variance: $0.252 million. Source: Contractor (data), GAO (presentation). [End of figure] Although the contractor began the year with a positive cumulative cost variance, overruns during the fiscal year of $10.7 million led the program to a total cumulative cost overrun of $10.3 million. The element's fiscal year schedule variance was slightly negative at an unfavorable $15 thousand, leaving its cumulative schedule variance largely unchanged at a favorable $0.3 million. The program attributes the cumulative cost variances to increased labor, procurement, and material costs as well as increased hardware and engineering drawings, and management oversight to resolve subcontractor inefficiencies. In addition, the program increased expenditures to resolve technical and schedule issues associated with the development of avionics subsystems. The planned date for the task order's main effort--completing the hover test--was delayed 2 months from its original date to December 2008 in part because of technical issues associated with the test vehicle's power unit and a software anomaly. These issues were resolved prior to the hover test being conducted. Based on its prior performance, the MKV contractor could overrun the budgeted cost of $48.0 million for the work under the contract at completion in January 2009 by between $5.7 million and $13.8 million. Sensors' Radar Experienced Fiscal Year Cost and Schedule Growth: As of September 2008, the Sensor's contractor had overrun its fiscal year budget by $2.2 million and was behind in completing $27.4 million worth of work. Considering prior years' performance, the contractor is performing under budget with a favorable cumulative cost variance of $22.0 million. However, the contractor has a cumulative unfavorable schedule variance of $9.6 million. The contractor reports the cumulative schedule variance is driven by delays in the manufacturing of the sixth radar and a software capability release that is 2 to 3 months behind schedule. Additionally, the contractor reports that its favorable cumulative cost variance is attributable to efficiencies in the second radar's manufacturing, design, development, and software. The Sensors contractor has not performed a realignment of its work since contract start in April 2003. See figure 16 for trends in the contractor's cost and schedule performance during the fiscal year. Figure 16: Sensors Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: $24.146 million; Cumulative Schedule Variance: $17.828 million. Month: October 07; Cumulative Cost Variance: $22.739 million; Cumulative Schedule Variance: $18.009 million. Month: November 07; Cumulative Cost Variance: $24.037 million; Cumulative Schedule Variance: $17.767 million. Month: December 07; Cumulative Cost Variance: $24.993 million; Cumulative Schedule Variance: $16.444 million. Month: January 08; Cumulative Cost Variance: $26.023 million; Cumulative Schedule Variance: $12.228 million; Month: February 08; Cumulative Cost Variance: $26.135 million; Cumulative Schedule Variance: $6.864 million. Month: March 08; Cumulative Cost Variance: $24.922 million; Cumulative Schedule Variance: $4.215 million. Month: April 08; Cumulative Cost Variance: $24.793 million; Cumulative Schedule Variance: $0.896 million. Month: May 08; Cumulative Cost Variance: $23.734 million; Cumulative Schedule Variance: -$3.31 million. Month: June 08; Cumulative Cost Variance: $15.282 million; Cumulative Schedule Variance: -$5.772 million. Month: July 08; Cumulative Cost Variance: $14.947 million; Cumulative Schedule Variance: -$7.484 million. Month: August 08; Cumulative Cost Variance: $16.51 million; Cumulative Schedule Variance: -$8.162 million. Months: September 08; Cumulative Cost Variance: $21.994 million; Cumulative Schedule Variance: -$9.607 million. Source: Contractor (data), GAO (presentation). [End of figure] The contractor reports that its unfavorable fiscal year schedule variance of $27.4 million is due to a decrease of previously earned positive schedule variances reaped from the manufacturing efficiencies leveraged from the Terminal High Altitude Area Defense radar hardware design. In addition, late delivery of components also contributed to the negative fiscal year schedule variances. The negative fiscal year cost variance of $2.2 million is largely due to a contract change related to its incentive fee. Our analysis predicts that if the contractor continues to perform as it has through fiscal year 2008, the work under the contract could cost from $25 million less to $9.1 million more than the budgeted cost of $1.1 billion at completion currently planned for December 2010. Technical Issues Drove STSS Cost Growth during the Fiscal Year: After a replan of work in October 2007, the Space Tracking and Surveillance System (STSS) contractor experienced an unfavorable cost variance of $87.9 million during the fiscal year. The replan was undertaken in order to extend the period of performance and delay the launch date of its demonstrator satellite. Despite fiscal year cost overruns, the contractor was able to make gains on the cumulative schedule variance by accomplishing $1.9 million more worth of work than was originally planned. Cumulatively, the program has both unfavorable cost and schedule variances at $319.3 million and $17.8 million, respectively. The program attributes cumulative cost variances and schedule variances to continual launch date schedule slippages. In addition, problems in the space segment portion of work also added to the cumulative cost variances. Figure 17 shows both cost and schedule trends during fiscal year 2008. Figure 17: STSS Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$231.41 million; Cumulative Schedule Variance: -$19.667 million. Month: October 07; Cumulative Cost Variance: -$246.178 million; Cumulative Schedule Variance: -$13.029 million. Month: November 07; Cumulative Cost Variance: -$249.783 million; Cumulative Schedule Variance: -$14.222 million. Month: December 07; Cumulative Cost Variance: -$256.466 million; Cumulative Schedule Variance: -$15.784 million. Month: January 08; Cumulative Cost Variance: -$263.5 million; Cumulative Schedule Variance: -$15.202 million; Month: February 08; Cumulative Cost Variance: -$268.303 million; Cumulative Schedule Variance: -$15.858 million. Month: March 08; Cumulative Cost Variance: -$274.939 million; Cumulative Schedule Variance: -$16.97 million. Month: April 08; Cumulative Cost Variance: -$282.524 million; Cumulative Schedule Variance: -$18.962 million. Month: May 08; Cumulative Cost Variance: -$288.702 million; Cumulative Schedule Variance: -$19.985 million. Month: June 08; Cumulative Cost Variance: -$295.914 million; Cumulative Schedule Variance: -$21.368 million. Month: July 08; Cumulative Cost Variance: -$305.378 million; Cumulative Schedule Variance: -$22.753 million. Month: August 08; Cumulative Cost Variance: -$313.263 million; Cumulative Schedule Variance: -$18.452 million. Months: September 08; Cumulative Cost Variance: -$319.261 million; Cumulative Schedule Variance: -$17.752 million. Source: Contractor (data), GAO (presentation). [End of figure] Program cost variances during the fiscal year were driven primarily by technical issues with hardware installed on the second space vehicle. These issues included an overheating flight communications box, a leak on the propulsion side of the satellite, and problems with the spacecraft processor that failed to send a critical command to the onboard computer. To resolve the issues with the processor, the program office initially recommended the removal of the entire computer from the spacecraft. However, after extensive research and testing, the program manager determined that the event with the spacecraft is an unverifiable failure with a low probability of occurrence and low mission impact and decided not to remove the computer from the spacecraft to resolve the issue. We estimate that if the contractor continues to perform as it has through fiscal year 2008, the work under the contract at completion in September 2011 could exceed its budgeted cost of $1.6 billion by between $621.7 million and $1.2 billion. Targets and Countermeasures Program's Rebaseline Positively Affected Fiscal Year Schedule Variances: In June 2008, a delivery order under the Targets and Countermeasures' element that is developing a new family of targets--the Flexible Target Family (FTF)--performed a rebaseline as a result of experiencing manufacturing delays to several components. The majority of the delays were from qualification failures, subsequent redesigns, and requalification efforts. The rebaseline was to realign the work under the contract to reflect realistic hardware delivery dates. This rebaseline did not affect cost variances, but did rebaseline major milestone delivery dates and, as a result, set some of the previously existing schedule variances to zero. The Targets and Countermeasures contractor made gains with a favorable $23.2 million fiscal year schedule variance due in part to the rebaseline in June 2008. However, the contractor ended the year with an unfavorable cumulative schedule variance of $6.4 million which was primarily driven by delays in the completion of the FTF qualification program.[Footnote 52] The program also ended the year with a cumulative cost variance of $52.8 million which the contractor attributed to costs associated with the FTF's avionics components integration and qualification issues, and more effort than expected required on motors for one of the targets in the program. See figure 18 below for an illustration of cumulative cost and schedule variances during the course of the fiscal year. Figure 18: Targets and Countermeasures Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$17.091 million; Cumulative Schedule Variance: -$29.577 million. Month: October 07; Cumulative Cost Variance: -$18.662 million; Cumulative Schedule Variance: -$34.647 million. Month: November 07; Cumulative Cost Variance: -$21.078 million; Cumulative Schedule Variance: -$37.97 million. Month: December 07; Cumulative Cost Variance: -$23.654 million; Cumulative Schedule Variance: -$34.392 million. Month: January 08; Cumulative Cost Variance: -$25.813 million; Cumulative Schedule Variance: -$32.862 million; Month: February 08; Cumulative Cost Variance: -$25.6 million; Cumulative Schedule Variance: -$32.2 million. Month: March 08; Cumulative Cost Variance: -$34.549 million; Cumulative Schedule Variance: -$30.243 million. Month: April 08; Cumulative Cost Variance: -$40.03 million; Cumulative Schedule Variance: -$32.813 million. Month: May 08; Cumulative Cost Variance: -$45.077 million; Cumulative Schedule Variance: -$31.551 million. Month: June 08; Cumulative Cost Variance: -$49.187 million; Cumulative Schedule Variance: -$3.288 million. Month: July 08; Cumulative Cost Variance: -$52.211 million; Cumulative Schedule Variance: -$4.232 million. Month: August 08; Cumulative Cost Variance: -$51.679 million; Cumulative Schedule Variance: -$5.196 million. Months: September 08; Cumulative Cost Variance: -$52.776 million; Cumulative Schedule Variance: -$6.383 million. Source: Contractor (data), GAO (presentation). [End of figure] The contractor attributes its unfavorable fiscal year cost variances of $35.7 million to the increased cost of completing the first four FTF 72- inch targets. Delays in completing component qualification extended the period of performance which invariably led to higher costs. In addition, the contractor cites cost increases to the failure of one of its targets in July 2008 that added mission assurance and testing cost to the follow-on mission using the same target configuration. We estimate that if the contractor continues to perform as it has in the past, it will overrun its budgeted cost of $1.1 billion at contract's end in December 2009 by between $63.7 million and $75.9 million. THAAD Contractor Spent More Money and Time Than Budgeted: The Terminal High Altitude Area Defense (THAAD) program experienced target issues during fiscal year 2008. The THAAD program performed a realignment in May 2008 to extend the flight test program after experiencing several delayed target deliveries. Because of the cost impact of these delayed targets, the program will increase its value of the work under its contract by approximately $80 million. The THAAD program performed a similar realignment in December 2006 as a result of delayed target deliveries as well. As a result of this realignment, the program extended its flight test program and added an estimated $121 million to the value of work under its contract. The THAAD contractor experienced downward trends in its cost and schedule performance during fiscal year 2008. The program overran its budgeted costs for the fiscal year by $33.5 million. It was also unable to accomplish $7.4 million worth of work during the fiscal year. Both of these unfavorable variances added to the negative cumulative cost and schedule variances of $228.7 million and $16.5 million, respectively, as shown in figure 19. Figure 19: THAAD Fiscal Year 2008 Cost and Schedule Performance: [Refer to PDF for image: multiple line graph] Month: September 07; Cumulative Cost Variance: -$195.232 million; Cumulative Schedule Variance: -$9.058 million. Month: October 07; Cumulative Cost Variance: -$196.647 million; Cumulative Schedule Variance: -$9.515 million. Month: November 07; Cumulative Cost Variance: -$187.51 million; Cumulative Schedule Variance: -$8.4 million. Month: December 07; Cumulative Cost Variance: -$187.962 million; Cumulative Schedule Variance: -$9.815 million. Month: January 08; Cumulative Cost Variance: -$187.929 million; Cumulative Schedule Variance: -$10.492 million; Month: February 08; Cumulative Cost Variance: -$192.491 million; Cumulative Schedule Variance: -$10.311 million. Month: March 08; Cumulative Cost Variance: -$199.843 million; Cumulative Schedule Variance: -$10.774 million. Month: April 08; Cumulative Cost Variance: -$206.191 million; Cumulative Schedule Variance: -$11.979 million. Month: May 08; Cumulative Cost Variance: -$213.162 million; Cumulative Schedule Variance: -$14.308 million. Month: June 08; Cumulative Cost Variance: -$217.408 million; Cumulative Schedule Variance: -$12.351 million. Month: July 08; Cumulative Cost Variance: -$221.086 million; Cumulative Schedule Variance: -$15.883 million. Month: August 08; Cumulative Cost Variance: -$223.567 million; Cumulative Schedule Variance: -$14.516 million. Months: September 08; Cumulative Cost Variance: -$228.743 million; Cumulative Schedule Variance: -$16.497 million. Source: Contractor (data), GAO (presentation). [End of figure] The THAAD prime contractor's fiscal year cost overrun of $33.5 million was primarily caused by the radar, missile, and launcher portions of work. Design problems delayed the prime power unit design review and slowed parts production, causing the radar's negative cost trend. In addition, the missile's negative cost trend for this same period was driven by design complexity, ongoing rework/retest of subsystems, unexpected qualification discoveries, and unfavorable labor variances at key subcontractors. Lastly, the launcher variances were driven by hardware and software complexities and higher-than-expected costs for transitioning a portion of this effort to a different facility for production. The contractor reports that its unfavorable fiscal year schedule variance of $7.4 million is primarily driven by the radar and missile components. The radar's negative schedule variance is associated with vendor delays in delivering trailers for both of the system's prime power units. The late delivery of the trailers has subsequently delayed delivery of the prime power units. Missile rework due to qualification test discoveries also negatively affected schedule performance. If the contractor continues to perform as it has through fiscal year 2008, we project that at the contract's completion currently scheduled for September 2009, the contractor could overrun its budgeted cost of $4.6 billion by between $252 million to $274 million. [End of section] Appendix III: FTG-04 Flight Test Cancellation: On May 23, 2008, the Senate Armed Services Committee requested that we review the reasons behind the cancellation of a GMD flight test designated FTG-04. Initially, on May 1, 2008, the Director, MDA decided to delay this test due to problems discovered in a telemetry device, the Pulse Code Modulation Encoder (PCME). This device does not affect operational performance, but rather is a critical component needed to transmit flight test data only. The PCME problems were due in large part to manufacturing defects, which the manufacturers and MDA concluded likely affected all the PCMEs. However, on May 8, the Director of MDA instead decided to cancel this flight test entirely, resulting in one less GMD end-to-end intercept flight test. MDA told us that delaying the flight test until the PCMEs could be repaired would cause delays in future tests since various test assets were shared. MDA officials therefore decided to cancel FTG-04 and transfer some test objectives to other tests, including a new non-intercept flight test, FTX-03, and an already planned intercept flight test, FTG-05. Also, for some remaining objectives not captured in FTG-05 and FTX-03, MDA stated that it planned a third intercept test, FTG-X. We were asked to investigate this test cancellation and answer the following questions: * Why did the MDA change its initial decision to delay FTG-04 until November 2008 and decide to cancel FTG-04 instead and what deliberative process did MDA follow in deciding to cancel FTG-04? * When and how, if ever, will each of the specific test objectives previously planned for FTG-04 be accomplished? * What are the implications of canceling this flight test on the ensuing test program, on demonstrating the capability of the GMD system, and on other programmatic decisions? Faulty Telemetry Component Caused Delay and Subsequent Cancellation of FTG-04: MDA initially delayed the FTG-04 flight test because of defects in the PCME, a telemetry component in the Exoatmospheric Kill Vehicle (EKV) only needed to gather test data. Although the PCME does not affect operational performance, it is needed for test assets to determine if design and operational issues have been resolved. The FTG-04 had four prior delays and was originally scheduled for the first quarter of fiscal year 2007. In responding to these delays, multiple tests over several years were affected. Several defects contributed to the problem, the first three of which are presumed to affect all PCMEs manufactured up to that point and all 24 fielded Test Bed/Capability Enhancement (CE) I EKVs: * The PCMEs experienced gold embrittlement due to lack of pretinning. * Insufficient oscillator stand-off height increased thermal stress. * Circuit board deflection caused by three washers missing from the board. In addition to these manufacturing defects, there were stress fractures in the solder of three PCMEs caused by the removal and replacement of a chip on the device. This chip was removed because a clock on a chip was asynchronous with another component's clock. It was estimated that there was an 18 to 48 percent chance of the loss of telemetry data at some point during a flight test due to the asynchronous chip problem. Again, all 24 fielded Test Bed/CE-I EKVs have the chip with this problem. This chip does not affect operational performance, but rather is a critical component needed to transmit flight test data only. See table 10 for timeline of events related to this cancellation. Table 10: Timeline of Events: 1/12/08-2/4/08; During early tests of the Payload 33 PCME at the subcontractor's facility, no failures were detected. 2/7/08; During final test readiness reviews at Vandenburg Air Force Base, the first failure was identified. 2/8/08-2/22/08; Trouble shooting isolates problem to PCME and EKV is returned to contractor for removal of PCME. 2/22/2008; MDA de-emplaced interceptor as a replacement (Payload 32). 3/3/08-3/26/08; Troubleshooting continues, failures are repeatable but intermittent. 3/28/08-4/2/08; Fault isolated to oscillator and solder joints. 4/30/08-5/1/08; Tiger Team formed to assess risk and presented risk assessment to the Director, MDA. 5/1/08; Director, MDA delays FTG-04 test. 5/8/08; Program Change Board recommends cancellation; Director, MDA makes the decision to cancel, replace it with FTX-03 and informs Congress. Source: GAO presentation; MDA documentation. [End of table] The contractor, Boeing, and the subcontractors, Raytheon and the manufacturer of the component, L-3, took actions to mitigate the problem. They eliminated the gold embrittlement problem by sending the oscillator out for pretinning, they designed custom washers for two already produced PCMEs and raised three bosses for new PCMEs to eliminate the need for washers, and they tightened tolerances on the board to eliminate the deflection issue. These first three PCME manufacturing improvements were finalized on May 16, 2008. They also made changes to correct the chip with the asynchronous clock problem in all newly manufactured PCMEs. None of the previously fielded GBIs will be refurbished with improved PCMEs needed for flight tests, but the GBIs emplaced starting in October 2008 and thereafter have the improved PCME. On May 1, 2008, MDA's Program Change Board considered five options. 1. Execute FTG-04 as scheduled, using payload (32) "as is": 2. Continue diagnostic testing of payload 32, but if decision was made that it was not ready, substitute payload 33, leading to a delay in the test schedule: 3. Refurbish payload 32, but if it did not improve, substitute payload 33: 4. Immediately replace payload 32 with 33 without further testing: 5. Immediately return payload 32 for repair: The Director, MDA chose option 5, delaying the FTG-04 into the November to December 2008 timeframe, but keeping the program on track to provide this intercept data as planned. MDA consulted the test community, Director, Operational Test and Evaluation (DOT&E) and the BMDS Operational Test Agency, on this initial decision to delay the test and both agreed with this decision. According to MDA, the Director also asked for options for a sensor test in the summer of 2008. On May 8, 2008, MDA's Program Change Board reconvened to consider three options for a sensor test (FTX-03) and canceling instead of delaying FTG-04: 1. Conduct FTX-03, with a baseline like the planned FTG-04, but without a live intercept attempt. 2. Conduct FTX-03 with a baseline like the FTG-05, an intercept flight test to be conducted in December 2008. 3. Similar to option 2, but with more sensor data collected. The Director MDA changed the May 1 decision to delay, refurbish and fly the planned FTG-04 test and chose instead to cancel FTG-04 and pursue the modified option 3 above. Choosing option 3 resulted in restructuring the intercept test into a test designed to assess multiple sensor integration capability. This new test benefited sensor modeling and simulation as post-flight reconstruction could occur now on two missions. According to MDA, it canceled the FTG-04 at the May 8, 2008 meeting instead of delaying it, in part, because rescheduling FTG-04 would have caused a major delay in another test, the Distributed Ground Test-03 (GTD-03). GTD-03 and FTG-04 required many of the same assets, so conducting FTG-04 would have delayed GTD-03 and thus the delivery of this new capability by four months. Ground tests assess the increased BMDS capability to be fielded next and GTD-03 was to provide the means by which a more realistic simulation of threats and scenarios and the means by which new software capability could be declared ready to move into the operational baseline. MDA consulted with BMDS Operational Test Agency officials on this decision and they supported it. DOT&E was not consulted on this decision and expressed concern that the elimination of any intercept test reduced the opportunity to gather additional data that might have increased confidence in models and simulations. DOT&E has repeatedly expressed concerns over the lack of test data needed to validate MDA's models and simulations. Most FTG-04 Test Objectives Will Be Allocated to Follow-on Tests: According to MDA, all FTG-04 test objectives were allocated to other flight tests. However, partly due to differences in how MDA describes test objectives, it is unclear whether all planned FTG-04 test objectives will be accomplished in follow-on tests. The loss of a primary objective, an intercept of a complex target scene, will slow MDA's efforts to build confidence in the EKV's ability to consistently achieve intercepts, unless an additional intercept is scheduled. In August 2008, MDA informed Congress that it planned to conduct a new intercept test called FTG-X in fiscal year 2009. However, in January 2009 MDA stated that the FTG-X intercept test was never formally approved and is no longer planned. In addition, some test objectives related to modeling and simulations have been redefined so it is unclear whether they will be fully tested. Models and simulations are critical to understanding and assessing the performance of the BMDS because flight tests are limited by their cost, complexity, and range safety constraints. Modeling and simulation is therefore the primary way to fully assess the overall performance of the BMDS and its various components. According to DOT&E, cancellation of FTG-04 reduced interceptor and EKV data available for modeling, leaving only two intercepts (FTG-3a and FTG-05) that have provided complete sets of information. In October 2008, MDA stated that modification of FTG-04 into a sensor test eliminated a second opportunity to anchor the models of EKV-fielded software. Test objectives in MDA planning documents describe modeling and simulation objectives at a high level. However, it is difficult to determine whether modeling and simulation objectives are addressed in the near term because the objectives are defined differently for each test. For example, the FTX-03 and FTG-05 objectives do not distinguish between primary and secondary objectives while the FTG-04 does. One objective that seems to be absent in the FTG-05 and FTX-03 is to collect data to support validation and anchoring of system-level (vs. element-level) simulations, MDA stated that the exclusion from the test objectives was inadvertent and it will be addressed by the tests. BMDS Operational Test Agency objectives related to the GBI engagement were not met, although the test agency officials indicate the majority of their non-intercept sensor related objectives for FTG-04 were met in the FTX-03 test. However, BMDS Operational Test Agency officials state that some of their intercept objectives may be addressed through a combination of previous intercept test, FTG-03a, and recently conducted FTG-05. Finally, several warfighter objectives for FTG-04, related to tactics, techniques, and procedures, will be met through the ground tests instead because, according to the warfighter representative at the BMDS Operational Test Agency, flight tests do not offer the best opportunity to assess this kind of objective. Cancellation Eliminates One of Few Opportunities to Demonstrate GMD Capabilities: The cancellation has increased the strain on the ensuing test program. GMD's current plans call for two intercept attempts in fiscal year 2009--FTG-05, which was conducted in December 2008, and FTG-06--and one booster verification test. This is an ambitious schedule as GMD has been able to conduct only one intercept flight test per year--FTG-02 in September 2006, FTG-03a in September 2007 and FTG-05 in December 2008. MDA had planned to conduct five intercept tests with varying stresses to assess the EKV capability between February 2007 and December 2008. Flight test failures and test plan revisions caused MDA to only carry out two intercept tests in that period--FTG-03a and FTG-05--both of which resulted in an intercept. In addition, the number of future flight tests planned has been reduced. MDA has not funded or scheduled an intercept replacement for FTG-04. In January 2008 MDA decided to merge two intercept tests--FTG- 06 and FTG-07--into one single intercept attempt. This merger removes another opportunity to gather end-game EKV performance data needed to assess capability. In January 2008 MDA also decided to accelerate a two- stage verification non-intercept test required to assess the European component. The cancellation of FTG-04 removed one chance to obtain end-game performance data needed to develop GMD models and to assess the capability of the CE-I EKV. The repetition of intercept-related objectives is important to build confidence in the intercept capability. These models are the primary way to fully assess the overall system performance, since flight tests are limited by their cost, complexity and range safety concerns. MDA planned to test the CE- I EKV against a dynamic target scene with countermeasures in both FTG- 04 and FTG-05. FTG-04 was canceled and an FTG-05 target anomaly affected this objective. According to MDA, no more CE-I EKV flight tests have been approved, although it is considering whether to conduct an intercept test using a CE-I EKV in the future. GMD developed some mitigations to various developmental issues, but realistic flight testing is needed to anchor the models and to determine the effectiveness of these mitigations. The test cancellation and target problems have reduced the knowledge that MDA expected to use for its upcoming end-to-end performance assessment. Performance assessments are annual system-level assessments to test, evaluate, and characterize the operational capability of the BMDS as of the end of the calendar year. Currently, MDA has only completed one--Performance Assessment 2007. Furthermore, acting on a joint recommendation between MDA and the Operational Test Agency, MDA officials canceled their 2008 performance assessment efforts in April 2008 because of developmental risks associated with modeling and simulations. Instead, MDA is focusing on testing and models for Performance Assessment 2009. However, the planned performance information available for Performance Assessment 2009 will be reduced. The FTG-04 cancellation reduced one set of data that was expected to be available. In addition, both FTX-03 and FTG-05 will be used to anchor data for Performance Assessment 2009, but target anomalies in each test precluded the completion of all planned test objectives. Neither target presented the complexity needed for advanced algorithm development. Manufacturing and emplacement continue unabated by reductions and delays in tests. Twenty-four CE-I GBIs have been fielded and the new CE- II GBIs are now being fielded without important knowledge about the systems capabilities expected to be gained through tests. The first CE- II GBI emplacement occurred prior to any flight testing of this configuration. The first flight test is FTG-06 currently scheduled to occur no earlier than the fourth quarter of fiscal year 2009. According to MDA, these CE-II GBIs will not be declared operational until after the successful completion of FTG-06. FTG-04 was also identified as a key source of data supporting a number of capabilities declarations. The cancellation of FTG-04, plus other testing delays, prompted MDA to defer some capabilities and to declare others based on previous tests. Conclusions: The cancellation of the FTG-04 flight test increases the risk to the GMD program and to the overall BMDS capability, since the lack of adequate intercept data adversely affects confidence that the system could perform as intended in a real-world situation. The GMD program has reduced its plans to assess operational performance of the fielded configuration between February 2007 and December 2008 from five to two intercept tests, leaving gaps in knowledge about the repeatability of the performance of fielded assets. In addition, the opportunity to obtain additional intercept data vital to the anchoring of models and simulations has been lost, unless the FTG-X flight test is conducted, adding to an existing concern expressed by DOT&E. Despite test reductions and effects on assessing system-level performance, production and fielding of assets continues as planned. [End of section] Appendix IV: Reduced Basis for Capability Declarations: The two tables below list test events supporting MDA capability declarations during fiscal year 2008 for certain engagement sequence groups in Blocks 1.0 through 3.0 (see table 11), as well as for the full completion of Bock 1.0 by the end of fiscal year 2009 (see table 12). Both tables illustrate that MDA reduced the basis for declaring certain engagement sequence groups as early or fully capable. The basis for declaring an early, partial, or full capability includes flight and ground tests as well as performance assessments. Table 11: Engagement Sequence Groups with Revised Basis for Fiscal Year 2008 Capability Declarations: Engagement sequence: Block 1.0: GBI Launch on COBRA DANE/Upgraded Early Warning Radar; Capability declaration: Early; Planned basis for 2008 declaration: PA 07; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 1.0: GBI Launch on COBRA DANE/Upgraded Early Warning Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTD-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Launch on COBRA DANE/Upgraded Early Warning Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTI-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Launch on COBRA DANE/Upgraded Early Warning Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTG-04-3. Engagement sequence: Block 1.0: GBI Engage on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: PA 07; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 1.0: GBI Engage on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTD-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Engage on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTI-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Engage on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTG-04-3. Engagement sequence: Block 1.0: GBI Launch on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: PA 07; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 1.0: GBI Launch on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTD-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Launch on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTI-02; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 1.0: GBI Launch on Sea-based X-band Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTG-04-3. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: PA 08; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: FTM-15; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-03; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-03; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: GTX-03[A]; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: FTM-10,-11,-12,-13. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTD-02. Engagement sequence: Block 2.0: SM-3 Engage on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTI-02. Engagement sequence: Block 2.0: SM-3 Launch on Remote shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTD-02; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Launch on Remote shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTI-02; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Launch on Remote shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: FTM-13; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 2.0: SM-3 Launch on Remote shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: FTM-14. Engagement sequence: Block 2.0: THAAD Engage on AN/TPY-2 (terminal mode); Capability declaration: Early; Planned basis for 2008 declaration: FTT-09; Revised basis for 2008 declaration[A]: [Check]. Engagement sequence: Block 2.0: THAAD Engage on AN/TPY-2 (terminal mode); Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: FTT-07,-08. Engagement sequence: Block 2.0: THAAD Engage on AN/TPY-2 (terminal mode); Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTI-02,-03. Engagement sequence: Block 2.0: THAAD Engage on AN/TPY-2 (terminal mode); Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTX-03a. Engagement sequence: Block 3.0: GBI Launch on shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: GTX-03[A]; Revised basis for 2008 declaration[A]: Dropped. Engagement sequence: Block 3.0: GBI Launch on shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTD-02. Engagement sequence: Block 3.0: GBI Launch on shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTI-02. Engagement sequence: Block 3.0: GBI Launch on shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: GTI-03. Engagement sequence: Block 3.0: GBI Launch on shipboard Aegis Radar; Capability declaration: Early; Planned basis for 2008 declaration: [Empty]; Revised basis for 2008 declaration[A]: FTX-03. Source: GAO analysis of MDA data. [A] Planned assessment or test was actually used for the capability declaration indicated by a "Check" in this column. [End of table] Table 12: Block 1.0 Engagement Sequence Groups with Revised Basis for Completion at End of Fiscal Year 2009: Engagement sequence: GBI Engage on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: PA-07; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: GTD-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: GTI-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Partial; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Partial; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Partial; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Partial; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: GTD-02. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Partial; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: FTI-02. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: PA-08; Revised basis for 2009 declaration[A]: PA-09 Quick Look. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on COBRA DANE (Beale Air Force Base, CA); Capability declaration: Full; Planned basis for 2008 declaration: FTG-05; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Empty]. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: [Empty]. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: PA-08; Revised basis for 2009 declaration[A]: PA-09 Quick Look. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-05; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Partial; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: PA-08; Revised basis for 2009 declaration[A]: PA-09 Quick Look. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-03; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on Sea-based X-band radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-05; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: PA 07; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: GTI-03. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: FTX-03. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: PA 07; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: GTI-03. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-02; Revised basis for 2009 declaration[A]: [Check]. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on forward-based AN/TPY-2 radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: FTX-03. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: PA 07; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: GTI-03. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-02; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-02; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Engage on shipboard Aegis radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: FTX-03. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: PA 07; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: GTI-03. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: GTD-02; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: GTI-02; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: FTG-04; Revised basis for 2009 declaration[A]: Dropped. Engagement sequence: GBI Launch on shipboard Aegis Radar; Capability declaration: Full; Planned basis for 2008 declaration: [Empty]; Revised basis for 2009 declaration[A]: FTX-03. Source: GAO analysis of MDA data. [A] Planned assessment or test for the capability declaration that hasn't changed is indicated by a "Check" in this column. [End of table] [End of section] Appendix V: Scope and Methodology: To examine the progress MDA made in fiscal year 2008 toward its cost, schedule, testing, and performance goals, we examined the efforts of 10 BMDS elements that MDA is developing and fielding. The elements included in our review collectively accounted for 80 percent of MDA's fiscal year 2008 research and development budget requests. In assessing each element, we examined the BMDS Fiscal Year 2008 Statement of Goals, Program Execution Reviews, test plans and reports, production plans, Contract Performance Reports, MDA briefings, and earned value management data. We developed data collection instruments that were completed by MDA and each element program office. The instruments gathered detailed information on planned and completed program activities including tests, design reviews, prime contracts, estimates of element performance, and challenges facing the elements. In addition, we discussed fiscal year 2008 progress and performance with officials in MDA's Agency Operations Office, each element program office, as well as the Office of DOD's Director, Operational Test and Evaluation, and DOD's Operational Test Agency. To assess each element's progress toward its cost goals, we reviewed Contract Performance Reports and, when available, the Defense Contract Management Agency's analyses of these reports. We applied established earned value management techniques to data captured in Contract Performance Reports to determine trends and used established earned value management formulas to project the likely costs of prime contracts at completion. To evaluate the sufficiency of MDA's modeling and simulation practices, we reviewed DOD and MDA policies, memos, flight and test plans related to modeling and simulations, the Acquisition Modeling and Simulation Master plan, as well as verification, validation and accreditation plans and reports for various elements, and MDA white papers discussing modeling and simulation techniques. We also interviewed officials in element program offices to discuss modeling and simulation plans and procedures particular to each. In assessing MDA's accountability, transparency, and management controls, we interviewed officials from the Office of the Under Secretary of Defense's Office for Acquisition, Technology, and Logistics, as well officials in the MDA Agency Operations Directorate. We also reviewed an Institute for Defense Analysis study, two Congressional Research Service reports, a Congressional Budget Office report, U.S. Code, DOD acquisition system policy, various DOD directives, the Missile Defense Executive Board charter, and various MDA statements and documents related to the agency's block structure. To ensure that MDA-generated data used in our assessment are reliable, we evaluated the agency's management control processes. We discussed these processes with MDA senior management. In addition, we confirmed the accuracy of MDA-generated data with multiple sources within MDA and, when possible, with independent experts. To assess the validity and reliability of prime contractors' earned value management systems and reports, we interviewed officials and analyzed audit reports prepared by the Defense Contract Audit Agency. Finally, we assessed MDA's internal accounting and administrative management controls by reviewing MDA's Federal Manager's Financial Integrity Report for Fiscal Years 2003, 2004, 2005, 2006, 2007, and 2008. Our work was performed primarily at MDA headquarters in Arlington, Virginia. At this location, we met with officials from the Aegis Ballistic Missile Defense Program Office; Airborne Laser Program Office; Command, Control, Battle Management, and Communications Program Office; MDA's Agency Operations Office; DOD's Office of the Director, Operational Test and Evaluation; and the Office of the Under Secretary of Defense for Acquisition, Technology and Logistics. In addition, in Huntsville, Alabama, we met with officials from the Ground-based Midcourse Defense Program Office, the Sensors Program Office, the Terminal High Altitude Area Defense Project Office, the Kinetic Energy Interceptors Program Office, the BMDS Kill Vehicles Program Office, the Targets and Countermeasures Program Office, and the Office of the Director for BMDS Tests. We also met with Space Tracking and Surveillance System officials in El Segundo, California. In December 2007, the conference report accompanying the National Defense Authorization Act for Fiscal Year 2008 noted the importance of DOD and MDA providing information to GAO in a timely and responsive manner to facilitate the review of ballistic missile defense programs. During the course this audit, we experienced significant delays in obtaining information from MDA. During the audit, MDA did not provide GAO staff with expeditious access to requested documents which delayed some audit analysis and contributed to extra staff-hours. Of the documents we requested, we received approximately 19 percent within the 10-15 business day protocols that were agreed upon with MDA. Pre- existing documentation took MDA on average about 50 business days to provide and many pre-existing documents took over 100 business days to be provided to GAO. Notwithstanding these delays, we were able to obtain the information needed to satisfy our objectives in accordance with generally accepted government auditing standards. We conducted this performance audit from May 2008 to March 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: Paul Francis (202) 512-4841 or francisp@gao.gov: Acknowledgments: In addition to the contact named above, David Best, Assistant Director; LaTonya Miller; Beverly Breen; Ivy Hübler; Tom Mahalek; Steven Stern; Claire Cyrnak; Isabella Johnson; Meredith Allen Kimmett; Kenneth E. Patton; Karen Richey; and Alyssa Weir made key contributions to this report. [End of section] Footnotes: [1] This initial BMDS capability was to defend the U.S. homeland, deployed troops, friends, and allies against ballistic missiles of all ranges in all phases of flight. MDA was tasked with carrying out the President's direction. [2] According to MDA, the agency was expected to field an initial increment of missile defense capability that provides initial protection of the entire United States from North Korea, partial protection of the United States from the Middle East threat and protection of deployed forces, allies, and friends with terminal defenses. MDA fielded a limited capability that included initial versions of Ground-based Midcourse Defense; Aegis Ballistic Missile Defense; Patriot Advanced Capability-3; and Command, Control, Battle Management, and Communications elements. MDA expected to enhance these initial capabilities and over time, produce an overarching BMDS capable of protecting the United States, deployed forces, friends, and allies from ballistic missile attacks of all ranges. [3] National Defense Authorization Act for Fiscal Year 2002, Pub. L. No. 107-107, § 232(g) (2001); John Warner National Defense Authorization Act for Fiscal Year 2007, Pub. L. No. 109-364, § 224 (2006); and National Defense Authorization Act for Fiscal Year 2008, Pub. L. No. 110-181, § 225 (2007). [4] We did not assess MDA's progress in fiscal year 2002 as the agency did not establish goals for that fiscal year. We delivered the following reports for fiscal years 2003 through 2007: GAO, Missile Defense: Actions Are Needed to Enhance Testing and Accountability, [hyperlink, http://www.gao.gov/products/GAO-04-409] (Washington, D.C.: Apr. 23, 2004); Defense Acquisitions: Status of Ballistic Missile Defense Program in 2004, [hyperlink, http://www.gao.gov/products/GAO-05-243] (Washington, D.C.: Mar. 31, 2005); Defense Acquisitions: Missile Defense Agency Fields Initial Capability but Falls Short of Original Goals, [hyperlink, http://www.gao.gov/products/GAO-06-327] (Washington, D.C.: Mar.15, 2006); Defense Acquisitions: Missile Defense Acquisition Strategy Generates Results but Delivers Less at a Higher Cost, [hyperlink, http://www.gao.gov/products/GAO-07-387] (Washington, D.C.: Mar.15, 2007); and Defense Acquisitions: Progress Made in Fielding Missile Defense, but Program Is Short of Meeting Goals, [hyperlink, http://www.gao.gov/products/GAO-08-448] (Washington, D.C.: Mar.14, 2008). [5] The BMDS also includes an 11th element, Patriot Advanced Capability- 3, which has been transferred to the Army for production, operation, and sustainment. This report does not evaluate Patriot Advanced Capability-3 because its initial development is complete and is now being managed by the Army. [6] Beginning in 2009, this report--previously referred to as the Statement of Goals--will be called the BMDS Accountability Report. [7] BMDS hardware and software are grouped into engagement sequence groups, each of which is the specific combination of all sensors, weapons, and C2BMC capability that are needed to detect, track, and intercept a particular threat. The engagement sequence group construct was created as an engineering tool to provide a simple representation of BMDS capabilities, integration, and functionality and is defined as a unique combination of detect-control-engage functions performed by BMDS subsystems used to engage a threat ballistic missile. [8] 10 U.S.C. § 2435 requires an approved program baseline for major defense acquisition programs. The BMDS program meets the definition of a major defense acquisition program, which is defined in 10 U.S.C. § 2430; however, the requirement to establish a baseline is not triggered until entry into system development and demonstration. Because the BMDS has not yet formally entered the acquisition cycle, it has not yet been required to meet the minimum requirements of section 2435. [9] The Fiscal Year 2005 National Defense Authorization Act, Pub. L. No. 108-375 § 234(e), required the Director, MDA, to establish and report annually to Congress a cost, schedule, and performance baseline for each block configuration being fielded. Modification to the baseline and variations against the baseline must also be reported. In addition, the National Defense Authorization Act for Fiscal Year 2008, Pub. L. No. 110-181 § 223(g) required that no later than the submittal of the budget for fiscal year 2009, MDA shall "establish acquisition cost, schedule and performance baselines" for BMDS elements that have entered the equivalent of system development and demonstration or are being produced and acquired for operation fielding. [10] 10 U.S.C. § 2432 requires DOD to submit to Congress a Selected Acquisition Report (SAR) for each major defense acquisition program that includes the program acquisition unit cost for each program. Unless waived by the Secretary of Defense, this requirement applies when funds have been appropriated for the program and the program has proceeded to system development and demonstration. Development programs that have not entered system development and demonstration may submit a limited SAR. MDA submits a limited SAR that does not include program acquisition unit costs. [11] 10 U.S.C. § 2433 (commonly referred to as Nunn-McCurdy). [12] National Defense Authorization Act for Fiscal Year 2008, Pub. L. No. 110-181 § 223(g) requires MDA to provide unit cost reporting data for each BMDS element that has entered the equivalent of the systems development and demonstration phase of acquisition or is being produced and acquired for operational use and secure independent estimation and verification of such cost reporting data. How MDA was to calculate these unit costs was not specified. [13] 10 U.S.C. § 2432. [14] Flyaway cost refers to the cost of procuring prime mission equipment (e.g., an aircraft, ship, tank, etc.). It is funded with Procurement appropriations and is part of the Investment cost category. This term includes the Work Breakdown Structure elements of Prime Mission Equipment, System Engineering/Program Management, System Test and Evaluation, Warranties, and Engineering Changes. [15] Earned Value Management is a program management tool that integrates the technical, cost, and schedule parameters of a contract. During the planning phase, an integrated baseline is developed by time phasing budget resources for defined work. As work is performed and measured against the baseline, the corresponding budget value is "earned." [16] According to the Over Target Baseline and Over Target Schedule Handbook, May 7, 2003. [17] The GMD program's replan began in fiscal year 2008 but is currently ongoing and, as such, was not totaled into table 4. [18] To determine if contractors are executing the work planned within the funds and time budgeted, each BMDS program office requires its prime contractor to provide monthly reports detailing cost and schedule performance. [19] We analyzed three task orders for the MKV program issued as part of an Indefinite Delivery/Indefinite Quantity contract as well as three contracts managed by the Aegis BMD element. We assessed one contract for each of the other eight elements. Indefinite Delivery/Indefinite Quantity contracts provide for an indefinite quantity, within stated limits, of supplies or services during a fixed period. [20] The current budgeted costs at completion are as-of September 30, 2008. [21] Situational awareness is defined as the degree to which the perception of the current environment mirrors reality. [22] According to program officials, an earlier test--FTG-02--provided limited data for assessment purposes. However, the data was incomplete and could not be used to fully verify and validate the models and simulations. [23] AN/TPY-2 was formerly known as Forward-Based X-Band Transportable radar. [24] The BMDS Operational Test Agency conducts independent operational assessments of BMDS capability to defend the United States, its deployed forces, friends, and allies against ballistic missiles of all ranges and in all phases of flight. MDA funds all BMDS Operational Test Agency activities. [25] There are various categories of knowledge points. For example, an element knowledge point is based on an element event that provides critical information for a key element program decision requiring the Program Manager's approval. Element knowledge points support one or more Director knowledge points, and may be supported by other knowledge points. [26] A sound business case demonstrates that (1) the identified needs are real and necessary and are best met with the chosen concept and (2) the chosen concept can be developed and produced with existing resources--such as technical, knowledge, funding, time and management capacity. [27] GAO, Defense Acquisitions: Sound Business Case Needed to Implement Missile Defense Agency's Targets Program, [hyperlink, http://www.gao.gov/products/GAO-08-1113] (Washington, D.C.: Sept. 26, 2008). [28] The accreditation of models and simulations is an official certification that a model or simulation is acceptable for use as their developers intended. Before a decision to accredit a model, MDA must first verify that the models and simulations operate as the designers conceptualized, and then validate that the models are sufficiently accurate representations of real-world conditions for their intended purposes. [29] An end-to-end simulation represents a complete BMDS engagement-- from enemy missile launch to attempted intercept by BMDS kill vehicle. [30] The BMDS Operational Test Agency provides an independent accreditation of MDA models and simulations. [31] The BMDS Operational Test Agency defines artificialities as BMDS architecture, targets, procedures, and conditions that exist in flight tests but would not exist in the real world. Flight test artificialities are introduced for a number of reasons, such as increased chances of success, range safety, data collection, and asset availability. [32] Weather conditions include rain, clouds, and snow. Severe sea states, ice loads, or winds could render tests unsafe to execute. [33] Post-flight reconstruction is the process of manually recreating and running a past flight test scenario in a simulated environment. [34] AN/TPY-2 radars were formerly known as Forward-Based X-Band- Transportable radars. According to MDA, an additional AN/TPY-2 radar has been provided and is undergoing Government ground testing. [35] The National Defense Authorization Act for Fiscal Year 2002, Pub. L. No. 107-107, § 234(c), states that "for ground-based midcourse interceptor systems, the Secretary of Defense shall initiate steps during fiscal year 2002 to establish a flight test capability of launching not less than three missile defense interceptors and not less than two ballistic missile targets to provide a realistic test infrastructure." Currently, GMD has not conducted this test. [36] 10 U.S.C. § 2399. [37] See app III which details reasons for the FTG-04 cancellation. [38] GTD-03 was delayed to support warfighter needs and resulted in delayed capability assessments and capability declaration later than planned. [39] The accreditation of models and simulations is an official certification that a model or simulation is acceptable for use as their developers intended. Before a decision to accredit a model, MDA must first verify that the models and simulations operate as the designers conceptualized, and then validate that the models are sufficiently accurate representations of real-world conditions for their intended purposes. [40] GTI-02 included models and simulations. [41] Our assessment of engagement sequence groups utilizes MDA's plans as of October 1, 2007. Commensurate with the new block structure, MDA devised a baseline in February 2008 for engagement sequence groups for Blocks 1.0, 2.0, 3.1, and 3.2. [42] [hyperlink, http://www.gao.gov/products/GAO-08-448]. [43] Section 223(b) of the National Defense Authorization Act of 2008, Pub. L. No. 110-181 (Jan. 28, 2008) specified a revised budget structure of the missile defense budget to be submitted in the President's budget no later than the first Monday in February. 31 U.S.C. §1105. [44] Lead services have already been designated for Aegis BMD, the AN/ TPY-2 radar, THAAD, GMD, ABL, the European radar, Cobra Dane, and upgraded early warning radars. [45] Before a program can enter the system development and demonstration phase of the acquisition cycle, statute requires that certain information be developed. 10 U.S.C. § 2366b. In 2002, the Secretary of Defense deferred the application of some of DOD's acquisition processes to BMDS. Therefore, MDA has not yet entered System Development and Demonstration which would trigger the statutes requiring the development of information that the Defense Acquisition Board uses to inform its decisions. Most major defense acquisition programs are also required by statute to obtain an independent verification of program cost prior to beginning system development and demonstration, and/or production and deployment. 10 U.S.C. § 2434. Statute also requires an independent verification of a system's suitability for and effectiveness on the battlefield before a major defense acquisition program can proceed beyond low-rate initial production. 10 U.S.C. § 2399. [46] [hyperlink, http://www.gao.gov/products/GAO-08-448]. [47] [hyperlink, http://www.gao.gov/products/GAO-07-387]. [48] Congressional Research Service, Defense Procurement: Full Funding Policy--Background, Issues, and Options for Congress (Oct. 20, 2006). [49] [hyperlink, http://www.gao.gov/products/GAO-08-1113]. [50] Earned Value Management is a program management tool that integrates the technical, cost, and schedule parameters of a contract. During the planning phase, an integrated baseline is developed by time phasing budget resources for defined work. As work is performed and measured against the baseline, the corresponding budget value is "earned." Using this earned value metric, cost and schedule variances can be determined and analyzed. [51] The total contract cost at completion is based on budgeted cost at completion for each contract we assessed. The budget at completion represents the total planned value of the contract. [52] The current budgeted costs at completion are as-of September 30, 2008. [53] MDA is developing the FTF to represent evolving threats of all ranges. MDA has narrowed its FTF development efforts, focusing on a single vehicle, the 72-inch LV-2 ground-launched target. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO‘s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO‘s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.