Missile Defense

Additional Knowledge Needed in Developing System for Intercepting Long-Range Missiles Gao ID: GAO-03-600 August 21, 2003

A number of countries hostile to the United States and its allies have or will soon have missiles capable of delivering nuclear, biological, or chemical weapons. To counter this threat, the Department of Defense's (DOD's) Missile Defense Agency (MDA) is developing a system to defeat ballistic missiles. MDA expects to spend $50 billion over the next 5 years to develop and field this system. A significant portion of these funds will be invested in the Ground-based Midcourse Defense (GMD) element. To field elements as soon as practicable, MDA has adopted an acquisition strategy whereby capabilities are upgraded as new technologies become available and is implementing it in 2-year blocks. Given the risks inherent to this strategy, GAO was asked to determine when MDA plans to demonstrate the maturity of technologies critical to the performance of GMD's Block 2004 capability and to identify the estimated costs to develop and field the GMD element and any significant risks with the estimate.

GMD is a sophisticated weapon system being developed to protect the United States against limited attacks by long-range ballistic missiles. It consists of a collection of radars and a weapon component--a three-stage booster and exoatmospheric kill vehicle--integrated by a centralized control system that formulates battle plans and directs the operation of GMD components. Successful performance of these components is dependent on 10 critical technologies. MDA expects to demonstrate the maturity of most of these technologies before fielding the GMD element, which is scheduled to begin in September 2004. However, the agency has accepted higher cost and schedule risks by beginning integration of the element's components before these technologies have matured. So far, MDA has matured two critical GMD technologies. If development and testing progress as planned, MDA expects to demonstrate the maturity of five other technologies by the second quarter of fiscal year 2004. The radar technologies are the least mature. MDA intends to demonstrate the maturity of an upgraded early warning radar in California in the first quarter of fiscal year 2005 and a sea-based radar in the Pacific Ocean in the fourth quarter of that year. Although MDA does not plan to demonstrate the maturity of the technology of the early warning radar in Alaska, which will serve as the primary fire control radar, through its own integrated flight tests, it may be able to do so through the anticipated launch of foreign test missiles. MDA estimates that it will spend about $21.8 billion between 1997 and 2009 to develop the GMD element. This estimate includes $7.8 billion to develop and field the GMD Block 2004 capability. For example, the funds will be used to install interceptors at two sites, upgrade existing radars and testing infrastructure, and develop the sea-based X-band radar. We found that MDA has incurred a greater risk of cost growth because for more than a year the agency was not able to rely fully on data from its primary tool for monitoring whether the GMD contractor has been performing work within cost and on schedule. In February 2002, MDA modified the prime contract to reflect an increased scope of work for developing GMD. It was not until July 2003 that the agency completed a review to ensure that the data was fully reliable.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-03-600, Missile Defense: Additional Knowledge Needed in Developing System for Intercepting Long-Range Missiles This is the accessible text file for GAO report number GAO-03-600 entitled 'Missile Defense: Additional Knowledge Needed in Developing System for Intercepting Long-Range Missiles' which was released on September 23, 2003. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Ranking Minority Member, Subcommittee on Financial Management, the Budget, and International Security, Committee on Governmental Affairs, U.S. Senate: United States General Accounting Office: GAO: August 2003: Missile Defense: Additional Knowledge Needed in Developing System for Intercepting Long-Range Missiles: GAO-03-600: GAO Highlights: Highlights of GAO-03-600, a report to the Ranking Minority Member, Subcommittee on Financial Management, the Budget, and International Security, Committee on Governmental Affairs, U.S. Senate Why GAO Did This Study: A number of countries hostile to the United States and its allies have or will soon have missiles capable of delivering nuclear, biological, or chemical weapons. To counter this threat, the Department of Defense‘s (DOD‘s) Missile Defense Agency (MDA) is developing a system to defeat ballistic missiles. MDA expects to spend $50 billion over the next 5 years to develop and field this system. A significant portion of these funds will be invested in the Ground-based Midcourse Defense (GMD) element. To field elements as soon as practicable, MDA has adopted an acquisition strategy whereby capabilities are upgraded as new technologies become available and is implementing it in 2-year blocks. Given the risks inherent to this strategy, GAO was asked to determine when MDA plans to demonstrate the maturity of technologies critical to the performance of GMD‘s Block 2004 capability and to identify the estimated costs to develop and field the GMD element and any significant risks with the estimate. What GAO Found: GMD is a sophisticated weapon system being developed to protect the United States against limited attacks by long-range ballistic missiles. It consists of a collection of radars and a weapon component” a three-stage booster and exoatmospheric kill vehicle”integrated by a centralized control system that formulates battle plans and directs the operation of GMD components. Successful performance of these components is dependent on 10 critical technologies. MDA expects to demonstrate the maturity of most of these technologies before fielding the GMD element, which is scheduled to begin in September 2004. However, the agency has accepted higher cost and schedule risks by beginning integration of the element‘s components before these technologies have matured. So far, MDA has matured two critical GMD technologies. If development and testing progress as planned, MDA expects to demonstrate the maturity of five other technologies by the second quarter of fiscal year 2004. The radar technologies are the least mature. MDA intends to demonstrate the maturity of an upgraded early warning radar in California in the first quarter of fiscal year 2005 and a sea-based radar in the Pacific Ocean in the fourth quarter of that year. Although MDA does not plan to demonstrate the maturity of the technology of the early warning radar in Alaska, which will serve as the primary fire control radar, through its own integrated flight tests, it may be able to do so through the anticipated launch of foreign test missiles. MDA estimates that it will spend about $21.8 billion between 1997 and 2009 to develop the GMD element. This estimate includes $7.8 billion to develop and field the GMD Block 2004 capability. For example, the funds will be used to install interceptors at two sites, upgrade existing radars and testing infrastructure, and develop the sea-based X-band radar. We found that MDA has incurred a greater risk of cost growth because for more than a year the agency was not able to rely fully on data from its primary tool for monitoring whether the GMD contractor has been performing work within cost and on schedule. In February 2002, MDA modified the prime contract to reflect an increased scope of work for developing GMD. It was not until July 2003 that the agency completed a review to ensure that the data was fully reliable. What GAO Recommends: GAO is recommending DOD (1) explore options to demonstrate effectiveness of the Cobra Dane radar and (2) establish procedures to help ensure data are reliable from MDA‘s monitoring system. DOD concurred with GAO‘s first recommendation and partially concurred with GAO‘s second. www.gao.gov/cgi-bin/getrpt?GAO-03-600. To view the full product, including the scope and methodology, click on the link above. For more information, contact Robert E. Levin at (202) 512-4841 or levinr@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: MDA Expects to Demonstrate the Maturity of Most GMD Technologies before September 2004: MDA Has Risked Cost Growth Because It Could Not Fully Rely on Data from Its System for Monitoring Contractor Performance: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Scope and Methodology: Appendix II: Comments from the Department of Defense: Appendix III: Technology Readiness Level Assessment Matrix: Appendix IV: Importance of Earned Value Management: Appendix V: GAO Contact and Staff Acknowledgments: Tables: Table 1: Technology Readiness Levels of GMD Critical Technologies: Table 2: Estimated Cost to Develop and Field GMD: Table 3: 32 Criteria for Earned Value Management Systems: Figures: Figure 1: Components of GMD: Figure 2: Notional GMD Concept of Operations: Figure 3: Tasks GMD Plans to Accomplish for the GMD Block 2004 Project: Abbreviations: BMDO: Ballistic Missile Defense Organization: CPR: Cost Performance Report: DCMA: Defense Contract Management Agency: EVM: Earned Value Management: GMD: Ground-based Midcourse Defense: IBR: integrated baseline review: IFT: integrated flight test: MDA: Missile Defense Agency: NMD: National Missile Defense: TRL: technology readiness level: United States General Accounting Office: Washington, DC 20548: August 21, 2003: The Honorable Daniel K. Akaka Ranking Minority Member Subcommittee on Financial Management, the Budget, and International Security Committee on Governmental Affairs United States Senate: Dear Senator Akaka: Hostile states, including those that sponsor terrorism, are investing significant resources to develop and deploy ballistic missiles of increasing range and sophistication that could be used against the United States, our deployed forces, and our allies. At least 25 countries now have, or are in the process of acquiring, missiles capable of delivering nuclear, biological, or chemical weapons. To counter this threat, the President of the United States in December 2002, directed the Department of Defense (DOD) to begin fielding a ballistic missile defense system in 2004. The Missile Defense Agency (MDA) within DOD is responsible for developing this system, including the Ground-based Midcourse Defense (GMD) element,[Footnote 1] which is being developed to protect the United States against long-range ballistic missiles. MDA is also building an integrated testing infrastructure--or "test bed"--with the newly designated GMD element as its centerpiece. MDA expects to spend nearly $50 billion in research and development funds between fiscal years 2004 and 2009 to develop and field a ballistic missile defense system. A significant percentage of the $50 billion will be invested in the GMD element. GMD is a sophisticated weapon system that will rely on state-of-the-art technologies that have been under development for a number of years. GMD will use space-based sensors to provide early warning of missile launches; ground-based radars to identify and refine the tracks of threatening warheads and associated objects; ground-based interceptors (each consisting of a three-stage booster and exoatmospheric kill vehicle) to destroy warheads; and a centralized control system that formulates battle plans and directs the operation of GMD components for carrying out the missile defense mission. To meet the technical challenge of developing both the integrated system and the GMD element, MDA has adopted a "capabilities-based" acquisition strategy and is implementing it in 2-year development blocks. This approach is designed to field elements as soon as practicable and to improve the effectiveness of fielded elements by upgrading their capability as new technologies become available or as the threat warrants. Block 2004 will be the first block fielded, followed by Blocks 2006 and 2008. Although GMD's Block 2004 capability is expected to be fielded beginning in September 2004, MDA plans to upgrade that capability through the end of 2005.[Footnote 2] Because development and fielding of GMD involves substantial technical challenges and a major investment, you asked us to review technical and cost issues related to the GMD element. Specifically, we determined when MDA plans to demonstrate the maturity[Footnote 3] of technologies critical to the performance of GMD's Block 2004 capability. We also identified the estimated costs to develop and field the GMD element and any significant risks associated with the estimate. Our scope and methodology are included in appendix I. Although we assessed the maturity of specific GMD critical technologies, the scope of this review did not include an evaluation of MDA's test plans for demonstrating GMD's ability to operate as a system overall. Our detailed assessment of GMD system-level testing is included in a classified report that we issued in June 2003 to other congressional requesters. Results in Brief: MDA expects to demonstrate the maturity of most of the ten technologies critical to GMD's initial performance before fielding of the element begins in September 2004. However, the agency has accepted a higher risk of cost growth and schedule slips by beginning the integration of the element's components before these technologies have been demonstrated. So far, MDA has matured two critical GMD technologies-- the infrared sensors of the kill vehicle[Footnote 4] and the fire control software of the battle management component.[Footnote 5] But if development and testing progress as planned, MDA expects to demonstrate the maturity of five others--resident in the kill vehicle, interceptor boosters, and the battle management component--by the second quarter of fiscal year 2004. MDA intends to demonstrate the maturity of an upgraded early warning radar--located at Beale Air Force Base, California--in the first quarter of fiscal year 2005 and a sea-based X- band radar, located in the Pacific Ocean, in the fourth quarter of that year. MDA does not plan to demonstrate through its own integrated flight tests the maturity of a technology resident in the Cobra Dane radar located in Alaska, which will serve as the element's primary radar when GMD is first fielded. Agency officials told us that they may be able to test the radar through the anticipated launch of foreign test missiles. However, it is not clear that testing Cobra Dane in this manner will provide all of the information that a dedicated test provides because MDA will not control the configuration of the target or the flight environment. MDA estimates that it will spend about $21.8 billion between 1997 and 2009 to develop the GMD element. This estimate includes $7.8 billion to develop and field the GMD Block 2004 capability and to develop the GMD portion of the test bed between 2002 and 2005. For example, the funds will be used to install interceptors at Fort Greely, Alaska, and Vandenberg Air Force Base, California; upgrade existing radars and the test bed infrastructure; and develop the sea-based X-band radar. MDA has incurred a greater risk of cost growth because for more than a year the agency was not able to rely fully on the data from its primary tool for monitoring whether the GMD contractor was performing work within cost and on schedule--the prime contractor's Earned Value Management (EVM) system.[Footnote 6] In February 2002, MDA modified GMD's contract to bring it into line with the agency's new capabilities-based acquisition strategy. It took several months to establish an interim cost baseline[Footnote 7] against which to measure the contractor's performance and 13 months to complete revisions to the baseline. Also, MDA and the contractor did not complete a review until July 2003 to ensure that the revised baseline was accurate and that contractor personnel were correctly using it to measure performance. This review was of particular importance because an earlier review revealed significant deficiencies in the contractor's development and use of the initial contract baseline. Until this review was completed, MDA did not know for sure whether it could rely fully on the data from its EVM system to recognize and correct potential problems in time to prevent significant cost increases and schedule delays. We are making recommendations that MDA (1) consider adding a test of the effectiveness of the radar in Alaska; and (2) ensure that procedures are in place that will increase MDA's confidence in data from its EVM system. DOD concurred with our first recommendation and partially concurred with the second. In commenting on the draft report, DOD stated that the feasibility of these procedures will be determined and that a portion of the work is already being accomplished. Background: The concept of using a missile to destroy another missile (hit-to-kill) has been explored since the mid-1950's, but it was not until 1984 that the first such intercept achieved its objective. Between the mid-1980's and late-1990's the United States conducted a number of experiments designed to demonstrate that it was possible to hit one missile with another. In 1997, the Ballistic Missile Defense Organization (BMDO) established the National Missile Defense (NMD) Joint Program Office. The program office was directed to demonstrate by 1999 a system that could protect the United States from attacks of intercontinental ballistic missiles and to be in a position to deploy the system if the threat warranted by 2003. The initial system consisted of space-and ground-based sensors, early warning radars, interceptors, and battle management functions. The program underwent additional changes as the new decade began. In September 2000, the President decided to defer deployment of the NMD system, but development of the system continued with the goal of being ready to deploy the system when directed. This action was followed in 2001 by BMDO's redirection of the prime contractor's efforts from developing and deploying an NMD system to developing an integrated test bed with the newly designated GMD system as its centerpiece. The Secretary of Defense, in January 2002, renamed BMDO as MDA and consolidated all ballistic missile defense programs under the new agency. Former missile defense acquisition programs became elements of a single ballistic missile defense system. These changes were followed in December 2002, by the President's directive to begin fielding in 2004 a ballistic missile defense system, which included components of the GMD element already under development. The GMD element is intended to protect the United States against long- range ballistic missiles in the midcourse phase of their flight. This is the point outside the atmosphere where the motors that boost an enemy missile into space have stopped burning and the deployed warhead follows a predictable path toward its target. Compared to the boost and terminal phases, this stage of flight offers the largest window of opportunity for interception and allows the GMD element a longer time to track and engage a target. As illustrated in figure 1, GMD will rely on a broad array of components to track and intercept missiles. Figure 2 provides a notional concept of how these components will operate once they are fully integrated into the GMD element. Figure 1: Components of GMD: [See PDF for image] [End of figure] Figure 2: Notional GMD Concept of Operations: [See PDF for image] Note: The concept of operations assumes weapons release authority has been previously granted by the President of the United States or the Secretary of Defense. Missile flight times may be too brief to ask for permission to launch interceptors and engage the enemy. [End of figure] MDA Expects to Demonstrate the Maturity of Most GMD Technologies before September 2004: MDA is gaining the knowledge it needs to have confidence that technologies critical to the GMD Block 2004 capability will work as intended. Two of the ten technologies essential to the Block 2004 capability have already been incorporated into actual prototype hardware and have been demonstrated to function as expected in an operational environment.[Footnote 8] Other technologies are reaching this level of maturity. If development and testing proceed as planned, MDA will demonstrate the maturity of five additional technologies by the second quarter of fiscal year 2004 and two critical radar technologies during fiscal year 2005. MDA believes that its best opportunity to demonstrate the maturity of the tenth technology, technology critical to GMD's primary radar, may come through the anticipated flight tests of foreign missiles. Our work over the years has found that making a decision to begin system integration of a capability before the maturity of all critical technologies have been demonstrated increases the program's cost, schedule, and performance risks. Because the President directed DOD to begin fielding a ballistic missile defense system in 2004, MDA began GMD system integration with technologies whose maturity has not been demonstrated. As a result, there is a greater likelihood that critical technologies will not work as intended in planned flight tests. If this occurs, MDA may have to spend additional funds in an attempt to identify and correct problems by September 2004 or accept a less capable system.[Footnote 9] Importance of Maturing Technology: Successful developers follow "knowledge-based acquisition" practices to get quality products to the customer as quickly and cost effectively as possible. As a part of meeting this goal, developers focus their technology programs on maturing technologies that have the realistic potential for being incorporated into the product under consideration. Accordingly, successful developers spend time to mature technology in a technology setting, where costs are typically not as great, and they do not move forward with product development--the initiation of a program to fully design, integrate, and demonstrate a product for production-- until essential technologies are sufficiently mature. An analytical tool--which has been used by DOD and the National Aeronautics and Space Administration, called technology readiness levels (TRLs),[Footnote 10] --can assess the maturity level of technology as well as the risk that technology poses if it is included in a product's development. The nine readiness levels are associated with progressing levels of technological maturity and demonstrated performance relative to a particular application--starting with paper studies of applied scientific principles (TRL 1) and ending with a technology that has been "flight proven" on an actual system through successful mission operations (TRL 9). Additional details on TRLs are shown in appendix III. TRLs provide a gauge of how much knowledge the program office has on the progress or status of a particular technology and are based on two principal factors: (1) the fidelity of demonstration hardware, including design maturity and level of functionality achieved; and (2) the extent and realism of the environment in which the technology has been demonstrated. MDA recognizes the value of beginning system integration with mature technology and of using TRLs to assess the maturity of technology proposed for a block configuration. In particular, MDA prefers to include new technology in a block configuration only if the technology has reached a TRL 7; that is, only if prototype hardware with the desired form, fit, and function has been proved in an operational environment. However; MDA retains the flexibility to include less mature technology in a block configuration if that technology offers a significant benefit in performance and the risk of retaining it is acceptable and properly managed. Readiness Levels of GMD Element Technologies: Through technical discussions with the GMD joint program office and its prime contractor, we identified ten critical GMD technologies and jointly assessed the readiness level of each. The critical technologies are resident in the exoatmospheric kill vehicle, the boosters, the battle management, command, and control component, and in the element's radars. In 7 of 10 cases, we agreed with the program office and the GMD prime contractor on the maturity level of the element's critical technologies. The differences in the remaining three cases, as discussed in detail below, were primarily due to interpretation of TRL definitions. The program office and its contractor rated the two booster technologies and one radar technology at higher readiness levels than, in our opinion, MDA had demonstrated. Most critical GMD technologies are currently at TRLs 5 and 6. At TRL 5, the technology's development is nearing completion, but it has not been applied or fitted for the intended product. At this point, the technology has been incorporated into a high-fidelity breadboard[Footnote 11] that has been tested in a laboratory or relevant environment[Footnote 12]. Although this demonstrates the functionality of the technology to some extent, the hardware is not necessarily of the form and fit (configuration) that would be integrated into the final product. A new application of existing technology is usually assessed at a TRL 5, because the technology has not been demonstrated in the relevant environment for the new application. TRL 6 begins the true "fitting" or application of the technology to the intended product. To reach this level, technology must be a part of a representative prototype that is very close to the form, fit, and function of that needed for the intended product. Reaching a TRL 6 requires a major step in a technology's demonstrated readiness, that is, the prototype must be tested in a high-fidelity laboratory environment or demonstrated in a restricted but relevant environment. Two of the ten GMD technologies were assessed at a TRL 7, the level that successful developers insist upon before initiating product development. To reach this level, a pre-production prototype of the technology must be demonstrated to its expected functionality in an operational environment. If development and testing proceed as planned by MDA, we judge that most of the technologies (7 of 10) will be at a TRL 7 after the completion of integrated flight test (IFT)-14,[Footnote 13] which is scheduled for the second quarter of fiscal year 2004. Table 1 summarizes our assessment of the TRL for each critical technology as of June 2003 and the date at which MDA anticipates each technology will reach TRL 7. A detailed discussion of each critical technology follows. Table 1: Technology Readiness Levels of GMD Critical Technologies: [See PDF for image] Source: GAO analysis of GMD data. Note: Information provided in the table--the configuration of flight test events and associated date--is as of June 2003 and is subject to change. [A] Assumes technology development and demonstrations will have been successful. [End of table] Exoatmospheric Kill Vehicle Technologies: The exoatmospheric kill vehicle is the weapon component of the GMD interceptor that attempts to detect and destroy the threat reentry vehicle through a hit-to-kill impact. The prime contractor identified three critical technologies pertaining to the operation of the exoatmospheric kill vehicle. They include the following: * Infrared seeker, which is the "eyes" of the kill vehicle. The seeker is designed to support kill vehicle functions like tracking and target discrimination. The primary subcomponents of the seeker are the infrared sensors, a telescope, and the cryostat that cools down the sensors. * On-board discrimination, which is needed to identify the true warhead from among decoys and associated objects. Discrimination is a critical function of the hit-to-kill mission that requires the successful execution of a sequence of functions, including target detection, target tracking, and the estimation of object features. As such, successful operation of the infrared seeker is a prerequisite for discrimination. * Guidance, navigation, and control subsystem, which is a combination of hardware and software that enables the kill vehicle to track its position and velocity in space and to physically steer itself into the designated target. All three kill vehicle technologies have been demonstrated to some extent in actual integrated flight tests on near-production- representative kill vehicles. The infrared seeker has reached a TRL 7, because a configuration very much like that to be fielded has been demonstrated in previous integrated flight tests, and only minor design upgrades are planned to reach the Block 2004 configuration. The remaining two kill vehicle technologies are at a TRL 6, because their functionality is being upgraded and the technologies have yet to be incorporated into the kill vehicle and demonstrated in an operational environment. The on-board discrimination technology has not yet reached TRL 7 because MDA has not tested a "knowledge database" that is expected to increase the kill vehicle's discrimination capability. The purpose of the database is to enable the kill vehicle to distinguish characteristics of threatening from non threatening objects. MDA expects to test the database for the first time in IFT-14. As a software-intensive technology, on-board discrimination performance under all flight conditions can only be evaluated through ground testing, but flight-testing is needed to validate the software's operation in a real world environment. The discrimination capability that will be tested in IFT-14 is expected to be fielded as part of the Block 2004 capability. Therefore, IFT-14 should demonstrate the technology's maturity if the test shows that the kill vehicle achieves its discrimination objective.[Footnote 14] Similarly, the guidance, navigation, and control technology will also increase to a TRL 7 if the technology achieves its objectives in IFT- 14. The inertial measurement unit, an important component of the guidance, navigation, and control subsystem that enables the kill vehicle to track its position and velocity, has not yet been tested in the severe environments (e.g., vibrations and accelerations) induced by the operational booster. This will be first attempted when one of the new operational boosters is used in IFT-14. In addition to testing the inertial measurement unit, IFT-14 will also test the upgraded divert hardware (used to actively steer the kill vehicle to its target) that is expected to be part of the Block 2004 configuration. Booster Technologies: The integrated booster stack is the part of the GMD interceptor that is composed of rocket motors needed to deliver and deploy the kill vehicle into a desired intercept trajectory. For all flight tests to date, a two-stage surrogate booster called the payload launch vehicle has been used. In July 1998, the GMD prime contractor began developing a new three-stage booster for the GMD program, known as the "Boost Vehicle", from commercial off-the-shelf components. However, the contractor encountered difficulty. By the time the booster was flight tested in August 2001, it was already about 18 months behind schedule. The first booster flight test met its objectives, but the second booster tested drifted off course and had to be destroyed 30 seconds after launch. Subsequently, MDA altered its strategy for acquiring a new booster for the interceptor. Instead of relying on a single contractor, MDA authorized the GMD prime contractor to develop a second source for the booster by awarding a subcontract to another contractor. If development of the boosters proceeds as planned, both boosters will be part of the Block 2004 capability. One booster is known as BV+ and the other as "OSC Lite.": The BV+ Booster: The prime contractor ultimately transferred development of the boost vehicle to a subcontractor who is currently developing a variant--known as "BV+"--for the GMD element. The program office and GMD contractor rated the BV+ at a TRL 7. The prime contractor reasoned that the extent of the legacy program and its one successful flight test should allow for this rating. However, given the limited testing to date, we assessed the BV+ booster currently at a TRL 6; that is, the technology has been demonstrated in a restricted flight environment using hardware close in form, fit, and function to that which will be fielded in 2004. We believe the contractor's assessment is too high at this time, because the step from TRL 6 to TRL 7 is significant in terms of the fidelity of the demonstration environment. However, the first test of a full configuration BV+ booster will occur with IFT-13A, which is scheduled for the first quarter of fiscal year 2004. In our opinion, the BV+ booster will reach TRL 7 at this time if the booster works as planned. The "OSC Lite" Booster: The second booster under development is referred to as "OSC Lite". This booster, which is essentially the Taurus Lite missile that carries satellites into low-earth orbit, will be reconfigured for the GMD element. Despite the fact that the booster was recently tested under restricted flight conditions, GMD's prime contractor believes that the legacy development of the Taurus Lite missile is sufficient to prove that the OSC Lite has reached TRL 7. However, in our opinion, because the test was conducted with hardware configured as it was in the Taurus missile, not as it will be configured for GMD's Block 2004, the booster's maturity level is comparable to that of the BV+. The first flight test of a full configuration OSC Lite booster is scheduled for IFT-13B in the first quarter of fiscal year 2004. We believe that if the booster performs as intended in this test, it will reach TRL 7. Battle Management Command, Control, and Communications Technologies: The battle management component is the integrating and controlling component of the GMD element. Prime contractor officials identified and assessed the following sub-components as critical technologies: * GMD fire control software, which analyzes the threat, plans engagements, and tasks components of the GMD element to execute a mission. * In-flight interceptor communications system, which enables the GMD fire control component to communicate with the exoatmospheric kill vehicle while in flight. The two battle management technologies have been demonstrated to some extent in actual integrated flight tests, and both are near their Block 2004 design. We determined that the GMD fire control software has currently achieved a TRL 7 and the in-flight interceptor communications system has reached a TRL 6. Prime contractor officials concur with our assessment. The fire control software is nearing expected functionality and prior software builds have been demonstrated in GMD flight tests. Only minor design changes will be made to address interfacing issues (linking the fire control component with other GMD components) before the software reaches the operational configuration of Block 2004. As a software-intensive technology, the performance of the fire control software throughout the entire "flight envelope" can only be evaluated through ground testing. Ground testing is well underway at both the Joint National Integration Center at Schriever Air Force Base, Colorado, and at the prime contractor's integration laboratory in Huntsville, Alabama. The second technology associated with the battle management component is the in-flight interceptor communications system. Even though the pointing accuracy and communications capability of this technology were demonstrated in previous flight tests, the operational hardware to be fielded by 2004 is expected to operate at a different uplink frequency than the legacy hardware used in these past flight tests.[Footnote 15] Accordingly, we assessed the in-flight interceptor communications system at a TRL 6. The first integrated flight test to include an operational-like build of this technology is IFT-14, and if the technology meets its objectives in this flight test, TRL 7 would be achieved. Radar Technologies: The GMD contractor initially identified the sea-based X-band radar as the only radar-related critical technology. Since its initial assessment in September 2002, the contractor has now agreed with us that the Beale upgraded early warning radar and the Cobra Dane radar are also critical technologies of the GMD element. The contractor and the GMD program office assessed the Beale and Cobra Dane radars at a TRL 5, because the technology, especially mission software, is still under development and has not yet been demonstrated in a relevant flight environment.[Footnote 16] The contractor assessed the sea-based X-band radar at a TRL 6. As discussed below, we agree with their assessment of the Beale and Cobra Dane radars but rated the sea-based X-band radar as a TRL 5. The early warning radar at Beale Air Force Base has participated in integrated flight tests in a missile-defense role using legacy hardware and developmental software. Design and development of operational builds of the software are progressing, but such builds have only been tested in a simulated environment. Therefore, we assessed the Beale radar technology at a TRL 5--an assessment driven by software considerations. The conversion of the early warning radar at Beale to an upgraded early warning radar, which consists of minor hardware and significant software upgrades, is planned for completion sometime during the middle of fiscal year 2004. After this time, the Beale radar can take part in flight-testing in its upgraded configuration. MDA currently plans to demonstrate the upgraded Beale technology in a non intercept flight test, known as a radar certification flight,[Footnote 17] in the first quarter of fiscal year 2005. The Beale radar will be demonstrated at a TRL 7 if the objectives of this flight test are achieved. The Cobra Dane radar is currently being used in a surveillance mode to collect data on selected intercontinental ballistic missile test launches out of Russia and does not require real-time data processing and communications capabilities. To achieve a defensive capability by September 2004, the Cobra Dane radar is being upgraded to perform both of these tasks. This upgrade, which requires a number of software modifications, is designed to enable Cobra Dane to detect and track enemy targets much as the Beale upgraded early warning radar does. Although the hardware component of the Cobra Dane radar is mature and will undergo only minor updating, Cobra Dane's mission software is being revised for this application. The revision includes reuse of existing software and development of new software so that the Cobra Dane radar can be integrated into the GMD architecture. Upgrades to the Cobra Dane radar are due to be completed at the beginning of 2004. After the software is developed and ground tested, the radar can reach a TRL 6, but it is uncertain when the radar will reach a TRL 7. Because of other funding and scheduling priorities, MDA has no plans through fiscal year 2007 for using this radar in integrated flight tests; such tests would require air-or sea-launched targets that are not currently part of the test program. Unless the current test program is modified, the only opportunities for demonstrating Cobra Dane in an operational environment would come from flight tests of foreign missiles. MDA officials anticipate that such opportunities will occur. However, it is not clear that testing Cobra Dane in this manner will provide all of the information that a dedicated test provides because MDA will not control the configuration of the target or the flight environment. The sea-based X-band radar is being built as part of the Block 2004 capability and scheduled for completion in 2005. It will be built from demonstrated technologies--a sea-based platform and the prototype X-band radar currently being used in the GMD test program. Prime contractor officials told us that they consider the risk associated with the construction and checkout of the radar as primarily a programmatic, rather than technical risk, and believe that the sea- based X-band radar has reached a TRL 6. The contractor also stated that the initial operational build of the radar software is developed and currently being tested at the contractor's integration laboratory. We assessed the sea-based X-band radar as a TRL 5 because the radar has not yet been built and because constructing a radar from an existing design and placing it on a sea-based platform is a new application of existing technology. For example, severe wind and sea conditions may affect the radar's functionality--conditions that cannot be replicated in a laboratory. As a result, developers cannot be sure that the sea- based X-band radar will work as intended until it is demonstrated in this new environment. However, both we and the contractor agree that the maturity level of the sea-based X-band radar will increase to a TRL 7 if it achieves its test objectives in IFT-18 (scheduled for the fourth quarter of fiscal year 2005). MDA Has Risked Cost Growth Because It Could Not Fully Rely on Data from Its System for Monitoring Contractor Performance: From the program's inception in 1997[Footnote 18] through 2009, MDA expects to spend about $21.8 billion to develop the GMD element. About $7.8 billion of the estimated cost will be needed between 2002 and 2005 to develop and field the Block 2004 GMD capability and to develop the GMD portion of the test bed.[Footnote 19] However, MDA has incurred a greater risk of cost increases because for more than a year MDA was not sure that it could rely fully upon data from the prime contractor's Earned Value Management (EVM) system,[Footnote 20] which provides program managers and others with early warning of problems that could cause cost and schedule growth. GMD Development Costs: Before the restructuring of the GMD program in 2002, about $6.2 billion was spent (between 1997 and 2001) to develop a ground-based defense capability. MDA estimates it will need an additional $7.8 billion between 2002 and 2005 to, among other tasks, install interceptors at Fort Greely, Alaska, and at Vandenberg Air Force Base, California; upgrade existing radars and test bed infrastructure; and develop the sea-based X-band radar that will be added in the fourth quarter of fiscal year 2005. In addition, MDA will invest an additional $7.8 billion between fiscal year 2004 and 2009 to continue efforts begun under Block 2004, such as enhancing capability and expanding the test bed. Table 2, below, provides details on the funding requirements by block and by fiscal year, and figure 3 provides examples of specific Block 2004 tasks. Table 2: Estimated Cost to Develop and Field GMD: [See PDF for image] Source: Ballistic Missile Defense Budget, Midcourse Defense Segment, February 2003. [End of table] Figure 3: Tasks GMD Plans to Accomplish for the GMD Block 2004 Project: [See PDF for image] [End of figure] MDA did not include the following costs is its Block 2004 estimate: * The cost to recruit, hire, and train military personnel to operate the initial defensive capability and provide site security at various locations, which MDA estimates to be an additional $13.4 million (half in fiscal year 2003 and half in 2004 each), will be needed to operate GMD and provide physical security. Additional costs to cover these personnel throughout the life of the program beginning in 2005 and beyond were also omitted. * The cost to maintain equipment and facilities was not included. * Systems engineering and national team costs--which benefit all elements, including GMD and cannot be divided among the elements--were not included in MDA's budget. MDA's Insight into Potential Cost Growth Was Limited by the Agency's Inability to Rely Fully on Data from Earned Value Management System: Because a significant portion of MDA's Block 2004 GMD cost estimate is the cost of work being performed by the element's prime contractor, MDA's ability to closely monitor its contractor's performance is critical to controlling costs. The tool that MDA, and many DOD entities, have chosen for this purpose is the EVM system. This system uses contractor reported data to provide program managers and others with timely information on a contractor's ability to perform work within estimated cost and schedule. It does so by examining variances reported in contractor cost performance reports between the actual cost and time of performing work tasks and the budgeted or estimated cost and time. While this tool can provide insightful information to managers, MDA's use of it has been hampered by several factors. Principally, although major contract modifications were made in February 2002, it took until July 2003 for MDA to complete a review to confirm the reliability of data from the EVM system. An earlier review of a similar nature revealed significant deficiencies in the contractor's formulation and collection of EVM data. Until a new review was completed, MDA could not be sure about its ability to rely fully upon this data to identify potential problems in time to prevent significant cost growth and schedule delays. Baseline Revised over 13-Month Period: An accurate, valid, and current performance management baseline is needed to perform useful analyses using EVM. The baseline identifies and defines work tasks, designates and assigns organizational responsibility for each task, schedules the work task in accordance with established targets, and allocates budget to the scheduled work. According to DOD guidance,[Footnote 21] a performance management baseline should be in place as early as possible after the contractor is authorized to proceed. Although the guidance does not define how quickly the contractor should establish a baseline, experts generally agree that it should be in place, on average, within 3 months after a contract is awarded or modified. About a year before the Secretary of Defense directed MDA to adopt an evolutionary acquisition strategy, the agency awarded a new contract for the development of a National Missile Defense system. In February 2002, MDA modified this contract to redirect the contractor's efforts. Instead of developing a missile defense system that met all of the requirements of the war fighter, as the initial contract required, the modification directed the contractor to develop the first GMD increment, or block, which was to be a ballistic missile test bed with GMD as its centerpiece. Following the contract's modification, the contractor in June 2002 established an interim baseline. This baseline was developed by adding budgets for near-term new work to the original baseline. Because the cost of the work being added to the baseline had not yet been negotiated, the contractor based the budgets on the cost proposed to MDA, as directed by DOD guidelines. The contractor implemented the baseline almost within the 3-month time frame recommended by experts. In the time between the modification and the development of the interim baseline, MDA authorized the contractor to begin work and spend a specified amount of money, and MDA paid the contractor about $390 million during this period. An option that MDA could have used to help validate the interim baseline was to have the Defense Contract Management Agency (DCMA)[Footnote 22] verify contractor work packages and track the movement of funds between the unpriced work account and the baseline. However, neither MDA nor DCMA initiated these actions. In its technical comments on a draft of this report, DOD pointed out that during the negotiation process, MDA reviews prime and subcontractor proposal data that include engineering labor hours, material, and cost estimates. DOD further noted that these estimates eventually form a basis for the work packages that make up the data for the performance management baseline. We agree that these costs will eventually be associated with the work packages that make up the baseline. However, a joint contractor and MDA review of the initial GMD baseline concluded that even though these costs were otherwise fair and reasonable, some work packages that the contractor developed for the original contract's baseline did not correctly reflect the work directed by MDA. An independent review of work packages included in the interim baseline would have increased the likelihood that the work packages were being properly developed and that their budget and schedule were appropriate. The contractor completed all revisions to the baseline for the prime contractor and all five subcontractors by March 2003, 3 months after negotiating the cost of the modification and 13 months after authorizing the work to begin. The contracting officer explained that it took until December 2002 to negotiate the 2002 contract change because the additional work was extremely complex, and, as a result, the modification needed to be vetted through many subcontractors that support the prime. Baseline Review Completed in July 2003: The DOD guidance states that an integrated baseline review (IBR) is to be conducted within 6 months of award of a new contract or major change to an existing contract.[Footnote 23] The review verifies the technical content of the baseline. It also ensures that contractor personnel understand and have been adequately trained to collect EVM data. The review also verifies the accuracy of the related budget and schedules, ensures that risks have been properly identified, assesses the contractor's ability to implement properly EVM, and determines if the work identified by the contractor meets the program's objectives. The government's program manager and technical staff carry out this review with their contractor counterparts. Completing an IBR of the new baseline has been of particular importance because the July 2001 IBR for the initial contract identified more than 300 deficiencies in the contractor's formulation and execution of the baseline. For example, the contractor had not defined a critical path for the overall effort, many tasks did not have sufficient milestones that would allow the contractor to objectively measure performance, and contractor personnel who were responsible for reporting earned value were making mistakes in measuring actual performance against the baseline. MDA began a review in March 2003 of the contractor's new baseline, which reflected the contract modification,. Completing this IBR took until July 2003 because of the complexity of the program and the many subcontractors that were involved. Although the review team found fewer problems with the contractor's formulation and execution of the new baseline, problems were identified. For example, the IBR showed that in some cases the baseline did not reflect the new statement of work. Also, both the prime contractor and subcontractors improperly allocated budget to activities that indirectly affect a work product (known as level of effort activities) when they could have associated these activities with a discrete end product. Because of the way these activities are accounted for, this designation could mask true cost variances. Management Reserve Used to Offset Expected Cost Overruns at Contract Completion: Before the IBR was underway, DCMA recognized another problem with the contractor's EVM reports. In its December 2002 cost performance report, the contractor reported that it expected no cost overrun at contract completion. This implied that the program was not experiencing any problems that could result in significant cost or schedule growth. However, DCMA stated that October 2002 was the second month in a row that the contractor had used management reserve funds to offset a significant negative cost variance.[Footnote 24] DCMA emphasized that this is not the intended purpose of management reserves. (Management reserves are a part of the total project budget intended to be used to fund work anticipated but not currently defined.) DCMA officials told us that while this is not a prohibited practice most programs wait until their work is almost completed, that is 80 to 90 percent complete, before making a judgment that the management reserve would not be needed for additional undefined work and could be applied to unfavorable contract cost variances. Conclusions: Because of the President's direction to begin fielding a ballistic missile defense system in 2004, the MDA took a higher risk approach by beginning GMD system integration before knowing whether its critical technologies were mature. If development and testing progress as planned, however, MDA expects to have demonstrated the maturity of 7 of the 10 critical GMD technologies before the element is initially fielded in September 2004 and 2 others during fiscal year 2005. If technologies do not achieve their objectives during testing, MDA may have to spend additional funds in an attempt to identify and correct problems by September 2004 or accept a less capable system. Because of other funding and scheduling priorities, MDA does not plan to demonstrate through integrated flight tests whether the Cobra Dane radar's software can process and communicate data on the location of enemy missiles in "real time." Although tests using sea-or air-launched targets before September 2004 would provide otherwise unavailable information on the software's performance, we recognize those tests would be costly and funds have not been allocated for that purpose. We also recognize that the most cost efficient means of testing the Cobra Dane radar is through launches involving foreign test missiles. However, we believe it would be useful for MDA to consider whether the increased confidence provided by a planned test event outweighs other uses for those funds. MDA is investing a significant amount of money to achieve an operational capability during the first block of GMD's development, and the agency expects to continue investing in the element's improvement over the next several years. Because MDA is also developing other elements and must balance its investment in each, it needs an accurate GMD cost estimate. If it is used as intended, the EVM system can be an effective means of monitoring one of GMD's largest costs, the cost of having a contractor develop the GMD system. It is understandable that the dynamic changes in MDA's acquisition strategy led to major contract modifications, which made it more difficult for the contractor to establish a stable baseline. However, in this environment, it is even more important that MDA find ways to ensure the integrity of the interim baselines and to quickly determine that revised baselines can be fully relied on to identify potential problems before they significantly affect the program's cost. Recommendations for Executive Action: To increase its confidence that the Ground-based Midcourse Defense element fielded in 2004 will operate as intended, we recommend that the Secretary of Defense direct the Director, Missile Defense Agency, to explore its options for demonstrating the upgraded Cobra Dane radar in its new ballistic missile defense role in a real-world environment before September 2004. To improve MDA's oversight of the GMD element and to provide the Congress with the best available information for overseeing the program, we recommend that the Secretary of Defense direct the Director, Missile Defense Agency, to: * ensure that when a contractor is authorized to begin new work before a price is negotiated that DCMA validate the performance measurement baseline to the extent possible by (1) tracking the movement of budget from the authorized, unpriced work account into the baseline, (2) verify that the work packages accurately reflect the new work directed, and (3) report the results of this effort to MDA; and: * strive to initiate and complete an integrated baseline review (IBR) of any major contract modifications within 6 months. Agency Comments and Our Evaluation: DOD's comments on our draft report are reprinted in appendix II. DOD concurred with our first recommendation. DOD stated that MDA is exploring its options for demonstrating, prior to 2004, the upgraded Cobra Dane radar in a real-world environment. However, DOD noted that because it takes considerable time to develop and produce targets and to conduct safety and environmental assessments, completing a Cobra Dane radar test before September 2004 would be very challenging. DOD concluded that "targets of opportunity" (flight tests of foreign missiles) and ground testing may provide the best means to demonstrate the radar's maturity in the near term. DOD partially concurred with our second recommendation. In responding to the first part of recommendation two, DOD stated that MDA and the DCMA will jointly determine the feasibility of tracking the budget for authorized, unpriced work into the baseline and will concurrently assess work package data while establishing the formal performance measurement baseline. DOD also stated that a selected portion of this work is already being accomplished by DCMA. We continue to believe in the feasibility of our recommendation. DCMA officials told us that they could monitor the movement of budget into the baseline and verify the work packages associated with the budget. In addition, the guidelines state that surveillance may be accomplished through sampling of internal and external data. We believe that if DCMA sampled the data as it is transferred into the baseline, the implementation of this recommendation should not be burdensome. In responding to the second part of recommendation two, DOD stated that MDA will continue to adhere to current DOD policy by starting an IBR of any major contract modification within 6 months. MDA correctly pointed out that DOD's Interim Defense Acquisition Guidebook only requires a review be initiated within 6 months (180 days) after a contract is awarded or a major modification is issued. However, DOD's Earned Value Management Implementation Guide states that such a review is conducted within 6 months. Similar language is found in the applicable clause from the GMD contract,[Footnote 25] which states that such reviews shall be scheduled as early as practicable and should be conducted within 180 calendar days after the incorporation of major modifications. While we understand the difficulty of conducting reviews within 180 days when the contract is complex and many subcontractors are involved, we believe that it is important for the government to complete an IBR as soon as possible to ensure accurate measurement of progress toward the program's cost, schedule, and performance goals. DOD also provided technical comments to this report, which we considered and implemented as appropriate. In its technical comments, for example, DOD expressed particular concern that our draft report language asserting MDA's inability to rely on the EVM system was unsupported and misleading. DOD also stated that its prime contractor's EVM system is reliable. It stated, for example, that MDA has reviewed, and continues to review on a monthly basis, the contractor's cost performance reports and that the prime contractor's EVM system and accounting systems have been fully certified and validated by DCMA. We modified our report to better recognize MDA's ability to use and trust the EVM system. However, we still believe that MDA would benefit from taking additional measures to increase its confidence in the accuracy of its interim baselines. Also, when the revised baseline is in place, a review of its formulation and execution is necessary before MDA can confidently and fully rely on data from the EVM system. We conducted our review from December 2001 through August 2003 in accordance with generally accepted government auditing standards. As arranged with your staff, unless you publicly announce its contents earlier, we plan no further distribution of this report until 30 days from its issue date. At that time, we plan to provide copies of this report to interested congressional committees, the Secretary of Defense, and the Director, Missile Defense Agency. We will make copies available to others upon request. In addition, the report will be available at no charge on the GAO Web site at http://www.gao.gov/. If you or your staff have any questions concerning this report, please contact me on (202) 512-4841. Major contributors to this report are listed in appendix V. Sincerely yours, Robert E. Levin Director Acquisition and Sourcing Management: Signed by Robert E. Levin: [End of section] Appendix I: Scope and Methodology: To determine when MDA plans to demonstrate the maturity of technologies critical to the performance of GMD's Block 2004 capability, we reviewed their critical technologies using technology readiness levels (TRLs) developed by the National Aeronautics and Space Administration and used by DOD. We did so by asking contractor officials at the Boeing System Engineering and Integration Office in Arlington, Virginia, to identify the most critical technologies and to assess the level of maturity of each technology using definitions developed by the National Aeronautics and Space Administration. We reviewed these assessments along with program documents, such as the results of recent flight tests and discussed the results with contractor and agency officials in order to reach a consensus, where appropriate, on the readiness level for each technology and identify the reasons for any disagreements. In reviewing the agency's current cost estimate to develop the first block of the GMD element and its test bed, we reviewed and analyzed budget backup documents, cost documents, and selected acquisition reports for the GMD program extending over a period of several years. We also met with program officials responsible for managing the development and fielding of the GMD Block 2004 capability. For example, we met with officials from the GMD Joint Program Office in Arlington, Virginia, and Huntsville, Alabama; and the Office of the Deputy Assistant for Program Integration at the MDA, Arlington, Virginia. To determine whether there were any significant risks associated with the estimate, we met with agency officials responsible for determining the cost of the GMD element to find out if there were costs that were omitted, but should have been included, in the estimate. We also analyzed data from cost performance reports that the GMD contractor developed for the MDA. We reviewed data from the GMD element and contracting officials and conducted interviews to discuss the data. Although we did not independently verify the accuracy of the cost performance reports we received from MDA, the data were assessed independently by DCMA. [End of section] Appendix II: Comments from the Department of Defense: OFFICE OF THE UNDER SECRETARY OF DEFENSE: 3000 DEFENSE PENTAGON WASHINGTON, DC 20301-3000: ACQUISITION, TECHNOLOGY AND LOGISTICS: 6 JUL 2003: Mr. R. E. Levin: Managing Director, Acquisition and Sourcing Management U.S. General Accounting Office: 441 G Street, NW Washington, D.C. 20548: Dear Mr. Levin: This is the Department of Defense (DoD) response to the General Accounting Office (GAO) draft report, "MISSILE DEFENSE: Additional Knowledge Needed in Developing System for Intercepting Long-Range Missiles," dated June 19, 2003 (GAO Code 120109/GAO 03-600). The Department appreciates the opportunity to comment on the draft report. The Department concurs with the first recommendation and partially concurs with the second. MDA will continue to adhere to current DoD policy by starting an integrated baseline review of any major modification within six (6) months after contract definitization. Specific comments for each recommendation are enclosed. We are also providing recommendations for factual corrections in a separate enclosure. My point of contact for this report is Lt Col Christina N. Walton, USAF, (703) 697-5385, christina.walton@osd.mil. Sincerely, Mark D. Schaeffer Principal Deputy Defense Systems: Signed by Mark D. Schaeffer: Enclosures: As Stated: GAO DRAFT REPORT - DATED JUNE 19, 2003 GAO CODE 120109/GAO-03-600: "MISSILE DEFENSE: ADDITIONAL KNOWLEDGE NEEDED IN DEVELOPING SYSTEM FOR INTERCEPTING LONG-RANGE MISSILES": DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS: RECOMENDATION 1: GAO recommends that "To increase its confidence that the Ground-based Midcourse Defense element fielded in 2004 will operate as intended, we recommend that MDA explore its options for demonstrating the upgraded Cobra Dane radar in its new ballistic missile defense role in a real-world environment before September 2004.": DOD RESPONSE: Concur. MDA is considering the addition of an integrated flight test prior to September 30, 2004, that would prove-out the upgrades that are underway to the Cobra Dane radar at Shemya, Alaska. However, the lead time for adding radar tests with dedicated targets is considerable and depends on the time needed for target development and/ or production, as well as the requisite safety and environmental assessments. As a result, accomplishing such testing before September 2004 would be very challenging. Consequently, targets-of-opportunity and ground testing may provide the best avenue for demonstrating the radar in the near term. RECOMENDATION 2: To improve MDA's oversight of the Ground-based Midcourse Defense element and to provide Congress with more reliable information for overseeing the program, we recommend that MDA: * Ensure that when a contractor is authorized to begin new work before a price is negotiated that DCMA validate the performance measurement baseline to the extent possible by (1) tracking the movement of budget from the authorized, un-priced work account into the baseline, (2) verify that the work packages accurately reflect the new work directed, and (3) report the results of this effort to MDA. * Strive to initiate and complete an integrated baseline review of any major contract modifications within 6 months. DOD RESPONSE: Partially Concur. MDA and DCMA will jointly determine the feasibility of performing reviews of the transfer of budget for authorized, un-priced work into the baseline and assess work package data concurrently with the establishment of the formal performance measurement baseline. A selected portion of this work is already being accomplished by DCMA contractor surveillance activities. During the period that the contractor is establishing the Performance Measurement Baseline (PMB), there are reviews of contractor estimating and accounting processes including earned value management. MDA will continue to adhere to current DoD policy by starting an integrated baseline review of any major modification within six (6) months after contract definitization. DCMA Boeing Anaheim Comments: DCMA early activities to validate the PMB included participation in all IBR activities (both at Prime and at Subcontractors), membership in all cost and financial Integrated Process Teams, attended program management reviews, reviewed prime and subcontractor's proposals, appointment of warranted Administrative Contract Officer, Certification of all the contractors ticketed system including; Accounting System, Anaheim Site Billing System, Anaheim Site, Estimating System, and Indirect & ODC System. [End of section] Appendix III: Technology Readiness Level Assessment Matrix: Technology readiness level (TRL): 1. Basic principles observed and reported; Description: Lowest level of technology readiness. Scientific research begins to be translated into applied research and development. Examples might include paper studies of a technology's basic properties; Hardware /software: None (paper studies and analysis); Demonstration environment: None. Technology readiness level (TRL): 2. Technology concept and/or application formulated; Description: Invention begins. Once basic principles are observed, practical applications can be invented. The application is speculative, and there is no proof or detailed analysis to support the assumption. Examples are still limited to paper studies; Hardware /software: None (paper studies and analysis); Demonstration environment: None. Technology readiness level (TRL): 3. Analytical and experimental critical function and/or characteristic proof of concept; Description: Active research and development is initiated. This includes analytical studies and laboratory studies to physically validate analytical predictions of separate elements of the technology. Examples include components that are not yet integrated or representative; Hardware / software: Analytical studies and demonstration of nonscale individual components (pieces of subsystem); Demonstration environment: Lab. Technology readiness level (TRL): 4. Component and/or breadboard. Validation in laboratory environment; Description: Basic technological components are integrated to establish that the pieces will work together. This is relatively "low fidelity" compared to the eventual system. Examples include integration of "ad hoc" hardware in a laboratory; Hardware /software: Low fidelity breadboard. Integration of nonscale components to show pieces will work together. Not fully functional or form or fit but representative of technically feasible approach suitable for flight articles; Demonstration environment: Lab. Technology readiness level (TRL): 5. Component and/or breadboard validation in relevant environment; Description: Fidelity of breadboard technology increases significantly. The basic technological components are integrated with reasonably realistic supporting elements so that the technology can be tested in a simulated environment. Examples include "high fidelity" laboratory integration of components; Hardware /software: High fidelity breadboard. Functionally equivalent but not necessarily form and/or fit (size, weight, materials, etc). Should be approaching appropriate scale. May include integration of several components with reasonably realistic support elements/ subsystems to demonstrate functionality; Demonstration environment: Lab demonstrating functionality but not form and fit. May include flight-demonstrating breadboard in surrogate aircraft. Technology ready for detailed design studies. Technology readiness level (TRL): 6. System/subsystem model or prototype demonstration in a relevant environment; Description: Representative model or prototype system, which is well beyond the breadboard tested for TRL 5, is tested in a relevant environment. Represents a major step up in a technology's demonstrated readiness. Examples include testing a prototype in a high fidelity laboratory environment or in simulated operational environment; Hardware / software: Prototype. Should be very close to form, fit, and function. Probably includes the integration of many new components and realistic supporting elements/subsystems if needed to demonstrate full functionality of the subsystem; Demonstration environment: High- fidelity lab demonstration or limited/restricted flight demonstration for a relevant environment. Integration of technology is well defined. Technology readiness level (TRL): 7. System prototype demonstration in an operational environment; Description: Prototype near or at planned operational system. Represents a major step up from TRL 6, requiring the demonstration of an actual system prototype in an operational environment, such as in an aircraft, on a vehicle or in space. Examples include testing the prototype in a test bed aircraft; Hardware / software: Prototype. Should be form, fit and function integrated with other key supporting elements/subsystems to demonstrate full functionality of subsystem; Demonstration environment: Flight demonstration in representative operational environment such as flying test bed or demonstrator aircraft. Technology is well substantiated with test data. Technology readiness level (TRL): 8. Actual system completed and "flight qualified" through test and demonstration; Description: Technology has been proven to work in its final form and under expected conditions. In almost all cases, this TRL represents the end of true system development. Examples include developmental test and evaluation of the system in its intended weapon system to determine if it meets design specifications; Hardware /software: Flight-qualified hardware; Demonstration environment: Developmental test and evaluation in the actual system application. Technology readiness level (TRL): 9. Actual system "flight proven" through successful mission operations; Description: Actual application of the technology in its final form and under mission conditions, such as those encountered in operational test and evaluation. In almost all cases, this is the end of the last "bug fixing" aspects of true system development. Examples include using the system under operational mission conditions; Hardware /software: Actual system in final form; Demonstration environment: Operational test and evaluation in operational mission conditions. Source: GAO and its analysis of National Aeronautics and Space Administration data. Note: GAO information based on U.S. General Accounting Office, Missile Defense: Knowledge-Based Decision Making Needed to Reduce Risks in Developing Airborne Laser, GAO-02-631 (Washington, D.C.: June 2002). [End of table] [End of section] Appendix IV: Importance of Earned Value Management: Pulling together essential cost, schedule, and technical information in a meaningful, coherent fashion is always a challenge for any program. Without this information, management of the program will be fragmented, presenting a distorted view of program status. For several decades, DOD has compared the value of work performed to the work's actual cost. This measurement is referred to as Earned Value Management (EVM). Earned value goes beyond the two-dimensional approach of comparing budgeted costs to actuals. It attempts to compare the value of work accomplished during a given period with the work scheduled for that period. By using the value of completed work as a basis for estimating the cost and time needed to complete the program, the earned value concept should alert program managers to potential problems early in the program. In 1996, in response to acquisition reform initiatives, DOD reemphasized the importance of earned value in program management and adopted 32 criteria for evaluating the quality of management systems. These 32 criteria are organized into 5 basic categories: organization, planning and budgeting, accounting considerations, analysis and management reports, and revisions and data maintenance. The 32 criteria are listed in table 1. In general terms, the criteria require contractors to (1) define the contractual scope of work using a work breakdown structure; (2) identify organizational responsibility for the work; (3) integrate internal management subsystems; (4) schedule and budget authorized work; (5) measure the progress of work based on objective indicators; (6) collect the cost of labor and materials associated with the work performed; (7) analyze any variances from planned cost and schedules; (8) forecast costs at contract completion; and (9) control changes. Table 3: 32 Criteria for Earned Value Management Systems: Criteria: Categories of Criteria: 1. Define the authorized work elements for the program. A work breakdown structure, tailored for effective internal management control, is commonly used in this process. Criteria: Categories of Criteria: 2. Identify the program organizational structure, including the major subcontractors responsible for accomplishing the authorized work, and define the organizational elements in which work will be planned and controlled. Criteria: Categories of Criteria: 3. Provide for the integration of the company's planning, scheduling, budgeting, work authorization, and cost accumulation processes with each other and, as appropriate, the program work breakdown structure and the program organizational structure. Criteria: Categories of Criteria: 4. Identify the company organization or function responsible for controlling overhead (indirect costs). Criteria: Categories of CriteriaPlanning and Budgeting: 5. Provide for integration of the program work breakdown structure and the program organizational structure in a manner that permits cost and schedule performance measurement by elements of either or both structures as needed. Categories of Criteria: Organization: Planning and Budgeting; Criteria: 6. Schedule the authorized work in a manner that describes the sequence of work and identifies significant task interdependencies required to meet the requirements of the program. Criteria: Categories of Criteria: 7. Identify physical products, milestones, technical performance goals, or other indicators that will be used to measure progress. Criteria: Categories of Criteria: 8. Establish and maintain a time- phased budget baseline, at the control account level, against which program performance can be measured. Budget for far-term efforts may be held in higher-level accounts until an appropriate time for allocation at the control account level. Initial budgets established for performance measurement will be based on either internal management goals or the external customer-negotiated target cost including estimates for authorized but undefinitized work. On government contracts, if an over target baseline is used for performance measurement reporting purposes, prior notification must be provided to the customer. Criteria: Categories of Criteria: 9. Establish budgets for authorized work with identification of significant cost elements (labor, material, etc.) as needed for internal management and for control of subcontractors. Criteria: Categories of Criteria: 10. To the extent it is practical to identify the authorized work in discrete work packages, establish budgets for this work in terms of dollars, hours, or other measurable units. Where the entire control account is not subdivided into work packages, identify the far term effort in larger planning packages for budget and scheduling purposes. Criteria: Categories of Criteria: 11. Provide that the sum of all work package budgets plus planning package budgets within a control account equals the control account budget. Criteria: Categories of Criteria: 12. Identify and control level of effort activity by time-phased budgets established for this purpose. Only that effort which is unmeasurable or for which measurement is impractical may be classified as level of effort. Criteria: Categories of Criteria: 13. Establish overhead budgets for each significant organizational component of the company for expenses that will become indirect costs. Reflect in the program budgets, at the appropriate level, the amounts in overhead pools that are planned to be allocated to the program as indirect costs. Criteria: Categories of Criteria: 14. Identify management reserves and undistributed budget. Criteria: Categories of CriteriaAccounting Considerations: 15. Provide that the program target cost goal is reconciled with the sum of all internal program budgets and management reserves. Categories of Criteria: Organization: Accounting Considerations; Criteria: 16. Record direct costs in a manner consistent with the budgets in a formal system controlled by the general books of account. Criteria: Categories of Criteria: 17. When a work breakdown structure is used, summarize direct costs from control accounts into the work breakdown structure without allocation of a single control account to two or more work breakdown structure elements. Criteria: Categories of Criteria: 18. Summarize direct costs from the control accounts into the contractor's organizational elements without allocation of a single control account to two or more organizational elements. Criteria: Categories of Criteria: 19. Record all indirect costs which will be allocated to the contract. Criteria: Categories of CriteriaAccounting Considerations: 20. Identify unit costs, equivalent units costs, or lot costs when needed. Categories of Criteria: Organization: Accounting Considerations; Criteria: 21. For EVMS, the material accounting system will provide for: (1) Accurate cost accumulation and assignment of costs to control accounts in a manner consistent with the budgets using recognized, acceptable, costing techniques. (2) Cost performance measurement at the point in time most suitable for the category of material involved, but no earlier than the time of progress payments or actual receipt of material. (3) Full accountability of all material purchased for the program including the residual inventory. Categories of Criteria: Organization: Analysis and Management Reports; Criteria: 22. At least on a monthly basis, generate the following information at the control account and other levels as necessary for management control using actual cost data from, or reconcilable with, the accounting system: (1) Comparison of the amount of planned budget and the amount of budget earned for work accomplished. This comparison provides the schedule variance. (2) Comparison of the amount of the budget earned and the actual (applied where appropriate) direct costs for the same work. This comparison provides the cost variance. Criteria: Categories of Criteria: 23. Identify, at least monthly, the significant differences between both planned and actual schedule performance and planned and actual cost performance, and provide the reasons for the variances in the detail needed by program management. Criteria: Categories of Criteria: 24. Identify budgeted and applied (or actual) indirect costs at the level and frequency needed by management for effective control, along with the reasons for any significant variances. Criteria: Categories of Criteria: 25. Summarize the data elements and associated variances through the program organization and/or work breakdown structure to support management needs and any customer reporting specified in the contract. Criteria: Categories of Criteria: 26. Implement managerial actions taken as the result of earned value information. Criteria: Categories of CriteriaRevisions and Data Maintenance: 27. Develop revised estimates of cost at completion based on performance to date, commitment values for material, and estimates of future conditions. Compare this information with the performance measurement baseline to identify variances at completion important to company management and any applicable customer reporting requirements including statements of funding requirements. Categories of Criteria: Organization: Revisions and Data Maintenance; Criteria: 28. Incorporate authorized changes in a timely manner, recording the effects of such changes in budgets and schedules. In the directed effort prior to negotiation of a change, base such revisions on the amount estimated and budgeted to the program organizations. Criteria: Categories of Criteria: 29. Reconcile current budgets to prior budgets in terms of changes to the authorized work and internal replanning in the detail needed by management for effective control. Criteria: Categories of Criteria: 30. Control retroactive changes to records pertaining to work performed that would change previously reported amounts for actual costs, earned value, or budgets. Adjustments should be made only for correction of errors, routine accounting adjustments, effects of customer or management directed changes, or to improve the baseline integrity and accuracy of performance measurement data. Criteria: Categories of Criteria: 31. Prevent revisions to the program budget except for authorized changes. Criteria: Categories of Criteria: 32. Document changes to the performance measurement baseline. Source: Interim Defense Acquisition Guidebook, app. 4. Note: In the Interim Defense Acquisition Guidebook, DOD states that these guidelines are reproduced from the American National Standards (ANSI) Institute/Electronic Industries Alliance (EIA) EVM System Standard (ANSI/EIA-748-98), Chapter 2 (May 19, 1998). [End of table] The criteria have become the standard for EVM and have also been adopted by major US government agencies, industry, and the governments of Canada and Australia. The full application of EVM system criteria is appropriate for large cost reimbursable contracts where the government bears the cost risk. For such contracts, the management discipline described by the criteria is essential. In addition, data from an EVM system have been proven to provide objective reports of contract status, allowing numerous indices and performance measures to be calculated. These can then be used to develop accurate estimates of anticipated costs at completion, providing early warning of impending schedule delays and cost overruns. The standard format for tracking earned value is through a Cost Performance Report (CPR). The CPR is a monthly compilation of cost, schedule and technical data which displays the performance measurement baseline, any cost and schedule variances from that baseline, the amount of management reserve used to date, the portion of the contract that is authorized unpriced work, and the contractor's latest revised estimate to complete the program. As a result, the CPR can be used as an effective management tool because it provides the program manager with early warning of potential cost and schedule overruns. Using data from the CPR, a program manager can assess trends in cost and schedule performance. This information is useful because trends tend to continue and can be difficult to reverse. Studies have shown that once programs are 15 percent complete the performance indicators are indicative of the final outcome. For example, a CPR showing a negative trend for schedule status would indicate that the program is behind schedule. By analyzing the CPR, one could determine the cause of the schedule problem such as delayed flight tests, changes in requirements, or test problems because the CPR contains a section that describes the reasons for the negative status. A negative schedule condition is a cause for concern, because it can be a predictor of later cost problems since additional spending is often necessary to resolve problems. For instance, if a program finishes 6 months later than planned, additional costs will be expended to cover the salaries of personnel and their overhead beyond what was originally expected. CPR data provides the basis for independent assessments of a program's cost and schedule status and can be used to project final costs at completion in addition to determining when a program should be completed. Examining a program's management reserve is another way that a program can use a CPR to determine potential issues early on. Management reserves, which are funds that may be used as needed, provide flexibility to cope with problems or unexpected events. EVM experts agree that transfers of management reserve should be tracked and reported because they are often problem indicators. An alarming situation arises if the CPR shows that the management reserve is being used at a faster pace than the program is progressing toward completion. For example, a problem would be indicated if a program has used 80 percent of its management reserve but only completed 40 percent of its work. A program's management reserve should contain at least 10 percent of the cost to complete a program so that funds will always be available to cover future unexpected problems that are more likely to surface as the program moves into the testing and evaluation phase. [End of section] Appendix V: GAO Contact and Staff Acknowledgments: GAO Contact: Barbara Haynes (256) 922-7500: Acknowledgments: In addition to the individual named above Yvette Banks, Myra Watts Butler, Cristina Chaplain, Roger Corrado, Jennifer Echard, Dayna Foster, Matt Lea, Karen Richey, and Randy Zounes made key contributions to this report. FOOTNOTES [1] In January 2002, the Secretary of Defense created the Missile Defense Agency and consolidated all ballistic missile defense programs under the new agency. Former missile defense acquisition programs are now referred to as elements of a single ballistic missile defense system. [2] The intended performance of the Block 2004 capability is described in a classified annex to this report. [3] Technological maturity for starting product development or systems integration is achieved when prototype hardware with the desired form, fit, and function has been proven in a realistic operational environment. See U.S. General Accounting Office, Best Practices: Better Management of Technology Development Can Improve Weapon System Outcomes, GAO/NSIAD-99-162 (Washington, D.C.: July 1999). [4] The kill vehicle is the weapon component of the GMD element that attempts to detect and destroy threat warheads through "hit-to-kill" impacts. [5] The battle management component is the integrating and controlling component of the GMD element. The fire control software plans engagements and tasks GMD components to execute a missile defense mission. [6] The EVM system is a management tool widely used by DOD to compare the value of contractor's work performed to the work's actual cost. The tool measures the contractor's actual progress against its expected progress and enables the government and contractor to estimate the program's remaining cost. [7] An interim baseline is often established by the contractor when the government has authorized work, but the requirements and terms of the work have not yet been negotiated. Until negotiations are completed, the contractor develops a baseline using proposed cost that has been divided among work packages with associated budgets and schedule. [8] An operational environment is a real-world environment (e.g., flight demonstration) that addresses all of the operational requirements and specifications demanded of the final product. [9] U.S. General Accounting Office, Missile Defense: Knowledge-Based Practices Being Adopted, but Risks Remain, GAO-03-441 (Washington, D.C.: Apr. 30, 2003). This report presents our analysis of MDA's new approach for developing missile defense technology. [10] U.S. General Accounting Office, Best Practices: Better Management of Technology Development Can Improve Weapon System Outcomes, GAO/ NSIAD-99-162 (Washington, D.C.: July 1999). [11] A breadboard is a collection of integrated components that provide a representation of a system/subsystem that can be used to determine concept feasibility and to develop technical data. A breadboard is typically configured for laboratory use to demonstrate the technical principals of immediate interest. [12] A relevant environment is defined as a testing environment that simulates key aspects of the operational environment. [13] Integrated flight tests of the GMD element are real-world demonstrations of system performance during which an interceptor is launched to engage and intercept a mock warhead above the atmosphere. [14] See classified annex for further details. [15] See classified annex for further details. [16] The hardware of the Beale and Cobra Dane radars is mature since both are currently in operation for other missions, namely, integrated tactical warning and technical intelligence, respectively. Adding the ballistic missile defense mission to these radars requires primarily software-related development and testing. [17] Ground testing of interim software builds to be mounted on the Beale radar is ongoing. [18] We calculated program cost from 1997 forward because the National Missile Defense program was established at that time. [19] The cost to develop and field the initial GMD capability and the ballistic missile defense test bed is funded in MDA's budget within the Defense Wide Research, Development, Test and Evaluation appropriation. MDA is not requesting any procurement, military construction, or military personnel funds for this effort. [20] The EVM system is a management tool widely used by DOD to compare the value of contractor's work performed to the work's actual cost. The tool measures the contractor's actual progress against its expected progress and enables the government and contractor to estimate the program's remaining cost. [21] Department of Defense, Earned Value Management Implementation Guide (Washington, D.C.: Dec. 1996, as revised, p. 10). [22] DCMA is the agency that DOD has given responsibility for validating contractors' Earned Value data. [23] Earned Value Management Implementation Guide, pp. 34 and 36. [24] Defense Contract Management Agency, Ground-Based Midcourse Defense Monthly Assessment Report Contract No. HQ0006-01-C-0001 for Missile Defense Agency (Seal Beach, Calif.: Dec. 2002, p. 10). DCMA reported that cost performance reports were giving "— a misleading feeling that everything in the program is OK. For the 2nd month in a row, [the prime contractor] has covered up a significant Variance-at-Completion (- $107,800K) — by taking money out of Management Reserve (MR). This is not the intended purpose of using MR funds. [The prime contractor] is reporting a $0 Variance-At-Completion [VAC] by subtracting $107,800K from MR to reduce VAC to $0. Based on prior performance to date, this could be an indication of a trend for growth of the EAC [estimate-at- completion]." [25] Defense Federal Acquisition Regulation Supplement clause 252.234- 7001, EVM System (March 1998). GAO's Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.