Defense Acquisitions

Space System Acquisition Risks and Keys to Addressing Them Gao ID: GAO-06-776R June 1, 2006

On April 6, 2006, we testified before Congress on the Department of Defense's (DOD) space acquisitions. In fiscal year 2007, DOD expects to spend nearly $7 billion to acquire space-based capabilities to support current military and other government operations as well as to enable DOD to transform the way it collects and disseminates information, gathers data on its adversaries, and attacks targets. Despite its growing investment in space, however, DOD's space system acquisitions have experienced problems over the past several decades that have driven up costs by hundreds of millions, even billions, of dollars; stretched schedules by years; and increased performance risks. In some cases, capabilities have not been delivered to the warfighter after decades of development. Within this context, Congress requested that we provide additional comments regarding the need for better program management, space acquisition policy, and DOD's Space Radar and Transformational Satellite Communications System acquisitions.

GAO provided information on the top obstacles to achieving program success from the point of view of program mangers. We found that the top obstacles are funding instability, requirements instability, staffing problems, excessive oversight and inexperienced leadership. Estimated costs have been high, and grown, for DOD's Space Based Infrared System (SBIRS)-High, the Evolved Expendable Launch Vehicle program, the Advanced Extremely High Frequency Satellite (AEHF) Program, the National Polar-orbiting Operational Environmental Satellite System (NPOESS), and the Space Based Infrared System-Low program also had high estimated costs which increased. Additionally, the Space Based Infrared System-Low program and SBIRS-High both overpromised capabilities. DOD has been taking actions to improve cost estimating and we are in the process of addressing these actions. To address the problem of low levels of technological maturity, DOD has committed to delay the development of one new major space program--the Transformational Satellite Communications System (TSAT)--until technology needs are better understood. It has also committed to deliver new space-based capabilities in an incremental fashion so that acquisition efforts can be more executable and the science and technology base can be more engaged in major space programs. DOD has also faced issues regarding the addition of new requirements well into the acquisition phase. Our past reports have pointed to requirements setting problems in the AEHF, NPOESS, and SBIRS-High programs, and noted that DOD could take further steps to strengthen requirements setting by implementing processes and policies, as needed, which stabilize requirements for acquisitions. GAO and DOD have disagreed on what Technological Readiness Levels (TRL) should beat major decision points for space system acquisitions, and will continue to disagree as long as GAO continues to base its reviews of space programs on best practices and DOD continues to use the wide leeway afforded regarding critical technologies and their maturity levels to initiate product development. We identified the main difference between TRL 6 and 7 as the testing environment. For TRL 6, the testing environment would be a laboratory or a simulated operational environment, and for TRL 7, the testing environment would be an operational environment. Achieving TRL 6 or 7 by the critical design review (CDR) is a matter of risk--if the critical technologies in question are supremely important and have no space-based heritage, then it is warranted to test the technologies in space before proceeding through CDR. To ensure its integration efforts are successful, the TSAT program is planning to demonstrate critical technologies at TRL 6 when key integration tests are conducted in fiscal year 2007, use the results of its first round of integration tests to refine testing during the second, more comprehensive round, conduct a series of independent tests to verify results of contractor testing, and assess the results of the main integration tests before making a decision to enter the production development phase. According to GAO's prior work on best practices, to ensure success integration, programs mangers should ensure that (1) the right validation events occur at the right times, (2) each validation event produces quality results, and (3) the knowledge gained from an event is used to improve the product. Finally, the program manager needs assurance that all testing that has been done is reflective of the capabilities that the program is trying to deliver.



GAO-06-776R, Defense Acquisitions: Space System Acquisition Risks and Keys to Addressing Them This is the accessible text file for GAO report number GAO-06-776R entitled 'Defense Acquisitions: Space System Acquisition Risks and Keys to Addressing Them' which was released on June 2, 2006. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. June 1, 2006: The Honorable Jeff Sessions: Chairman: The Honorable Bill Nelson: Ranking Minority Member: Subcommittee on Strategic Forces: Committee on Armed Services: United States Senate: Subject: Defense Acquisitions: Space System Acquisition Risks and Keys to Addressing Them: On April 6, 2006, we testified before the subcommittee on the Department of Defense's (DOD) space acquisitions. In fiscal year 2007, DOD expects to spend nearly $7 billion to acquire space-based capabilities to support current military and other government operations as well as to enable DOD to transform the way it collects and disseminates information, gathers data on its adversaries, and attacks targets. Despite its growing investment in space, however, DOD's space system acquisitions have experienced problems over the past several decades that have driven up costs by hundreds of millions, even billions, of dollars; stretched schedules by years; and increased performance risks. In some cases, capabilities have not been delivered to the warfighter after decades of development. Within this context, you requested that we provide additional comments regarding the need for better program management, space acquisition policy, and DOD's Space Radar and Transformational Satellite Communications System acquisitions. Your specific questions and our answers are discussed below. Question: What are the top obstacles to achieving program success from the point of view of program managers? As part of a 2005 review[Footnote 1] on program management best practices, we surveyed DOD's major weapon program managers, including some managing space programs, who cited the following as "top" obstacles to achieving successful outcomes in an open ended question: ² funding instability (about 36 percent), ² requirements instability (13 percent), ² staffing problems (8 percent), ² excessive oversight (7 percent), and: ² inexperienced leadership (7 percent). Although the majority of respondents to our survey believed that the initial baselines of their programs were reasonable, a significant group, about 24 percent, responded that their program parameters were not reasonable at the start, and 45 program managers responded that their program had been rebaselined one or more times for cost and schedule increases. In addition, 18 percent said one or more key technologies fell below best practice standards for maturity. Our reviews of space programs are consistent with these views--we have found technologies to be immature at program start for major space programs. Further, in delving deeper into the root causes behind these problems, we have found that competition for funding has incentivized programs to produce optimistic cost and schedule estimates, over promise on capability, suppress bad news, and forsake the opportunity to identify potentially better alternatives. In addition, because DOD starts more weapons programs than it can afford, it invariably finds itself in the position of having to shift funds to sustain programs-- often to the point of undermining well-performing programs to pay for poorly performing ones. We also have found that DOD starts its space programs too early, that is, before it has assurance that the capabilities it is pursuing can be achieved within available resources (time, money technology, people, etc.) and time constraints, and it allows new requirements to be added well into the acquisition phase, a course of action that can further stretch technology challenges. This is encouraged by the funding process, as acquisition programs tend to attract the majority of research, development, test, and evaluation (RDT&E) dollars. Many officials working within the space community agreed that these were key underlying causes of acquisition problems during a review we conducted last year.[Footnote 2] In addition, officials we spoke with also cited pressures resulting from having a diverse array of officials and organizations involved with the space acquisition process, tensions between the science and technology (S&T) and acquisition communities as to who is better suited to translate technology concepts into reality, pressures resulting from short tenures among staff critical to achieving acquisition success, and difficulties in overseeing contractors. Question: Do you believe that the Air Force is addressing these obstacles? The Air Force has recently taken steps to put is Transformational Satellite Communications System (TSAT) program on a more executable track by reducing its expectations in the level of sophistication for the first two satellites so that it can meet its schedule goals. It is also holding off entering product development of the first increment until critical technologies are proven. If the Air Force adheres to this commitment for TSAT and applies it to Space Radar, as it has also informally committed to do, then it would be addressing some of the obstacles noted above. For example, it would reduce the risk of funding instability since cost estimates would be more realistic. In addition, the Air Force has committed to estimating cost and funding new acquisitions to an 80-percent confidence level, strengthening systems engineering, and strengthening the acquisition workforce. And for some specific programs, the Air Force has applied additional mechanisms to regulate requirements. These actions could also remove obstacles, if effectively implemented. However, as we testified, such actions should be accompanied by an investment strategy for space, and ultimately DOD's entire weapons portfolio, to separate wants from needs and to alleviate long-standing pressures associated with competition within DOD to win funding. DOD could also instill the best practices it is now embracing into its space acquisition policy. In addition, we have recommended that DOD, as a whole, take steps to hold people and programs accountable when best practices are not pursued. This will require DOD to empower program managers to make decisions related to funding, staffing, and moving into subsequent phases and to match program manager tenure with delivery of a product. It may also require DOD to tailor career paths and performance management systems to provide incentives for longer tenures. By embracing a model that incorporates all these elements, DOD can achieve better outcomes for its space programs. By not doing so, there will still be incentives and allowances to overpromise capability, underestimate cost and schedule, and to start programs prematurely, which, in turn, can eventually undo other improvement efforts. Question: DOD starts more space and weapons programs than it can afford, which, according to GAO, "pressures programs to underestimate costs and over promise capabilities." Can you provide a few examples of this problem in space programs and [say] if and how the problem is being addressed? Actual costs for nearly every major space acquisition we review each year as part of our annual weapon system assessment have greatly exceeded earlier estimates--a clear indication that programs consistently underestimate costs. For example, the Space Based Infrared System (SBIRS)-High cost estimate climbed from about $4 billion as of October 1996 to over $10 billion in September 2005, and costs are expected to increase further. Estimated costs for the Evolved Expendable Launch Vehicle (EELV) program have climbed from about $15 billion in October 1998 to $27 billion in August 2005 with 43 fewer launches to be purchased than anticipated. Estimated costs for the Advanced Extremely High Frequency Satellite program (AEHF) increased from $5.6 billion as of October 2001 to $6.2 billion as of August 2005, with quantities decreasing from five to three satellites. Estimated costs for the National Polar-orbiting Operational Environmental Satellite System (NPOESS) grew from $5.9 billion in August 2002 to nearly $8 billion in September 2005. Our past reports have also identified cases where programs have overpromised capabilities. For example, the Space Based Infrared System (SBIRS)-Low program started under the assumption that the satellites would be able to detect and track multiple objects and differentiate a threatening warhead from decoys, even though that technology challenge was exceedingly high. In fact, the program was never able to achieve this capability. It was eventually shut down in the face of cost and schedule overruns that came with addressing technology challenges. The SBIRS-High program began with the assumption that there would be four satellites in geosynchronous orbit, but more than 10 years later, DOD plans to reduce the number of satellites it will procure and still does not have the assurance it needs that the missile detection capability can be achieved in time to replace the existing detection system. In addition, DOD has initiated efforts to develop a parallel competing capability with the SBIRS-High program. Similarly, the NPOESS program is now considering dropping some of its planned capability because of technology and design-related challenges. DOD has been taking actions to improve cost estimating and we are in the process of assessing these actions. As mentioned above, for example, it has committed to estimating cost and funding new acquisitions to an 80-percent confidence level. In addition, the Air Force is requiring the use of independent cost estimates--rather than estimates produced by a program office or a contractor. It is also committed to strengthening its cost-estimating capabilities--in terms of people, methodologies, and tools. In regard to the issue of overpromising capability, the Air Force has deferred pursuing some of its more ambitious capabilities on its TSAT program, so that the program can be better positioned to meet its schedule. We do not know at this point whether it will be doing the same for its new Space Radar program. As we underscored in our testimony, it is important that these and other individual actions be made within a framework of broader, systemic improvements to DOD's overall acquisition process, the acquisition workforce, and an overall investment strategy. Question: The second problem is that DOD "starts its space programs too early, that is, before it is sure that the capabilities it is pursuing can be achieved within available resources and time constraints." Can you provide a few examples of this problem in space programs and [indicate] if and how the problem is being addressed? Many of our annual reviews of major space acquisitions show that programs have started with relatively low levels of technology maturity--meaning DOD does not have assurance that the technologies can work as intended. This includes, AEHF, NPOESS, SBIRS-High, and SBIRS- Low--now known as the Space Tracking and Surveillance System. Exceptions include the Navy's Mobile User Objective System, or MUOS (though the program later added two additional technologies that did not meet best practices standards for maturity) and the Global Positioning System Block IIF. At times, we have found that key sensors to be included in new satellites were not fully tested, or even prototyped, before being included in a program. In other cases, technologies used to support the health of the overall satellite, such as cooling systems, were immature. And in other cases, software needs were vastly underestimated. In the case of AEHF, technical resources to support security needs were underestimated. Many programs we have studied felt the need to start the acquisition process before such needs were better understood because acquisition programs tend to attract more funding than science and technology efforts. In addition, in the case of space, programs have historically attempted to satisfy all requirements in a single step, regardless of the design challenge or the maturity of the technologies to achieve the full capability. While this is partly attributable to a desire to speed delivery of capability, it has perversely slowed down programs, since programs were at increased risk of facing costly and disruptive technical and design problems. As noted previously, DOD has committed to delay the development of one new major space program--TSAT--until technology needs are better understood. It has also committed to deliver new space-based capabilities in an incremental fashion so that acquisition efforts can be more executable and the science and technology base can be more engaged in major space programs. It has not taken such action yet on other new programs, notably Space Radar, though it has informally committed to. In addition, DOD's space acquisition policy still allows major acquisitions to begin without demonstrating that technology can work as intended. Question: A third issue is that DOD has "allowed new requirements to be added well into the acquisition phase." I would also add that sometimes the original requirements may be unrealistic or unaffordable and that this may be part of the problem. Can you provide a few examples of the requirements problem in space programs and [indicate] if and how the problem is being addressed? Our past reports have pointed to requirements setting problems in the AEHF, NPOESS, and SBIRS-High programs. In the case of SBIRS-High, we pointed to problems related to not adequately defining requirements up front. These were further detailed in subsequent DOD studies, including those by the SBIRS-High Independent Review Team and the Defense Science Board. Both noted that the acquisition approach the Air Force was following, known as Total System Performance Responsibility, placed too much responsibility on the part of the contractor to negotiate requirements, and that the process eventually broke down. In the case of NPOESS, we reported in the early phases of the program that the Air Force and the National Oceanic and Atmospheric Administration had difficulty resolving diverging requirements. In the case of AEHF, we reported that DOD substantially and frequently altered requirements and design in the early phases of the program. While considered necessary, some changes increased costs by hundreds of millions of dollars and caused scheduling delays on a program that DOD was trying to accelerate in order to address a potential capability gap. DOD has since rejected the acquisition approaches that led to requirements-setting problems on both SBIRS-High and AEHF. It has also instituted control mechanisms to regulate requirements on SBIRS-High. In our testimony, we noted that DOD could take further steps to strengthen requirements setting by implementing processes and policies, as needed, which stabilize requirements for acquisitions, like NPOESS, that are being shared with other agencies. We have also reported on programs that took on unrealistic or potentially unaffordable requirements. The SBIRS-Low program's pursuit of discrimination capability is an older example of such a program. More recently, we have pointed to affordability and feasibility issues related to Space Radar and the TSAT programs, which together, have been preliminarily estimated to cost about $40 billion. Specifically, we have stated that DOD was planning to start these acquisitions even when many of their critical technologies were still immature, and it was pursuing a highly ambitious path in terms of the technology push. Given that these systems were among the most complex programs ever undertaken for space, they were being counted on to enable wider DOD transformation efforts, and DOD was already contending with highly problematic space efforts, we believed that DOD could not afford to pursue such risky approaches for TSAT and Space Radar. As noted earlier, DOD has taken steps to ensure it is pursuing realistic requirements for TSAT, and it has informally committed to do the same for Space Radar. Question: Is there a clear definition of each Technology Readiness Level (TRL) that all of you agree on (GAO and DOD) and that exists in writing and that clearly applies to space programs? The National Aeronautics and Space Administration (NASA) developed the original ranking and definitions of technology maturity levels. GAO and DOD agree on the TRL definitions--in its reports, GAO continues to reference the TRL scale for assessing critical technologies from DOD's Interim Defense Acquisition Guidebook (app 6, dated October 30, 2002). However, for space system acquisitions, GAO and DOD have disagreements on what the TRLs should be at major decision points. According to our work on best practices, product development should be initiated after critical technologies have been incorporated into a system prototype and tested in an operational environment--meaning the cold-radiated vacuum of space. Our prior reports have recognized that space systems are uniquely difficult to test in a true operational environment. However, DOD has found ways to test sensors and other critical technologies on experimental satellites. Nonetheless, DOD continues to stand up formal space system acquisitions too early--before critical technologies have been tested in operational or relevant environments- -that is, before DOD has assurance that the capabilities it is pursuing can be achieved. This causes DOD to extend technology invention to its acquisitions, which have reverberating effects and require large amounts of time and money to fix. In these cases, DOD points to its National Security Space Acquisition Policy, which allows it to take such an approach--unlike DOD's acquisition policy for non-space acquisitions, where TRL 7 (testing in an operational environment) is preferred before product development is initiated (TRL 6 is required). As long as GAO continues to base its reviews of space programs on best practices and DOD continues to use the wide leeway afforded in its space acquisition policy regarding critical technologies and their maturity levels to initiate product development, GAO and DOD will continue to have disagreements in this area. Question: What is the difference between TRL 6 and 7 and what is the advantage or disadvantage of being at level 6 or 7 at the [Critical] Design Review? The main difference between TRL 6 and 7 is the testing environment. For TRL 6, the testing environment would be a laboratory or a simulated operational environment, and for TRL 7, the testing environment would be an operational environment--meaning in space. According to GAO's work on best practices, achieving a high level of technology maturity at program start is an important indicator of whether available resources in terms of knowledge, time, money and capacity match the customer's requirements. In addition, the key measure for a successful critical design review (CDR) is when 90-percent of the design drawings have been submitted to manufacturing. When space programs reach CDR and TRLs are below 6, it is unlikely that a high percentage of design drawings would have been released to manufacturing, thereby increasing program risk at this juncture. Another key point to remember is that CDR is the point at which programs begin ordering long-lead parts to build the first few satellites. This investment in hardware is at risk if the technologies do not prove out to work as intended. Achieving TRL 6 or 7 by CDR is a matter of risk--if the critical technologies in question are supremely important and have no space-based heritage, then it is warranted to test the technologies in space before proceeding through CDR. For TSAT, some critical technologies have a heritage of being tested or operated in space, and they are all slated to be at TRL 6 at the time of CDR--an approach that GAO did not fault. Question: The Transformational Communications Satellite program, though still very early in the process, appears to have begun to adopt some of the recommendations of the GAO as well as the Young Panel and is focusing on technology maturity. Integration of the satellite appears to be the next difficult step for the TSAT program. What plans are in place to ensure successful integration? The TSAT program is taking several steps to ensure its integration efforts are successful. First, according to program officials, the plan is to demonstrate critical technologies at TRL 6 when key integration tests are conducted in fiscal year 2007. Second, the program plans to use the results of its first round of integration tests to refine the testing to be conducted during a second round of more comprehensive integration testing. Third, the program is conducting a series of independent tests to verify results of contractor testing as it incrementally builds toward the two main integration tests facing the program--tests of the Next Generation Processor Router and Optical Standards Validation Suite. The program office plans to have knowledge on how these two major subcomponents work to reduce risk by uncovering technical problems before awarding the space segment contract for the design and assembly of the satellites. Finally, the TSAT program also plans to assess the results of the main integration tests before making a decision to enter the production development phase. Question: What actions would you recommend to the programs managers to ensure successful integration? According to GAO's prior work on best practices, leading firms ensure that (1) the right validation events--tests, simulations, and other means for demonstrating product maturity--occur at the right times, (2) each validation event produces quality results, and (3) the knowledge gained from an event is used to improve the product. Fully disclosing the results of tests (from low-level brass board tests to the main integration tests) and documenting the actions taken to address shortcomings further validates product knowledge. It is imperative that problems are fully addressed before rushing efforts to begin the next round of testing. It is also important that program managers use the test and evaluation parameters originally established, and any changes should be fully disclosed along with the reasons for doing so. Finally, the program manager needs assurance that all testing that has been done is reflective of the capabilities that the program is trying to deliver. Rigorous and sophisticated testing early and often will uncover problems when they are relatively easy and inexpensive to fix. Waiting too long to fully stress and test components will put the program in a risky position. In preparing answers to your questions, we relied on our prior work on DOD's space acquisition policy, best practices in weapon system acquisitions, and our reviews of specific space acquisitions as well as DOD studies. In addition, for specific space systems development and cost growth, we relied on our annual assessment of selected major weapon programs. Because we relied on previously issued work, we did not obtain comments from DOD on a draft of this letter. We conducted our work from April 2006 through May 2006 in accordance with generally accepted government auditing standards. We are sending copies of this letter to the Secretaries of Defense and the Air Force and interested congressional committees. We will also make copies available to others upon request. In addition, the report will be available at no charge on the GAO web site at [Hyperlink, http://www.gao.gov]. If you or your staff have any questions concerning these comments, please contact me at (202) 512-4841. Sincerely yours, Signed by: Cristina Chaplain: Acting Director: Acquisition and Sourcing Management: cc: (120561): FOOTNOTES [1] GAO, Best Practices: Better Support of Weapon System Program Managers Needed to Improve Outcomes, GAO-06-110 (Washington, D.C.: Nov. 30, 2005). [2] GAO, Defense Acquisitions: Incentives and Pressures That Drive Problems Affecting Satellite and Related Acquisitions, GAO-05-570R (Washington, D.C.: June 23, 2005). GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.