Military Readiness

DOD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System Gao ID: GAO-09-518 September 25, 2009

The Department of Defense (DOD) reports data about the operational readiness of its forces. In 1999, Congress directed DOD to create a comprehensive readiness system with timely, objective, and accurate data. In response, DOD started to develop the Defense Readiness Reporting System (DRRS). After 7 years, DOD has incrementally fielded some capabilities, and, through fiscal year 2008, reported obligating about $96.5 million. GAO was asked to review the program including the extent that DOD has (1) effectively managed and overseen DRRS acquisition and deployment and (2) implemented features of DRRS consistent with legislative requirements and DOD guidance. GAO compared DRRS acquisition disciplines, such as requirements development, test management, and DRRS oversight activities, to DOD and related guidance, and reviewed the system's current and intended capabilities relative to legislative requirements and DOD guidance. We did not evaluate DOD's overall ability to assess force readiness or the extent that readiness data reflects capabilities, vulnerabilities, or performance issues.

DOD has not effectively managed and overseen the DRRS acquisition and deployment, in large part because of the absence of rigorous and disciplined acquisition management controls and an effective governance and accountability structure for the program. In particular, system requirements have not been effectively developed and managed. For example, user participation and input in the requirements development process was, until recently, limited, and requirements have been experiencing considerable change, are not yet stable, and have not been effectively controlled. In addition, system testing has not been adequately performed and managed. For example, test events for already acquired system increments, as well as currently deployed and operating increments, were not based on well-defined plans or structures, and test events have not been executed in accordance with plans or in a verifiable manner. Moreover, DRRS has not been guided by a reliable schedule of work to be performed and key activities to occur. These program management weaknesses can, in part, be attributed to long-standing limitations in program office staffing and program oversight and accountability. Despite being a DOD-wide program, until April, 2009 DRRS was not accountable to a DOD-wide oversight body, and it was not subject to DOD's established mechanisms and processes for overseeing business systems. Collectively, these acquisition management weaknesses have contributed to a program that has fallen well short of expectations, and is unlikely to meet future expectations. DOD has implemented DRRS features that allow users to report certain mission capabilities that were not reported under the legacy system, but these features are not fully consistent with legislative requirements and DOD guidance; and DOD has not yet implemented other features. The geographic combatant commands are currently reporting their capabilities to execute most of their operations and major war plans in DRRS, and DOD is reporting this additional information to Congress. However, because DRRS does not yet fully interface with legacy systems to allow single reporting of readiness data, the military services have not consistently used DRRS's enhanced capability reporting features. For example, as of May 2009, the Army and Navy had developed interfaces for reporting in DRRS, while the Marine Corps required units to only report in their legacy system. Recently, the Marine Corps also began developing an interface and has done limited reporting in DRRS. In addition, DRRS has not fully addressed the challenges with metrics that led Congress to require a new readiness reporting system. DRRS metrics are less objective and precise, and no more timely than the legacy system metrics. Users have also noted that DRRS lacks some of the current and historical data and connectivity with DOD's planning systems necessary to manage and deploy forces. Until these limitations are fully addressed, DRRS will not have the full complement of features necessary to meet legislative and DOD requirements, and users will need to rely on legacy reporting systems to support mission-critical decisions.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-09-518, Military Readiness: DOD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System This is the accessible text file for GAO report number GAO-09-518 entitled 'Military Readiness: DOD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System' which was released on September 25, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Subcommittee on Readiness and Management Support, Committee on Armed Services, U.S. Senate: United States Government Accountability Office: GAO: September 2009: Military Readiness: DOD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System: Military Readiness: GAO-09-518: GAO Highlights: Highlights of GAO-09-518, a report to the Subcommittee on Readiness and Management Support, Committee on Armed Services, U.S. Senate. Why GAO Did This Study: The Department of Defense (DOD) reports data about the operational readiness of its forces. In 1999, Congress directed DOD to create a comprehensive readiness system with timely, objective, and accurate data. In response, DOD started to develop the Defense Readiness Reporting System (DRRS). After 7 years, DOD has incrementally fielded some capabilities, and, through fiscal year 2008, reported obligating about $96.5 million. GAO was asked to review the program including the extent that DOD has (1) effectively managed and overseen DRRS acquisition and deployment and (2) implemented features of DRRS consistent with legislative requirements and DOD guidance. GAO compared DRRS acquisition disciplines, such as requirements development, test management, and DRRS oversight activities, to DOD and related guidance, and reviewed the system‘s current and intended capabilities relative to legislative requirements and DOD guidance. We did not evaluate DOD‘s overall ability to assess force readiness or the extent that readiness data reflects capabilities, vulnerabilities, or performance issues. What GAO Found: DOD has not effectively managed and overseen the DRRS acquisition and deployment, in large part because of the absence of rigorous and disciplined acquisition management controls and an effective governance and accountability structure for the program. In particular, system requirements have not been effectively developed and managed. For example, user participation and input in the requirements development process was, until recently, limited, and requirements have been experiencing considerable change, are not yet stable, and have not been effectively controlled. In addition, system testing has not been adequately performed and managed. For example, test events for already acquired system increments, as well as currently deployed and operating increments, were not based on well-defined plans or structures, and test events have not been executed in accordance with plans or in a verifiable manner. Moreover, DRRS has not been guided by a reliable schedule of work to be performed and key activities to occur. These program management weaknesses can, in part, be attributed to long- standing limitations in program office staffing and program oversight and accountability. Despite being a DOD-wide program, until April, 2009 DRRS was not accountable to a DOD-wide oversight body, and it was not subject to DOD‘s established mechanisms and processes for overseeing business systems. Collectively, these acquisition management weaknesses have contributed to a program that has fallen well short of expectations, and is unlikely to meet future expectations. DOD has implemented DRRS features that allow users to report certain mission capabilities that were not reported under the legacy system, but these features are not fully consistent with legislative requirements and DOD guidance; and DOD has not yet implemented other features. The geographic combatant commands are currently reporting their capabilities to execute most of their operations and major war plans in DRRS, and DOD is reporting this additional information to Congress. However, because DRRS does not yet fully interface with legacy systems to allow single reporting of readiness data, the military services have not consistently used DRRS‘s enhanced capability reporting features. For example, as of May 2009, the Army and Navy had developed interfaces for reporting in DRRS, while the Marine Corps required units to only report in their legacy system. Recently, the Marine Corps also began developing an interface and has done limited reporting in DRRS. In addition, DRRS has not fully addressed the challenges with metrics that led Congress to require a new readiness reporting system. DRRS metrics are less objective and precise, and no more timely than the legacy system metrics. Users have also noted that DRRS lacks some of the current and historical data and connectivity with DOD‘s planning systems necessary to manage and deploy forces. Until these limitations are fully addressed, DRRS will not have the full complement of features necessary to meet legislative and DOD requirements, and users will need to rely on legacy reporting systems to support mission-critical decisions. What GAO Recommends: GAO is making recommendations to address the risks facing DOD in acquiring and developing DRRS and increase the chance of success. DOD agreed or partially agreed with three of our eight recommendations, and disagreed with the remaining five because it stated that it was already taking actions in these areas. View [hyperlink, http://www.gao.gov/products/GAO-09-518] or key components. For more information, contact Sharon Pickup at (202) 512- 9619 or pickups@gao.gov, or Randolph C. Hite at (202) 512-3439 or hiter@gao.gov. [End of section] Contents: Letter: Background: DOD Disagreed with GAO's Prior Recommendation to Develop an Implementation Plan to Guide DRRS Development: DOD Has Not Effectively Managed and Overseen the Acquisition and Deployment of DRRS: Some DRRS Features Are Operational but Are Not Fully Consistent with Legislative Requirements and DOD Guidance: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Objectives, Scope, and Methodology: Appendix II: Detailed Analysis of DRRS Acquisition and Deployment Management and Oversight: Appendix III: Comments from the Department of Defense: Appendix IV: GAO Contacts and Staff Acknowledgments: Tables: Table 1: DRRS Capability Modules: Table 2: Organizations Interviewed during Our Review: Table 3: DRRS Satisfaction of Nine Schedule-Estimating Key Practices: Figures: Figure 1: Air Force and Marine Corps Dual Reporting Requirements to Meet Readiness Reporting Guidelines: Figure 2: Schedule for Developing and Implementing DRRS Capabilities: Figure 3: Changes in Estimated Dates for DRRS Capabilities: Abbreviations: CJCSI: Chairman of the Joint Chiefs of Staff Instruction: DIO: DRRS Implementation Office: DOD: Department of Defense: DRRS: Defense Readiness Reporting System: ESORTS: Enhanced Status of Resources and Training System: GSORTS: Global Status of Resources and Training System: JITC: Joint Interoperability Test Command: MAIS: Major Automated Information System: OSD: Office of the Secretary of Defense: SIPRNet: Secure Internet Protocol Router Network: TEMP: Test and Evaluation Master Plan: USD (P&R): Under Secretary of Defense for Personnel and Readiness: [End of section] United States Government Accountability Office: Washington, DC 20548: September 25, 2009: The Honorable Evan Bayh: Chairman: The Honorable Richard Burr: Ranking Member: Subcommittee on Readiness and Management Support: Committee on Armed Services: United States Senate: To assess the ability of U.S. forces to execute the wartime missions for which they were designed, as well as other assigned missions, and to make decisions about deploying forces, the Department of Defense (DOD) relies heavily on readiness information derived from multiple information systems. Over the years, we and others have identified shortcomings in DOD's readiness assessment and reporting, such as limitations in the completeness and precision of readiness data and a tendency to focus on examining the status of personnel, equipment, and other resources rather than broader capabilities. Congress addressed DOD's readiness reporting in the Strom Thurmond National Defense Authorization Act for Fiscal Year 1999[Footnote 1] by adding section 117 to Title 10 of the U.S. Code, directing the Secretary of Defense to establish a comprehensive readiness reporting system to measure, in an "objective, accurate, and timely manner," the capability of the armed forces to carry out the National Security Strategy prescribed by the President, the defense planning guidance provided by the Secretary of Defense, and the National Military Strategy prescribed by the Chairman of the Joint Chiefs of Staff. In June 2002, the Deputy Secretary of Defense directed[Footnote 2] the Under Secretary of Defense for Personnel and Readiness (USD P&R) to develop, field, maintain, and fund the Enhanced Status of Resources and Training System (ESORTS), which is the automated readiness reporting system within the Defense Readiness Reporting System (DRRS).[Footnote 3] He also directed that DRRS build upon existing processes and readiness assessment tools to establish a capabilities-based, adaptive, near-real-time readiness reporting system. In addition, in June 2004, the Secretary of Defense directed USD (P&R) to develop DRRS in a manner that would support the data requirements of various users of readiness information, such as the Chairman of the Joint Chiefs of Staff, the combatant commanders, the Secretaries of the military departments, and the Chief of the National Guard Bureau, including their requirements for data on the availability, readiness, deployment, and redeployment of forces.[Footnote 4] USD (P&R) established a DRRS Implementation Office (DIO) to manage the system's acquisition, including managing system development and engaging the user community. The DIO has used support contractors to develop the system and reported obligating about $96.5 million for DRRS from fiscal year 2001 through fiscal year 2008. Some fielded system capabilities are currently being used to varying degrees by the user community. The DIO originally estimated that DRRS would achieve full operational capability in fiscal year 2007, but currently expects DRRS to reach full capability in 2014. In September 2008, the DIO projected that it would spend about $135 million through fiscal year 2014. Recognizing that DRRS was not yet fully deployed or operational and in light of our prior work on readiness-related issues, you asked us to review DOD's efforts to develop and implement DRRS, including the program's status, and the extent that DRRS addresses the challenges that led Congress to require a new system, such as the availability of information on capabilities, and the precision, timeliness, reliability, and objectivity of readiness metrics. In addressing these issues, we assessed the extent to which (1) DOD has effectively managed and overseen DRRS acquisition and deployment, and (2) features of DRRS have been implemented and are consistent with legislative requirements and DOD guidance. To determine the extent that DOD has effectively managed and overseen DRRS acquisition and deployment, we analyzed a range of program documentation and interviewed cognizant program and contractor officials relative to the following acquisition management disciplines: requirements development and management, test management, schedule reliability, and human capital. For each discipline, we compared key program documentation, such as a requirements management plan, test and evaluation master plans, and test reports to relevant DOD, federal, and related guidance, and we analyzed a statistical sample of individual requirements and test cases to determine consistency among them. In addition, we attended meetings of organizations established to monitor or govern DRRS development and reviewed information from meetings that we did not attend and interviewed officials associated with these meetings. To determine the extent to which the features of DRRS have been implemented and are consistent with legislative requirements and DOD guidance, we reviewed criteria such as the legislation that mandated a comprehensive DOD readiness reporting system, the DOD directive that established DRRS, program documentation and USD (P&R) guidance memorandums, DIO briefings to the readiness community, other departmental instructions, and directives and memorandums related to DRRS requirements and implementation. From these documents, we identified desired features of DRRS and compared them to documentary and testimonial evidence concerning system performance during meetings with a full range of officials responsible for developing and using the system. To obtain the developer's perspective, on numerous occasions throughout our review we met with officials from USD (P&R), the DIO, and the three current DRRS contractors. To obtain user perspectives, we met with and surveyed by questionnaire officials from the Joint Staff, the geographic and functional combatant commands, and the services, and also met with officials from USD (P&R). We also attended meetings of organizations established to monitor or govern DRRS development and analyzed information from meetings that we did not attend. We also directly observed the system's capabilities through our own limited use of the system and by observing others using the system. We did not evaluate the department's overall ability to assess the readiness of its forces or the extent that data contained in any of its readiness reporting systems, including DRRS and GSORTS, reflect capabilities, deficiencies, vulnerabilities, or performance issues. Our review focused on acquisition and program management issues, such as requirements management, schedule and human capital requirements, the current usage of DRRS, and the extent to which DRRS' features address legislative requirements and DOD guidance. We conducted this performance audit from April 2008 through August 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Additional details on our scope and methodology are in appendix I. Background: Historically, DOD has used its readiness assessment system to assess the ability of units and joint forces to fight and meet the demands of the national security strategy. DOD's readiness assessment and reporting system is designed to assess and report on military readiness at three levels--(1) the unit level; (2) the joint force level; and (3) the aggregate, or strategic, level. Using information from its readiness assessment system, DOD prepares and sends legislatively mandated Quarterly Readiness Reports to Congress. DRRS is DOD's new readiness reporting system that is intended to capture information from the previous system, as well as information about organizational capabilities to perform a wider variety of missions and mission essential tasks. DRRS is also intended to capture readiness information from defense agencies and installations, which were not required to report under the previous system. Some DRRS features are currently fielded and being used to varying degrees by the user community. DOD Collects a Wide Range of Readiness Information to Support Decision Makers Within and Outside DOD: Laws, directives, and guidance, including a DOD directive, Chairman of the Joint Chiefs of Staff Instruction (CJCSI), Secretary of Defense and USD (P&R) memorandums, and service regulations and messages, show that readiness information and data are needed to support a wide range of decision makers. These users of readiness data include Congress, the Secretary of Defense, the Chairman of the Joint Chiefs of Staff, the combatant commanders, the Secretaries of the military departments, and the Chief of the National Guard Bureau. The directives and guidance also list roles and responsibilities for collecting and reporting various types of readiness data. For example, CJCSI 3401.02A[Footnote 5] assigns the service chiefs responsibility for ensuring required global status of resources and training system (GSORTS) reports are submitted. GSORTS is DOD's legacy, resource-based readiness reporting system that provides a broad assessment of unit statuses based on units' abilities to execute the missions for which they were organized or designed as well as the current missions for which they may be employed. The information in the required GSORTS reports includes units' abilities to execute the missions for which they were organized or designed, as well as the status of their training, personnel, and equipment.[Footnote 6] In addition, DOD directive 7730.65, which established DRRS as DOD's new readiness reporting system, assigns the Secretaries of the military departments and the commanders of the combatant commands responsibilities for developing mission essential tasks for all of their assigned missions. [Footnote 7] Requirements and Intended Characteristics of DOD's New Readiness Reporting System: Prior to 1999, we identified challenges with DOD's existing readiness reporting system, GSORTS, and in 1999, Congress directed the Secretary of Defense to establish a comprehensive readiness reporting system. [Footnote 8] The legislation requires the system to measure in an objective, accurate, and timely manner the capability of the armed forces to carry out (1) the National Security Strategy prescribed by the President, (2) the defense planning guidance provided by the Secretary of Defense, and (3) the National Military Strategy prescribed by the Chairman of the Joint Chiefs of Staff. To address the requirements established by Congress, the Office of the Deputy Under Secretary of Defense (Readiness) began in 2001 to build consensus among DOD's senior readiness leaders for an improved readiness assessment system. For example, the Deputy's office distributed a list of key characteristics of the improved readiness assessment system to the leaders in advance of scheduled meetings. The system's key desired characteristics included allowing near-real-time access to readiness data and trends, enabling rapid, low-cost development using classified Internet technology, and reducing the reporting burdens on people. Since then various directives and memorandums have been issued regarding DRRS responsibilities, requirements, and related issues. For example: * On June 3, 2002, the Deputy Secretary of Defense established DOD's new readiness reporting system, as directed by Congress, by signing DOD Directive 7730.65. According to this directive, DRRS is intended to build upon DOD's existing processes and readiness assessment tools to establish a capabilities-based, near-real-time readiness reporting system. The DRRS directive assigned USD (P&R) responsibilities for developing, fielding, maintaining, and funding ESORTS (the tool to collect capability, resource, and training information) and overseeing DRRS to ensure accuracy, completeness, and timeliness of its information and data, its responsiveness, and its effective and efficient use of modern practices and technologies. In addition, the USD P&R is responsible for ensuring that ESORTS information, where appropriate, is integrated into DOD's planning systems and processes. The directive also states that until ESORTS becomes fully operational, the Chairman of the Joint Chiefs of Staff shall maintain the GSORTS database. * On June 25, 2004, the Secretary of Defense issued a memorandum, which directed USD (P&R) to develop DRRS to support data requirements identified by the Chairman of the Joint Chiefs of Staff, the combatant commanders, the Secretaries of the Military Departments, and the Chief, National Guard Bureau to include availability, readiness, deployment, and redeployment data.[Footnote 9] * On November 2, 2004, USD (P&R) issued a DRRS interim implementation guidance memorandum.[Footnote 10] In this memorandum, the undersecretary noted that he had established a DIO to provide reporting assistance for units. The memorandum also stated that combatant commanders would begin reporting readiness by mission essential tasks by November 30, 2004. The memorandum also directed the services to develop detailed implementing guidance for reporting and assessing mission essential task readiness in ESORTS within their respective services, and set a goal for the services to implement the mission essential task reporting process by September 30, 2005. To meet these mission essential task reporting requirements, USD (P&R) directed commanders to rate their organizational capabilities as (1) yes or "Y", (2) qualified yes or "Q", or (3) no or "N." A "Y" indicates that an organization can accomplish the rated tasks or missions to prescribed standards and conditions in a specified environment. It should reflect demonstrated performance in training or operations. A "Q" indicates that performance has not been demonstrated, and, although data may not readily support a "Y," the commander believes the organization can accomplish the rated task or mission to standard under most conditions. An "N" indicates that an organization cannot accomplish the rated task or mission to prescribed standards in the specified environment at the time of the assessment. * The November 2004 memorandum also stated that the expected transition from GSORTS to ESORTS was scheduled to begin in fiscal year 2005. According to the 2004 memorandum, the ESORTS module of DRRS would provide, among other things, visibility of the latest GSORTS information reported by units, and detailed resource information from authoritative data sources with the capability to aggregate or separate the data. This memorandum signaled a change in program direction. Although the 2002 DOD directive stated that DRRS is intended to build upon DOD's existing processes and readiness assessment tools, the 2004 memorandum indicated that DRRS was to replace GSORTS, as the ESORTS module of DRRS captured both capabilities and resource data. Overview of DRRS Program Management and Oversight Structure: Since its establishment, the DIO has operated within the Office of USD (P&R) and has relied on multiple contractors. To provide governance of DRRS, and enhance communication between the development community, represented by the DIO and contractors, and the user community, which includes the Joint Staff, military services, and combatant commands, USD (P&R) established various bodies with representatives from the user community, including military services, combatant commands, and the defense agencies. Representatives from the Office of USD (P&R) and the Joint Staff currently serve as cochairs of the various bodies. DRRS Battle Staffs comprise colonels, Navy captains, and similar-graded civilians. They track DRRS development and identify issues with the system. At the one-star level, the DRRS General and Flag Officer Steering Committee discusses issues raised by the Battle Staff. In December 2007, USD (P&R) created a committee at the three-star level, referred to as the DRRS Executive Committee. Its charter, finalized about a year later in January 2009, calls for the committee to review and approve proposals and plans to establish policy, processes, and system requirements for DRRS, including approving software development milestones required to reach objectives. To ensure that leadership is provided for the direction, oversight, and execution of DOD's business transformation efforts, including business systems modernization efforts such as DRRS, DOD relies on several entities. These entities include the Defense Business Systems Management Committee, which is chaired by the Deputy Secretary of Defense and serves as the department's highest-ranking governance body and the approval authority for business systems modernization activities; the Investment Review Boards, which are chartered by the Principal Staff Assistants--senior leaders from various offices within DOD--and serve as the review and certification bodies for business system investments in their respective areas of responsibility;[Footnote 11] and the Business Transformation Agency, which is responsible for supporting the Investment Review Boards and for leading and coordinating business transformation efforts across the department. Among other things, the Business Transformation Agency supports the Office of the Under Secretary of Defense, Acquisition, Technology and Logistics in conducting system acquisition risk assessments.[Footnote 12] Disciplined System Acquisition and Oversight Are Keys to Program Success: Our research and evaluations of information technology programs, including business systems modernization efforts within DOD, have shown that delivering promised system capabilities and benefits on time and within budget largely depends on the extent to which key program management disciplines are employed by an adequately staffed program management office. Among other things, these disciplines include a number of practices associated with effectively developing and managing system requirements, adequately testing system capabilities, and reliably scheduling the work to be performed. They also include proactively managing the program office's human capital needs, and promoting program office accountability through executive-level program oversight. DOD acquisition policies and guidance, along with other relevant guidance, recognize the importance of these management and oversight disciplines.[Footnote 13] As we have previously reported, not employing these and other program management disciplines increases the risk that system acquisitions will not perform as intended and require expensive and time-consuming rework. DOD Disagreed with GAO's Prior Recommendation to Develop an Implementation Plan to Guide DRRS Development: In 2003, we reported that, according to USD (P&R) officials, DRRS was a large endeavor, and that development would be challenging and require buy-in from many users.[Footnote 14] We also reported that the program was only a concept without detailed plans to guide its development and implementation. Based on the status of the program at that time and DOD's past record on readiness reporting initiatives, we recommended that the Secretary of Defense direct the Office of USD (P&R) to develop an implementation plan that identified: * performance goals that are objective, quantifiable, and measurable; * performance indicators to measure outcomes; * an evaluation plan to compare program results with established goals; and: * milestones to guide DRRS development to the planned 2007 full capability date. DOD did not agree with our recommendation, stating that it had established milestones, cost estimates, functional responsibilities, expected outcomes, and detailed plans for specific information technology requirements and progress indicators. In evaluating the DOD comments, we noted that DOD had established only two milestones-- initial capability in 2004 and full capability in 2007--and did not have a road map explaining the steps needed to achieve full capability by 2007.[Footnote 15] DOD Has Not Effectively Managed and Overseen the Acquisition and Deployment of DRRS: DOD has not effectively managed and overseen the acquisition and deployment of DRRS in accordance with a number of key program management disciplines that are recognized in DOD acquisition policies and guidance, along with other relevant guidance, and are fundamental to delivering a system that performs as intended on time and within budget. In particular, DRRS requirements have not been effectively developed and managed, and DRRS testing has not been adequately performed and managed. Further, DRRS has not been guided by a reliable schedule of the work needed to be performed and the key activities and events that need to occur. These program management weaknesses can be attributed in part to long-standing limitations in program office staffing and oversight. As a result, the program has not lived up to the requirements set for it by Congress, and the department has not received value from the program that is commensurate with the time and money invested--about 7 years and $96.5 million. Each of these weaknesses are summarized below and discussed in detail in appendix II. DRRS Requirements Have Not Been Effectively Developed and Managed: According to DOD and other relevant guidance, effective requirements development and management includes, among other things, (1) effectively eliciting user needs early and continuously in the system life-cycle process, (2) establishing a stable baseline set of requirements and placing the baseline under configuration management, (3) ensuring that system requirements are traceable backward to higher level business or operational requirements (e.g., concept of operations) and forward to system design documents (e.g., software requirements specification) and test plans, and (4) controlling changes to baseline requirements. However, none of these conditions have been met on DRRS. Specifically, key users have only recently become fully engaged in developing requirements, and requirements have been experiencing considerable change and are not yet stable. Further, different levels of requirements and related test cases have not been aligned with one another, and changes to requirements have not been effectively controlled. As a result, efforts to develop and deliver initial DRRS capabilities have taken longer than envisioned and these capabilities have not lived up to user expectations. These failures increase the risk of future DRRS capabilities not meeting expectations and increase the likelihood that expensive and time-consuming system rework will be necessary. Recent Actions Have Been Taken to Address Limited User Involvement in Developing and Managing Requirements: Until recently, key users were not fully or effectively engaged in DRRS requirements development and management. One of the leading practices associated with effective requirements development is engaging system users early and continuously in the process of defining requirements. However, DIO officials and representatives from the military services and the Joint Staff agree that until recently, key users were not effectively engaged in DRRS requirements development and management, although they disagree at to why user involvement has suffered. Regardless, DRRS Executive Committee direction has improved the situation. Specifically, in January 2008, the committee directed the Joint Staff to conduct an analysis of DRRS capabilities, referred to as the "capabilities gap analysis," which involved the entire readiness community and resulted in 530 additional user requirements. In our view, this analysis is a positive step in addressing long-standing limited involvement by key DRRS users in defining requirements that has contributed to significant delays in the program, as discussed later in the report. DRRS Requirements Are Not Stable: As of April 2009, DRRS requirements continued to be in a state of flux. Establishing an authoritative set of baseline requirements prior to system design and development provides a stable basis for designing, developing, and delivering a system that meets its users' operational needs. However, the fact that these 530 user requirements have recently been identified means that the suite of requirements documentation associated with the system, such as the detailed system requirements, will need to change and thus is not stable. To illustrate, these 530 requirements have not been fully evaluated by the DIO and the DRRS governance boards and according to program officials, have not yet been approved, and thus their impact on the program is not clear. Compounding this instability in the DRRS requirements is the fact that additional changes are envisioned. According to program officials, the changes resulting from the gap analysis and reflected in the latest version of the DRRS Concept of Operations, which was approved by the DRRS Executive Committee in January 2009, have yet to be reflected in other requirements documents, such as the detailed system requirements. Although defining and developing requirements is inherently an iterative process, having a baseline set of requirements that are stable is a prerequisite to effective and efficient system design and development. Without them, the DIO has not been able to deliver a system that meets user needs on time, and it is unlikely that future development and deployment efforts will produce better results. Alignment among Requirements and Related System Design and Testing Artifacts Has Not Been Demonstrated: During our review, DIO officials could not demonstrate that requirements and related system design and testing artifacts are properly aligned. One of the leading practices associated with developing and managing requirements is maintaining bidirectional traceability from high-level operational requirements through detailed lower-level requirements and design documents to test cases. We attempted on three separate occasions to verify the traceability of system requirements backwards to higher-level requirements and forward to lower-level software specifications and test cases, and each time we found that traceability did not exist. DIO and contractor officials attributed the absence of adequate requirements traceability to the ongoing instability in requirements and efforts to update program documentation. Without traceable requirements, the DIO does not have a sufficient basis for knowing that the scope of the design, development, and testing efforts will produce a system solution on time and on budget and that will meet users' operational needs and perform as intended. As a result, the risk is significant that expensive and time- consuming system rework will be required. Changes to Requirements Are Not Being Effectively Controlled: Since the inception of the program in 2002, DRRS has been developed and managed without a formally documented and approved process for managing changes to system requirements. Adopting a disciplined process for reviewing and accepting changes to an approved baseline set of requirements in light of the estimated costs, benefits, and risk of each proposed change is a recognized best practice. However, requirements management and change-control plans developed in 2006 by the DRRS software development contractor, according to DIO officials, were not adequate. To address this, the Joint Staff developed what it referred to as a conceptual requirements change-control process in February 2008, as a basis for the DIO to develop more detailed plans that could be implemented. In January 2009, the DIO drafted more detailed requirements management and configuration management plans, the latter of which the DIO updated in March 2009. However, the plans have yet to be approved and implemented. Until the DIO effectively controls requirements changes, it increases the risk of needed DRRS capabilities taking longer and costing more to deliver than necessary. DRRS Testing Has Not Been Adequately Performed and Key Test Management Structures and Controls Have Not Been Established: According to DOD and other relevant guidance, system testing should be progressive, meaning that it should consist of a series of test events that first focus on the performance of individual system components, then on the performance of integrated system components, followed by system-level tests that focus on whether the system (or major system increments) are acceptable, interoperable with related systems, and operationally suitable to users. For this series of related test events to be conducted effectively, each test event needs to be executed in accordance with well-defined test plans, the results of each test event need to be captured and used to ensure that problems discovered are disclosed and corrected, and all test events need to be governed by a well-defined test management structure. However, the DIO cannot demonstrate that it has adequately tested any of the DRRS increments, referred to as system releases and subreleases, even though it has already acquired and partially deployed a subset of these increments. Moreover, the DIO has yet to establish the test management structures and controls needed to effectively execute DRRS testing going forward. More specifically, the test events for already acquired, as well as currently deployed and operating, DRRS releases and subreleases were not based on well-defined plans. For example, the test plan did not include a schedule of activities to be performed or defined roles and responsibilities for performing them. Also, the test plan did not consistently include test entrance and exit criteria, a test defect management process, and metrics for measuring progress. Further, test events have not been fully executed in accordance with plans, or executed in a verifiable manner, or both. For example, although increments of DRRS functionality have been put into production, the DIO has not performed system integration testing, system acceptance testing, or operational testing on any DRRS release or subrelease. Moreover, the results of all executed test events have not been captured and used to ensure that problems discovered were disclosed to decision makers, and ultimately corrected. For example, the DIO has not captured the test results for at least 20 out of 63 DRRS subreleases. Test results that were captured did not include key elements, such as entrance/exit criteria status and unresolved defects and applicable resolution plans. The DIO has also not established an effective test management structure to include, for example, a clear assignment of test management roles and responsibilities, or a reliable schedule of planned test events. Compounding this absence of test management structures and controls is the fact that the DIO has yet to define how the development and testing to date of a series of system increments (system releases and subreleases) relate to the planned development and testing of the 10 system modules established in January 2009. (See table 1 for a list and description of these modules.) Collectively, this means that it is unlikely, that already developed and deployed DRRS increments can perform as intended and meet user operational needs. Equally doubtful are the chances that the DIO can adequately ensure that yet-to-be developed DRRS increments will meet expectations. Table 1: DRRS Capability Modules: Module: 1. Joint Mission Readiness; Definition: Enables selected users to see and assess the highest strategic-level "joint" readiness data. Module: 2. Unit Mission Readiness; Definition: Captures unit-level readiness data, as well as authoritative resources data (e.g., personnel, equipment) fed from external systems owned by military services. Module: 3. Asset Visibility; Definition: Allows authoritative resources data from military services to be provided in near-real-time, and certifies them. Module: 4. Business Intelligence; Definition: Provides the ability to query and analyze readiness and asset visibility data, based on the desires and needs of the user. Module: 5. Installation Readiness; Definition: Allows readiness reporting by installations, training/firing ranges and other infrastructure facilities. Module: 6. Readiness Reviews; Definition: Enables forcewide readiness reviews to be conducted, such as the Joint Combat Capability Assessment review process. Module: 7. Planning/Execution Support; Definition: Allows DRRS to interact with other planning systems and processes, such as the Global Force Management and Adaptive Planning and Execution communities. Module: 8. Historical Data; Definition: Focuses on the historical collection and presentation of information. Also includes the Continuity of Operations (COOP) capability. Module: 9. System Architecture; Definition: Meets the underlying information technology system specifications, such as standards for information assurance, interoperability, bandwidth, and other issues. Module: 10. End-User Usability; Definition: Fulfills end-user support of the DRRS application to include training, customer support, and documentation. Source: DOD. [End of table] DRRS Has Not Been Guided by a Reliable Schedule: The success of any program depends in part on having a reliable schedule that defines, among other things, when work activities will occur, how long they will take, and how they are related to one another. From its inception in 2002 until November 2008, the DIO did not have an integrated master schedule, and thus has long been allowed to proceed without a basis for executing the program and measuring its progress. In fact, the only milestone that we could identify for the program prior to November 2008 was the date that DRRS was to achieve full operational capability, which was originally estimated to occur in fiscal year 2007, but later slipped to fiscal year 2008 and then fiscal year 2011, and is now fiscal year 2014--a 7-year delay. Moreover, the DRRS integrated master schedule that was first developed in November 2008, and was updated in January 2009 and again in April 2009 to address limitations that we identified and shared with the program office, is still not reliable. Specifically, our research has identified nine practices associated with developing and maintaining a reliable schedule.[Footnote 16] These practices are (1) capturing all key activities, (2) sequencing all key activities, (3) assigning resources to all key activities, (4) integrating all key activities horizontally and vertically, (5) establishing the duration of all key activities, (6) establishing the critical path for all key activities, (7) identifying float between key activities, (8) conducting a schedule risk analysis, and (9) updating the schedule using logic and durations to determine the dates for all key activities.[Footnote 17] The program's latest integrated master schedule does not address three of the practices and only partially addresses the remaining six. For example, the schedule does not establish a critical path for all key activities, nor does it include a schedule risk analysis, and it is not being updated using logic and durations to determine the dates for all key activities. In addition, the schedule introduces considerable concurrency across key activities and events for several modules, which introduces increased risk. These limitations in the program's latest integrated master schedule, coupled with the program's 7-year slippage to date and continued requirements instability, make it likely that DRRS will incur further delays. DIO Lacks Adequate Staffing and an Effective Approach to Meeting Its Human Capital Needs: The DIO does not currently have adequate staff to fulfill its system acquisition and deployment responsibilities, and it has not managed its staffing needs in an effective manner. Effective human capital management should include an assessment of the core competencies and essential knowledge, skills, and abilities needed to perform key program management functions, an inventory of the program's existing workforce capabilities, an analysis of the gap between the assessed needs and the existing capabilities, and plans for filling identified gaps. DIO performs a number of fundamental DRRS program management functions, such as acquisition planning, performance management, requirements development and management, test management, contractor tracking and oversight, quality management, and configuration management. To effectively perform such functions, program offices, such as the DIO, need to have not only well-defined policies and procedures and support tools for each of these functions, but also sufficient human capital to implement the processes and use the tools throughout the program's life cycle. However, the DIO is staffed with only a single full-time government employee--the DIO Director. All other key program office functions are staffed by either contractor staff or staff temporarily detailed, on an as-needed basis, from other DOD organizations. In addition, key positions, such as performance manager and test manager, have either not been established or are vacant. According to DIO and contractor officials, they recognize that additional program management staffing is needed but stated that requests for additional staff had not been approved by USD (P&R) due to competing demands for staffing. Further, they stated that the requests were not based on an assessment of the program's human capital needs and the gap between these needs and its onboard workforce capabilities. Until DIO adopts a strategic, proactive approach to managing its human capital needs, it is unlikely that it will have an adequate basis for obtaining the people it needs to effectively and efficiently manage DRRS. DOD Executive Oversight of DRRS Has Been Limited: A key principle for acquiring and deploying system investments is to establish a senior-level governance body to oversee the investment and hold program management accountable for meeting cost, schedule, and performance commitments. Moreover, for investments that are organization wide in scope and introduce new ways of doing business, like DRRS, the membership of this oversight body should represent all stakeholders and have sufficient organizational seniority to commit their respective organizations to any decisions reached. For significant system investments, the department's acquisition process provides for such executive governance bodies. For example, Major Automated Information Systems, which are investments over certain dollar thresholds or that are designated as special interest because of, among other things, their mission importance, are reviewed at major milestones by a designated milestone decision authority.[Footnote 18] These authorities are supported by a senior advisory group, known as the Information Technology Acquisition Board, which comprises senior officials from the Joint Staff, the military departments, and staff offices within the Office of the Secretary of Defense. In addition, all business system investments in DOD that involve more than $1 million in obligations are subject to review and approval by a hierarchy of DOD investment review boards that comprise senior DOD leaders, including the Defense Business Systems Management Committee, which is chaired by the Deputy Secretary of Defense. Through these executive oversight bodies and their associated processes, programs are to be, among other things, governed according to corporate needs and priorities, and program offices are to be held accountable for meeting cost, schedule, and performance expectations. Until April 2009, DRRS was not subject to any of DOD's established mechanisms and processes for overseeing information technology systems. As previously discussed, USD (P&R) established the DRRS Battle Staff, which is a group of midlevel military officers and civilians from DRRS stakeholder organizations, and it established a higher-ranked General and Flag Officer Steering Committee, consisting of stakeholder representatives. However, neither of these entities had specific oversight responsibilities or decision-making authority for DRRS. Moreover, neither was responsible for holding the program office accountable for results. According to meeting minutes and knowledgeable officials, these entities met on an irregular basis over the last several years, with as much as a 1-year gap in meeting time for one of them, to discuss DRRS status and related issues. In December 2007, USD (P&R) recognized the need for a more senior-level and formal governance body, and established the DRRS Executive Committee. Since January 2008, this committee, which consists of top- level representatives from stakeholder organizations, has met at least seven times. In January 2009, the DRRS Executive Committee's charter was approved by the Deputy Under Secretary of Defense (Readiness) and the three-star Director of the Joint Staff. According to the charter, the committee is to review and approve proposals and plans to establish policy, processes, and system requirements for DRRS, including approving software development milestones required to reach objectives. Consistent with its charter, the committee has thus far made various program-related decisions, including approving a DRRS concept of operations to better inform requirements development, and directing the Joint Staff to conduct an analysis to identify any gaps between DRRS requirements and user needs. However, the committee has not addressed the full range of acquisition management weaknesses previously discussed in this report, and it has not taken steps to ensure that the program office is accountable for well-defined program baseline requirements. More recently, the DOD Human Resources Management Investment Review Board and the Defense Business Systems Management Committee reviewed DRRS and certified and approved, respectively, the program to invest $24.625 million in fiscal years 2009 and 2010. These entities comprise senior leadership from across the department, including the Deputy Secretary of Defense as the Defense Business Systems Management Committee Chair, military service secretaries, the defense agency heads, principal staff assistants, and representatives from the Joint Staff and combatant commands. However, neither the Investment Review Board's certification nor the Defense Business Systems Management Committee's approval was based on complete and accurate information from USD (P&R). Specifically, the certification package submitted to both oversight bodies by the USD (P&R) precertification authority (Office of Readiness Programming and Assessment) stated that DRRS was on track for meeting its cost, schedule, and performance measures and highlighted no program risks despite the weaknesses discussed in this report. According to the chairwoman of the Investment Review Board, the board does not have a process or the resources to validate the information received from the programs that it reviews.[Footnote 19] Moreover, the chairwoman stated that program officials did not make the board aware of the results of our review that we shared with the DIO prior to either the Investment Review Board or Defense Business Systems Management Committee reviews. Since we briefed the chairwoman, the Investment Review Board has requested that the DIO provide it with additional information documenting DRRS compliance with applicable DOD regulations and statutes. According to USD (P&R) and DIO officials, DRRS was not subject to department executive-level oversight for almost 6 years because, among other things, they did not consider DRRS to be a more complex information technology system. Furthermore, because of the nature of the authority provided to the USD (P&R) in the DRRS charter, they did not believe it was necessary to apply the same type of oversight to DRRS as other information systems within DOD. This absence of effective oversight has contributed to a void in program accountability and limited prospects for program success. Some DRRS Features Are Operational but Are Not Fully Consistent with Legislative Requirements and DOD Guidance: DOD has implemented DRRS features that allow users to report certain mission capabilities that were not reported under the legacy system, but these features are not fully consistent with legislative requirements and DOD guidance; and DOD has not yet implemented other envisioned features of the system. While some users are consistently reporting enhanced capability information, reporting from other users has been inconsistent. In addition, DRRS has not fully addressed the challenges with metrics that were identified prior to 1999 when Congress required DOD to establish a new readiness reporting system. Users have also noted that DRRS lacks some of the current and historical data and connectivity with DOD's planning systems necessary to manage and deploy forces. DOD Is Providing Congress with DRRS Capability Data from the Combatant Commands, but Services Have Not Consistently Reported Capability Data: The geographic combatant commands are capturing enhanced capability data in DRRS, and DOD's quarterly readiness reports to Congress currently contain this information, as well as information that is drawn from DOD's legacy readiness reporting system, GSORTS. However, the military services have not consistently used the enhanced capability reporting features of DRRS. Because DRRS does not yet fully interface with legacy systems to allow single reporting of readiness data, the Army and Navy developed additional system interfaces and are reporting in DRRS. Until May 2009, the Marine Corps directed its units to report only in the legacy system to avoid the burden of dual reporting. The Air Force chose not to develop an interface and instructed its units to report in both DRRS and the legacy system. DRRS and GSORTS both contain capabilities information and resource (numbers of personnel, equipment availability, and equipment condition) and training data. However, DRRS currently provides more capabilities data than GSORTS. When Congress directed DOD to establish a new readiness reporting system, GSORTS was already providing capability information concerning unit capabilities to perform the missions for which they were organized or designed.[Footnote 20] More recently, some of the military services began reporting limited capability information on unit capabilities to perform missions other than those that they were organized or designed to perform into GSORTS.[Footnote 21] However, DRRS is designed to capture capabilities on a wider variety of missions and mission essential tasks. For example, organizations can report their capabilities to conduct missions associated with major war plans and operations such as Operation Iraqi Freedom into DRRS, as well as their capabilities to perform the missions for which they were organized or designed. DRRS also captures capability information from a wider range of organizations than GSORTS. Although the primary (monthly) focus is on operational units and commands, DRRS collects and displays readiness information from defense agencies and installations. Geographic combatant commands--such as U.S. Central Command, U.S. Pacific Command, and U.S. Northern Command--are currently reporting their commands' capabilities to execute most of their operations and major war plans in DRRS. DOD reports this enhanced capability information from the geographic combatant commands in its Quarterly Readiness Report to Congress. The geographic combatant commands are also using DRRS to report their capabilities to perform headquarters- level, joint mission essential tasks, and some of these commands utilize DRRS as their primary readiness reporting tool. For example, U.S. Northern Command uses DRRS to assess risk and analyze capability gaps, and U.S. Pacific Command identifies critical shortfalls by evaluating mission essential task assessments that are captured in DRRS. While DRRS currently has the necessary software to collect and display these enhanced capability data from organizations at all levels throughout DOD, a variety of technical and other factors have hindered service reporting of capability data.[Footnote 22] As a result, the services have either developed their own systems to report required readiness data or have delayed issuing implementing guidance that would require their units to report standardized mission essential task data in DRRS. By 2005, DRRS was able to collect and display mission essential task information from any organizations that had access to a Secure Internet Protocol Router Network (SIPRNet) workstation.[Footnote 23] In August 2005, USD (P&R) issued a memorandum that directed the services to ensure that all of their GSORTS-reporting units were reporting mission essential task capabilities in DRRS by September 30, 2005.[Footnote 24] The memorandum stated that, for tactical units, mission essential tasks were to be drawn from the Service Universal Task List and standardized across like-type entities, such as tank battalions, destroyers, or F-16 squadrons.[Footnote 25] However, two factors that have hindered compliance with the memorandum's direction to report mission essential task capabilities in DRRS are discussed below. Lack of Direct Access to DRRS: While DRRS has been able to collect and display mission essential task data since 2005, some Army and Navy users did not have the means to directly access DRRS and update mission essential task assessments. For example, some ships lacked hardware necessary to be able to transmit their mission essential task data directly into DRRS while at sea. In addition, many National Guard units lacked, and still lack, direct access to the SIPRNet workstations that are necessary to directly input mission essential task data directly into DRRS. However, the Army and the Navy have developed systems, respectively designated DRRS-A and DRRS-N that interface with DRRS and thus allow all of their units to report mission essential task data. After Army and Navy units report mission essential task data in their respective DRRS-A and DRRS-N service systems, the services transmit these data to DRRS. As a result, Army and Navy officials told us that they are currently complying with the requirement to ensure that all their GSORTS-reporting units report mission essential task data in DRRS. Delays in Developing Software Tools to Input Data: Unlike the Army and the Navy, the Marine Corps and the Air Force have not developed their own systems to allow their units to use a single tool to enter readiness data to meet Office of the Secretary of Defense, Chairman of the Joint Chiefs of Staff, and service readiness reporting requirements. While the DIO has developed the software for users to enter mission essential task data into DRRS, the DIO has been unsuccessful in attempts to develop a tool that would allow Air Force and Marine Corps users to enter readiness data to meet all of their readiness reporting requirements through DRRS. As a result, rather than reducing the burden on reporting units, DRRS has actually increased the burden on Air Force and Marine Corps units because they are now required to report readiness information in both DRRS and GSORTS. On September 29, 2005, USD (P&R) issued a memorandum stating that DRRS is the single readiness reporting system for the entire Department of Defense and that legacy systems, such as GSORTS and associated service readiness systems, should be phased out.[Footnote 26] Since that time, officials have discussed whether to phase out GSORTS and tentative dates for this action have slipped several times. In 2001, the Office of the Deputy Undersecretary of Defense (Readiness) listed reducing reporting burdens as a key characteristic of its envisioned improved readiness assessment system. In an effort to eliminate this burden of dual reporting, the DIO began to develop a "current unit status" tool as a means for users to manage unit-specific readiness data and submit required reports in support of all current reporting guidelines.[Footnote 27] The tool was to minimize the burden associated with dual reporting by collecting, displaying, and integrating resource data from service authoritative data sources with GSORTS and DRRS. However, in December 2007, the DIO reported that it was unable to deliver the intended functionality of the "current unit status" tool. Instead, the DIO decided to develop an interim reporting tool, known as the SORTSREP tool, which would not provide the type of new capabilities envisioned for the "current unit status" tool, but would simply replicate the functionality of the input tool that the Air Force and Marines already used to input data into GSORTS. After delays, and 10 months of effort, the DIO delivered the SORTSREP tool to the Marine Corps for review. Based on this review, in December, 2008, the Marine Corps provided the developers and the DIO with 163 pages of detailed descriptions and graphics to explain the SORTSREP tool's deficiencies. It then informed the DIO that it would no longer expend energy and resources to review future versions of the SORTSREP tool and would instead look at leveraging the Army's or Navy's DRRS-A or DRRS-N systems. The Air Force also informed the DIO that it was no longer interested in the SORSTSREP tool, and said efforts should be focused on the "current unit status" tool instead. As a result, the Air Force and Marine Corps are currently faced with dual reporting requirements, as illustrated in figure 1. Figure 1: Air Force and Marine Corps Dual Reporting Requirements to Meet Readiness Reporting Guidelines: [Refer to PDF for image: illustration] Air Force units: RAS-IT: *GSORTS (Designed mission capabilities, resources, and training); DRRS (Missions and mission essential tasks). Marine Corps units: RAS-IT: *GSORTS (Designed mission capabilities, resources, and training); DRRS (Missions and mission essential tasks). TRAS-IT: Readiness Assessment System Input Tool. Source: GAO based on Air Force and Marine Corps information. [End of figure] On March 3, 2009, the Air Force Deputy Chief of Staff (Operations, Plans and Requirements) issued a memorandum that updated the Air Force's previous implementing guidance and directed all GSORTS- reporting units to begin assessing readiness in DRRS based on standardized core task lists within 90 days.[Footnote 28] As a result, Air Force units will report readiness in both DRRS and GSORTS until the DIO is able to deliver the intended functionality of the "current unit status" tool. While some Marine Corps units are reporting their capabilities in DRRS, the Marine Corps had not yet directed its units to report in the system as of May 2009. The Commandant of the Marine Corps had stated that he supported the development and implementation of DRRS, but that he would not direct units to assess mission essential tasks in DRRS until the system met its stated requirements and was accepted as the single readiness reporting system of record. Marine Corps officials said that they did not want to place a burden on operational units, which were fighting or preparing to fight a war, by requiring that they report readiness in two different systems. After we completed our audit work, on May 12, 2009, the Marine Corps issued an administrative message that required that units assess their mission essential tasks and missions in DRRS. The message stated that doing so would improve familiarity with DRRS, which will lead to an easier transition when the Marine Corps fields DRRS-Marine Corps (DRRS-MC).[Footnote 29] Without a viable tool for inputting data, DRRS is not fully integrated with GSORTS or with the service readiness reporting systems and it is not capable of replacing those systems since it does not capture the required data that are contained in those systems. DRRS Enhanced Capability Data Are Not Likely to Fully Address the Challenges with Metrics That Existed Prior to the Establishment of the System: While DRRS is being used to provide Congress with enhanced capability information, the quality of DRRS metrics still faces the same challenges, including limitations in timeliness, precision, and objectivity that existed prior to 1999 when Congress directed DOD to establish a new readiness reporting system. Section 117 of Title 10 of the U.S. Code directed the Secretary of Defense to establish a comprehensive readiness reporting system to measure the capability of the armed forces in an "objective, accurate, and timely manner." However, the enhanced capability data that are captured in DRRS and reported to Congress are no more timely than the readiness data that were being provided to Congress in 1999 using GSORTS. Furthermore, the metrics that are being used to capture the enhanced capability information are less objective and precise than the metrics that were used to report readiness in 1999. Timeliness: The statute directing the development of a new readiness reporting system requires that the reporting system measure in a timely manner the capability of the armed forces to carry out the National Security Strategy, the Secretary of Defense's defense planning guidance, and the National Military Strategy. The legislation also lists a number of specific requirements related to frequency of measurements and updates. For example, the law requires that the capability of units to conduct their assigned wartime missions be measured monthly, and that units report any changes in their overall readiness status within 24 hours of an event that necessitated the change in readiness status.[Footnote 30] In its DRRS directive, DOD assigned USD (P&R) responsibility for ensuring the timeliness of DRRS information and data, and it specified that DRRS was to be a near-real-time readiness reporting system. While DOD is reporting readiness information to Congress on a quarterly basis as required, and units are measuring readiness on a monthly basis, DRRS is not a near-real-time reporting system. Specifically, in DRRS, as in GSORTS, operational commanders assess the readiness of their organizations on a monthly basis or when an event occurs that changes the units' overall reported readiness. Thus, DRRS has not improved the timeliness of the key readiness data that are reported to Congress. According to USD (P&R) officials, DRRS data will be more timely than GSORTS data because DRRS will update underlying data from authoritative data sources between the monthly updates.[Footnote 31] However, DRRS is not yet capturing all the data from the authoritative data sources, and according to service officials, the service systems that support GSORTS also draw information from their service authoritative data sources between the monthly updates. Furthermore, the source and currency of some of the authoritative data that are currently in DRRS are not clearly identified. As a result, some users told us that they are reluctant to use DRRS data to support their decisions. Precision and Objectivity: We previously reported that the readiness information that DOD provided to Congress lacked precision, noting that GSORTS readiness measures that differed by 10 percentage points or more could result in identical ratings, with DOD often not reporting the detailed information behind the ratings outside of the department.[Footnote 32] For example, units that were at 90 and 100 percent of their authorized personnel strengths both were reported as P-1 in DOD's reports to Congress. In 2003, USD (P&R) recognized the imprecision of the reported metrics from GSORTS and noted that its efforts to develop DRRS would allow enhancements to reported readiness data. As previously noted, the DRRS capability information that DOD is reporting to Congress covers a broader range of missions than the GSORTS information that was provided in the past. However, when comparing the DRRS and GSORTS assessments of units' capabilities to perform the missions for which the units were organized or designed, DRRS assessments are actually less precise than the GSORTS assessments. Specifically, within GSORTS, overall capability assessments are grouped into four categories based on four percentage ranges for the underlying data. For example, commanders compare on-hand and required levels of personnel and equipment. Within DRRS, mission essential task assessments are reported on a scale that includes only three ratings-- "yes", "no", and "qualified yes," which can include any assessments that fall between the two extremes. The law directing DOD to establish a new readiness reporting system also requires that the system measure readiness in an objective manner. GSORTS assessments of units' capabilities to execute the missions for which they were organized or designed are based on objective personnel and equipment data and training information that may include both objective and subjective measures. Furthermore, the overall capability assessment in GSORTS is based on an objective rule that calls for the overall assessment to be the same level as the lowest underlying resource or training data level. For example, if a unit reported the highest personnel level (P-1) and the lowest training level (T-4), the rules in the GSORTS guidance instruct the commander to rate the unit's overall capability at the C-4 level. Because GSORTS contains these objective measures and rules, it is easy to evaluate reported readiness to see if it aligns with established reporting criteria.[Footnote 33] Within DRRS, organizations rate their capabilities based on mission essential tasks. These mission essentials tasks have conditions and standards associated with them. The conditions specify the types of environments that units are likely to face as they execute the tasks, such as weather conditions and political or cultural factors. Standards describe what it means for the unit to successfully execute the task under specified conditions. For example, a unit may have to achieve a 90 percent success rate for measures associated with the task being assessed. In spite of these conditions and standards, DRRS mission assessments are often subjective rather than objective. In DRRS program guidance, DOD has defined mission essential tasks as tasks that are approved by the commander and that, based on mission analysis, are "absolutely necessary, indispensable, or critical to mission success." In prior briefings and reports to Congress, we have noted examples that highlight the subjective nature of DRRS mission assessments. For example, we noted that one commander used his professional judgment to decide that his command was "qualified" to execute a mission even though the preponderance of the "indispensable" tasks that supported that mission were rated as "no." In contrast, other commanders used their professional judgments to rate missions as "qualified" based on one or more "qualified" tasks among many "yes" tasks. DRRS Lacks the Complete Resource and Training Data and System Connectivity Some Users Need to Manage and Deploy Forces: DRRS does not have all of the resource, training, readiness data, and connectivity with the department's operations planning and execution system that the services, Joint Staff, and certain combatant commands need to manage and deploy forces. As a result, DRRS is not yet able to operate as the department's single readiness reporting system, as intended. The Secretary of Defense's and the Under Secretary of Defense's guidance documents recognize that DRRS needs to support the data requirements of multiple users. For example, the Secretary of Defense's June 25, 2004, memorandum directed USD (P&R) to develop DRRS to support the data requirements identified by the Chairman of the Joint Chiefs of Staff, the combatant commanders, the Secretaries of the military departments, and the Chief of the National Guard Bureau. [Footnote 34] Furthermore, the 2002 DRRS directive noted that DRRS was to build upon DOD's existing processes and readiness assessment tools and that ESORTS information (capability, resource, and training), where appropriate, is integrated into DOD's planning systems and processes. It also directed the Chairman of the Joint Chiefs of Staff to maintain the GSORTS database until key capabilities of DRRS become fully operational. Complete and Accurate Current and Historical Data: Officials with U.S. Joint Forces Command and U.S. Special Operations Command reported that historical data are needed to manage forces and provide users the ability to analyze readiness trends.[Footnote 35] Similarly, service officials stated a need for historical data so they can manage their forces and take action to address readiness issues. In 2005, USD (P&R) reported that unit resource data, including detailed inventory and authorization data on personnel, equipment, supply, and ordnance were available in DRRS. However, in response to a survey we conducted in December 2008, the services and certain combatant commands stated that necessary current and historical resource and training data were not available in DRRS. For example, officials from all four services responded that DRRS, at that time, contained less than half of their GSORTS resources and training data. In addition, officials from U.S. Joint Forces Command, U.S. Special Operations Command, the U.S. Strategic Command, and the U.S. Transportation Command all responded that historical resource data were not available in DRRS. We confirmed that this information was still not available when we concluded our review, and in April, 2009, the DIO said it was still working on this data availability issue. Furthermore, user organizations have reported inaccuracies in the data that are available in DRRS. Marine Corps and U.S. Special Operations Command officials stated that inconsistencies between DRRS data and the data in other readiness systems have caused them to adjudicate the inconsistencies by contacting their subordinate units directly. Army officials noted that searches of DRRS data can produce different results than searches in the Army's data systems. For example, they noted that a DRRS search for available personnel with a particular occupational specialty produced erroneously high results because DRRS did not employ the appropriate business rules when conducting the search. Specifically, DRRS did not apply a business rule to account for the fact that an individual person can possess multiple occupational specialty codes but can generally fill only one position at a time. DIO officials informed us that they intend to correct issues with the accuracy of data drawn from external databases. However, the current version of the DRRS Integrated Master Schedule indicates that the ability of DRRS to provide the capability to correctly search, manipulate, and display current and historical GSORTS and mission essential task data will not be complete until June 2010. As a result, the reliability of the DRRS data is likely to remain questionable and a number of DOD organizations will likely continue to rely on GSORTS and other sources of readiness data to support their decision making. Connections to Planning Systems: One important DRRS function is integration with DOD's planning systems. Specifically, the 2002 DRRS directive requires USD (P&R) to ensure that, where appropriate, ESORTS information (capability, resource, and training) is compatible and integrated into DOD's planning systems and processes. Global force management is one of the DOD planning processes that is to be integrated with DRRS. Global Force Management is a process to manage, assess, and display the worldwide disposition of U.S. forces, providing DOD with a global view of requirements and availability of forces to meet those requirements. The integration of DRRS with global force management planning processes is supposed to allow DOD to link force structure, resources, and capabilities data to support analyses, and thus help global force managers fill requests for forces or capabilities. Officials from the four organizations with primary responsibilities for providing forces (U.S. Joint Forces Command, U.S. Special Operations Command, U.S. Strategic Command, and U.S. Transportation Command) all stated that they are unable to effectively use DRRS to search for units that will meet requested capabilities. These commands also reported that DRRS does not currently contain the information and tools necessary to support global force management. For example, officials from U.S. Northern Command told us that when they used DRRS to search for available helicopters of a certain type, they found thousands, but when U.S. Joint Forces Command did the same DRRS search they found hundreds. The current version of the DRRS Integrated Master Schedule indicates that DRRS will not be able to fully support global force management until March 2011. As a result, these commands continue to rely on GSORTS rather than DRRS to support their planning and sourcing decisions. Conclusions: DRRS is not currently and consistently providing timely, objective, and accurate information, and it is not exactly clear where the department stands in its efforts to meet this expectation because system requirements remain in a state of flux, and the program office lacks disciplined program management and results information due to a long- standing lack of rigor in its approach to acquiring and deploying system capabilities. This situation can be attributed, in part, to long- standing limitations in the program office's focus on acquiring human capital skills needed to manage such a complex initiative. It can also be linked to many of years of limited program office oversight and accountability. Although program oversight has recently increased, oversight bodies have not had sufficient visibility into the program's many management weaknesses. DRRS is providing Congress and readiness users with additional mission and mission essential task capability data that were not available in GSORTS. However, after investing about 7 years and about $96.5 million in developing and implementing DRRS, the system's schedule has been extended, requirements are not stable, and the system still does not meet congressional and DOD requirements for a comprehensive readiness reporting system to assess readiness and help decision makers manage forces needed to conduct combat and contingency operations around the world. Given DRRS performance and management weaknesses, it is critical that immediate action be taken to put the program on track and position it for success. Without this action, it is likely that DRRS will cost more to develop and deploy than necessary and that DOD will not have a comprehensive reporting system that meets the needs of all the decision makers who rely on accurate, timely, and complete readiness information. Recommendations for Executive Action: To address the risks facing DOD in its acquisition and deployment of DRRS, and to increase the chances of DRRS meeting the needs of the DOD readiness community and Congress, we recommend that the Secretary of Defense direct the Deputy Secretary of Defense, as the Chair of the Defense Business Systems Management Committee, to reconsider the committee's recent approval of DRRS planned investment for fiscal years 2009 and 2010, and convene the Defense Business Systems Management Committee to review the program's past performance and the DIO's capability to manage and deliver DRRS going forward. To fully inform this Defense Business Systems Management Committee review, we also recommend that the Secretary direct the Deputy Secretary to have the Director of the Business Transformation Agency, using the appropriate team of functional and technical experts and the established risk assessment methodology, conduct a program risk assessment of DRRS, and to use the findings in our report and the risk assessment to decide how to redirect the program's structure, approach, funding, management, and oversight. In this regard, we recommend that the Secretary direct the Deputy Secretary to solicit the advice and recommendations of the DRRS Executive Committee. We also recommend that the Secretary, through the appropriate chain of command, take steps to ensure that the following occur: 1. DRRS requirements are effectively developed and managed with appropriate input from the services, Joint Staff, and combatant commanders, including (1) establishing an authoritative set of baseline requirements prior to further system design and development; (2) ensuring that the different levels of requirements and their associated design specifications and test cases are aligned with one another; and (3) developing and instituting a disciplined process for reviewing and accepting changes to the baseline requirements in light of estimated costs, benefits, and risk. 2. DRRS testing is effectively managed, including (1) developing test plans and procedures for each system increment test event that include a schedule of planned test activities, defined roles and responsibilities, test entrance and exit criteria, test defect management processes, and metrics for measuring test progress; (2) ensuring that all key test events are conducted on all DRRS increments; (3) capturing, analyzing, reporting, and resolving all test results and test defects of all developed and tested DRRS increments; and (4) establishing an effective test management structure that includes assigned test management roles and responsibilities, a designated test management lead and a supporting working group, and a reliable schedule of test events. 3. DRRS integrated master schedule is reliable, including ensuring that the schedule (1) captures all activities from the work breakdown structure, including the work to be performed and the resources to be used; (2) identifies the logical sequencing of all activities, including defining predecessor and successor activities; (3) reflects whether all required resources will be available when needed and their cost; (4) ensures that all activities and their duration are not summarized at a level that could mask critical elements; (5) achieves horizontal integration in the schedule by ensuring that all external interfaces (hand-offs) are established and interdependencies among activities are defined; (6) identifies float between activities by ensuring that the linkages among all activities are defined; (7) defines a critical path that runs continuously to the program's finish date; (8) incorporates the results of a schedule risk analysis to determine the level of confidence in meeting the program's activities and completion date; and (9) includes the actual start and completion dates of work activities performed so that the impact of deviations on downstream work can be proactively addressed. 4. The DRRS program office is staffed on the basis of a human capital strategy that is grounded in an assessment of the core competencies and essential knowledge, skills, and abilities needed to perform key DRRS program management functions, an inventory of the program office's existing workforce capabilities, and an analysis of the gap between the assessed needs and the existing capabilities. 5. DRRS is developed and implemented in a manner that does not increase the reporting burden on units and addresses the timeliness, precision, and objectivity of metrics that are reported to Congress. To ensure that these and other DRRS program management improvements and activities are effectively implemented and that any additional funds for DRRS implementation are used effectively and efficiently, we further recommend that the Secretary direct the Deputy Secretary to ensure that both the Human Resources Management Investment Review Board and the DRRS Executive Committee conduct frequent oversight activities of the DRRS program, and report any significant issues to the Deputy Secretary. Agency Comments and Our Evaluation: In written comments on a draft of this report, signed by the Deputy Under Secretary of Defense (Military Personnel Policy) performing the duties of the Under Secretary of Defense (Personnel and Readiness), DOD stated that the report is flawed in its assessment of DRRS, noting that DRRS is a net-centric application that provides broad and detailed visibility on readiness issues, and that achieving data sharing across the DOD enterprise was groundbreaking work fraught with barriers and obstacles, many of which have now been overcome. In addition, DOD stated that it was disappointed that the report did not address cultural impediments that it considers to be the root cause of many of the issues cited in the report and of many previous congressional concerns on readiness reporting. DOD further stated that the report instead focuses on past acquisition process and software development problems that it believes have now been remedied According to the department, this focus, coupled with inaccurate and misleading factual information included in the report, led us to develop an incomplete picture of the program. Notwithstanding these comments, DOD agreed with two of our recommendations and partially agreed with a third. However, it disagreed with the remaining five recommendations, and provided comments relative to each recommendation. DOD's comments are reprinted in their entirety in appendix III. In summary, we do not agree with DOD's overall characterization of our report or the positions it has taken in disagreeing with five of our recommendations, finding them to be inconsistent with existing guidance and recognized best practices on system acquisition management, unsupported by verifiable evidence, and in conflict with the facts detailed in our report. Further, we recognize that developing DRRS is a significant and challenging undertaking that involves cultural impediments. As a result, our report explicitly focuses on the kind of program management rigor and disciplines needed to address such impediments and successfully acquire complex systems, including effective requirements development and management and executive oversight. We also disagree that our report focuses on past issues and problems. Rather, it provides evidence that demonstrates a long- standing and current pattern of system acquisition and program oversight weaknesses that existed when we concluded our audit work and that DOD has not provided any evidence to demonstrate has been corrected. In addition, we would emphasize that we defined our objectives, scope, and methodology, and executed our audit work in accordance with generally accepted government auditing standards, which require us to subject our approach as well as the results of our audit work to proven quality assurance checks and evidence standards that require us to seek documentation rather than relying solely on testimonial evidence. While we support any departmental efforts, whether completed or ongoing, that would address the significant problems cited in our report, we note that DOD, in its comments, did not specifically cite what these efforts are or provide documentation to support that they have either been completed or are ongoing. Therefore, we stand by our findings and recommendations. Moreover, we are concerned that in light of the program's significant and long-standing management weaknesses, the department's decision not to pursue recommendations aimed at corrective actions for five of our eight recommendations will further increase risk to achieving program success, and is not in the best interests of the military readiness community or the U.S. taxpayer. Accordingly, we encourage the department to reconsider its position when it submits its written statement of the actions taken on our recommendations to the Senate Committee on Homeland Security and Governmental Affairs and the House Committee on Oversight and Government Reform , as well has the House and Senate Committees on Appropriations, as required under 31 U.S.C. 720. DOD's specific comments on each recommendation, along with our responses to its comments follow. * The department did not agree with our recommendation for the Deputy Secretary of Defense, as the Chair of the Defense Business Systems Management Committee, to reconsider the committee's recent approval of DRRS planned investment for fiscal years 2009 and 2010, and to convene the Defense Business Systems Management Committee to review the program's past performance and the DIO's capability to manage and deliver DRRS going forward in deciding how best to proceed. In this regard, DOD stated that the Investment Review Board certification and Defense Business Systems Management Committee approval were granted in compliance with the established processes. It also added that oversight of the specific issues identified in this report are the responsibility of the DRRS Executive Committee, which it stated has and will continue to provide appropriate governance for this effort. It also stated that USD (P&R) will ensure that the program is compliant with all acquisition requirements prior to submission for further certifications. We do not question whether the Investment Review Board certification and Defense Business Systems Management Committee approval were provided in accordance with established processes, as this is not relevant to our recommendation. Rather, our point is that the Investment Review Board and Defense Business Systems Management Committee were provided, and thus based their respective decisions, on erroneous and incomplete information about DRRS progress, management weaknesses, and risks. Moreover, neither the Investment Review Board nor the Defense Business Systems Management Committee were informed about the findings in our report, even though we shared each of them with the DRRS program director and other DIO officials prior to both the Investment Review Board and the Defense Business Systems Management Committee deliberations. Therefore, while the Investment Review Board certification and the Defense Business Systems Management Committee approval were granted in accordance with established processes, they were not based on a full disclosure of facts. Moreover, while we support DOD's comment that it will ensure that the program is in compliance with all acquisition requirements prior to further certifications, nothing precludes the board or the committee from reconsidering their respective decisions in light of our report. With respect to DOD's comment that the DRRS Executive Committee has and will continue to provide appropriate governance for this effort, we do not disagree that the DRRS Executive Committee has an oversight role. However, the DRRS Executive Committee should not be solely responsible for oversight of the specific issues in our report. Both the Investment Review Board and the Defense Business Systems Management Committee provide additional layers of oversight pursuant to law[Footnote 36] and DOD policy.[Footnote 37] Accordingly, we stand by our recommendation as it appropriately seeks to have the Investment Review Board and Defense Business Systems Management Committee, in collaboration with the DRRS Executive Committee, act in a manner that is consistent with their respective roles as defined in law. In doing so, our intent is to promote accountability for DRRS progress and performance, and prompt action to address the many risks facing the program. * The department agreed with our recommendation for the Deputy Secretary of Defense, as the chair of the Defense Business Systems Management Committee, to have the Business Transformation Agency conduct a risk assessment of DRRS, and with the advice and recommendation of the DRRS Executive Committee, to use the results of this assessment and the findings in our report to decide how to redirect the program. In this regard, the department stated that this assessment will be complete by the middle of fiscal year 2010. * The department did not agree with our recommendation for ensuring that DRRS requirements are effectively developed and managed. In this regard, it stated that the program has an authoritative set of baseline requirements established with an effective governance process for overseeing the requirements management process, to include biweekly reviews as part of the DRRS configuration control process. We do not agree. At the time we concluded our work, DRRS requirements were not stable, as evidenced by the fact that an additional 530 requirements had been identified that the DIO was still in the process of reviewing and had yet to reach a position on their inclusion, or process them through the DRRS change control governance process. Moreover, when we concluded our work, this change control process had yet to be approved by the DRRS governance structure. As we state in our report, the introduction of such a large number of requirements provided a compelling basis for concluding that requirements had yet to progress to the point that they could be considered sufficiently complete and correct to provide a stable baseline. Our recommendation also noted that the Secretary should take steps to ensure that the different levels of requirements be aligned with one another. DOD's comments did not address this aspect of our recommendation. * The department did not agree with our recommendation for ensuring that DRRS testing is effectively managed. In this regard, it stated that DRRS testing is already in place and performing effectively, and stated, among other things, that (1) the DIO goes through a rigorous testing regimen that includes documenting test plans with user test cases for each incremental release to include utilizing system integration, acceptance, interoperability, and operational testing; (2) user test cases and functionality are validated by designated testers independent of the developers prior to a deployment; and (3) for interoperability testing the DIO has a designated test director and the Joint Interoperability Test Command (JITC) is the designated interoperability and operational test activity. We do not agree. As our report concludes, DRRS testing has not been effectively managed because it has not followed a rigorous testing regimen that includes documented test plans, cases, and procedures. To support this conclusion, our report cites numerous examples of test planning and execution weaknesses, as well as the DIO's repeated inability to demonstrate through requisite documentation that the testing performed on DRRS has been adequate. Our report shows that test events for already acquired, as well as currently deployed and operating, DRRS releases and subreleases were not based on well-defined plans and DOD had not filled its testing director vacancy. Further, our report shows that test events were not fully executed in accordance with plans that did exist, or executed in a verifiable manner, or both. For example, although increments of DRRS functionality had been put into production, the program had no documentation (e.g., test procedures, test cases, test results) to show that the program office had performed system integration testing, system acceptance testing, or operational testing on any DRRS release or subrelease, even though the DIO's test strategy stated that such tests were to be performed before system capabilities became operational. Moreover, evidence showed that the results of all executed test events had not been captured and used to ensure that problems discovered were disclosed to decision makers, and ultimately corrected. With respect to DOD's comments that JITC is the designated lead for interoperability and operational testing, our report recognizes that JITC is to conduct both interoperability and operational testing before the system is deployed and put into production (i.e., used operationally). However, during the course of our audit, the DIO could not produce any evidence to show that interoperability and operational testing of all operating system increments had been conducted. * The department did not agree with our recommendation for ensuring that the DRRS integrated master schedule is reliable. In this regard, it stated that a process is already in place to ensure that the schedule is current, reliable, and meets all the criteria outlined in the recommendation. We do not agree. As our report states, an integrated master schedule for DRRS did not exist until November 2008, which was 2 months after we first requested one. Moreover, following our feedback to the DIO on limitations in this initial version, a revised integrated master schedule was developed in January 2009, which was also not reliable. Subsequently, a revised integrated master schedule was developed in April 2009. However, as we detail in our report, that version still contained significant weaknesses. For example, it did not establish a critical path for all key activities or include a schedule risk analysis, and was not being updated using logic and durations to determine the dates for all key activities. These practices are fundamental to producing a sufficiently reliable schedule baseline that can be used to measure progress and forecast slippages. In addition, the schedule introduced considerable concurrency across key activities and events for several modules, which introduces increased risk. Therefore, we stand by our recommendation. * The department partially agreed with our recommendation for ensuring that it has an effective human capital strategy. In this regard, it stated that actions are underway to add more full-time civilian support to the DIO, and that plans exist to convert some contractor to civilian billets during the 2010/2011 time frame. We support the department's actions and plans described in its comments to address the DIO human capital management limitations discussed in our report, but would note that they do not go far enough to systematically ensure that the program has the right people with the right skills to manage the program in both the near term and the long term. To accomplish this, the department needs to adopt the kind of strategic and proactive approach to DRRS workforce management that our report describes and our recommendation embodies. As our evaluations and research show, failure to do so increases the risk that the program office will not have the people it needs to effectively and efficiently manage DRRS. Therefore, we believe that the department needs to fully implement our recommendation. * The department did not agree with our recommendation to take steps to ensure that DRRS is developed and implemented in a manner that does not increase the reporting burden on units and addresses the timeliness, precision, and objectivity of metrics that are reported to Congress. In this regard, it stated that one of the primary tenets of DRRS has been to reduce reporting requirements on the war fighter. It also stated that DRRS is already using state-of-the-art technology to ensure that near-real-time data are available for the war fighters. Finally it stated that the DRRS governance structure that is currently in place ensures that DRRS development does not deviate from these core principles. While we recognize that a goal of DRRS is to reduce a reporting burden on the war fighter, we disagree with the department's position because the system has not yet achieved this goal. As our report states, while the DIO has developed the software for users to enter mission essential task data into DRRS, the DIO has been unsuccessful in attempts to develop a tool that would allow Air Force and Marine Corps users to enter readiness data to meet all of their readiness reporting requirements through DRRS. As a result, rather than reducing the burden on reporting units, DRRS actually increased the burden on Air Force and Marine Corps units because they were required to report readiness information in both DRRS and GSORTS. Without a viable tool for inputting data, DRRS is not fully integrated with GSORTS or with the service readiness reporting systems and it is not capable of replacing those systems since it does not capture the required data that are contained in those systems. In addition, the DRRS readiness data that are currently reported to Congress are not near-real-time data. Specifically, the periodicity for DRRS capability assessments is the same as the legacy GSORTS system's readiness reports--monthly or when an event occurs that changes a unit's overall readiness. Furthermore, our report shows that DRRS mission assessments are often subjective and imprecise because they are reported on a scale that includes only three ratings--"yes," "no," and "qualified yes," which can include any assessments that fall between the two extremes. Therefore, because additional actions are still needed to reduce reporting burdens and improve the timeliness, precision, and objectivity of the DRRS data that are reported to Congress, we stand by our recommendation. * The department agreed with our recommendation for ensuring that both the Human Resources Management Investment Review Board and the DRRS Executive Committee conduct frequent oversight activities of the DRRS program and report any significant issues to the Deputy Secretary of Defense. In this regard, the department stated that the USD (P&R) component acquisition executive is working with the program to ensure that it becomes fully compliant with all acquisition requirements. In addition, it stated that the acquisition executive will certify to the Human Resources Investment Review Board and the Deputy Chief Management Officer of compliance prior to submission of future certification requests. Further, it stated that the current DRRS governance process will provide sustained functional oversight of the program and that issues that arise in any of these areas will be elevated for review, as appropriate. We believe these are positive steps. We are sending copies of this report to the appropriate congressional committees; the Secretary of Defense; and the Director, Office of Management and Budget. The report will also be available at no charge on the GAO Web site at [hyperlink, http://www.gao.gov]. If you or your staffs have questions about this report, please contact us at pickups@gao.gov or hiter@gao.gov or at our respective phone numbers, (202) 512-9619 and (202) 512-3439. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made key contributions to this report are listed in appendix IV. Signed by: Sharon L. Pickup: Director, Defense Capabilities and Management: Signed by: Randolph C. Hite: Director, Information Technology Architecture and Systems Issues: [End of section] Appendix I: Objectives, Scope, and Methodology: Our objectives were to assess the extent to which (1) the Department of Defense (DOD) has effectively managed and overseen DRRS acquisition and deployment, and (2) features of the Defense Readiness Reporting System (DRRS) have been implemented and are consistent with legislative requirements and DOD guidance. We did not evaluate the department's overall ability to assess the readiness of its forces or the extent that data contained in any of its readiness reporting systems, including DRRS and the Global Status of Resources and Training System (GSORTS), reflect capabilities, deficiencies, vulnerabilities, or performance issues. Our review focused on acquisition and program management issues, such as requirements management, schedule, and human capital requirements; the current usage of DRRS; and the extent to which DRRS' features address legislative requirements and DOD guidance. To determine the extent to which the DRRS acquisition and deployment has been effectively managed and overseen, we focused on the following acquisition management areas: (1) requirements development and management, (2) test planning and execution, (3) DRRS schedule reliability, and (4) human capital planning. In doing so, we analyzed a range of program documentation, such as high-level and detailed-level requirements documentation, test plans and reports, the current DRRS schedule, and program management documentation and interviewed cognizant program and contractor officials. * To determine the extent to which the program had effectively implemented requirements development and management, we reviewed relevant program documentation, such as the concept of operations document, capability requirements document, software requirements document, requirements traceability matrix, configuration management plan, and the program management plan, as well as minutes of change control board meetings, and evaluated them against relevant guidance.[Footnote 38] Moreover, we reviewed briefing slides from meetings of DRRS oversight bodies in order to identify concerns about DRRS expressed by representatives from the DRRS community of users, as well as the efforts by the Joint Staff (at the direction of DRRS Executive Committee) to identify and address any gaps identified by users in the development of DRRS requirements. To determine the extent to which the program has maintained traceability backward to high-level business operation requirements and system requirements, and forward to system design specifications and test plans, we randomly selected 60 program requirements and traced them both backward and forward. This sample was designed with a 5 percent tolerable error rate at the 95 percent level of confidence so that, if we found zero problems in our sample, we could conclude statistically that the error rate was less than 5 percent. In addition, we interviewed program and development contractor officials to discuss the requirements development and management process. * To determine if the DRRS Implementation Office (DIO) is effectively managing DRRS testing, we reviewed relevant documentation, such as the DRRS Test and Evaluation Master Plans and test reports and compared them to DOD and other relevant guidance.[Footnote 39] Further, we reviewed developmental test plans and procedures for each release/ subrelease that to date has either been developed or fielded and compared them with best practices to determine whether test activities had been adequately documented. We also examined test results and reports for the already acquired, as well as currently deployed and operating, DRRS releases and subreleases and compared them against plans to determine whether they had been executed in accordance with plans. Moreover, we reviewed key test documentation, such as the Software Version Descriptions, and compared them against relevant guidance to determine whether defect data were being captured, analyzed, prioritized, and reported. We also interviewed program and contractor officials to gain clarity beyond what was included in the program documentation, including the Defense Information Systems Agency's Joint Interoperability Test Center in order to determine the results of their efforts to independently test DRRS interoperability. In addition, to determine the extent to which the program had effectively tested its system requirements, we observed the DIO's efforts to demonstrate the traceability of 60 program requirements to test cases and results. This sample was designed with a 5 percent tolerable error rate at the 95 percent level of confidence so that, if we found zero problems in our sample, we could conclude statistically that the error rate was less than 5 percent. * To determine the extent to which the program's schedule reflects key estimating practices that are fundamental to having a reliable schedule, we reviewed the DRRS integrated master schedules and schedule estimates and compared them with relevant guidance.[Footnote 40] We also used schedule analysis software tools to determine whether the latest schedule included key information, such as the activities critical to on-time completion of DRRS, a logical sequence of activities, and evidence that the schedule was periodically updated. We also reviewed the schedule to determine the time frames for completing key program activities and to determine any changes to key milestones. In addition, we shared the results of our findings with program and contractor officials and asked for clarifications. We then reviewed the revised schedule, prepared in response to the weaknesses we found, and compared it with relevant guidance. * To evaluate whether DOD is adequately providing for the DRRS program's human capital needs, we compared the program's efforts against relevant criteria and guidance, including our own framework for strategic human capital management.[Footnote 41] In doing so, we reviewed key program documentation, such as the program management plan and the DIO organizational structure to determine whether it reflected key acquisition functions and identified whether these functions were being performed by government or contractor officials. We interviewed key officials to discuss workforce analysis and human capital planning efforts. * To determine the level of oversight and governance available to the DRRS community of users, we attended relevant meetings, met with officials responsible for program certification, and reviewed relevant guidance and program documentation. Specifically, we attended Battle Staff meetings and analyzed briefing slides and meeting minutes from the DRRS Executive Committee, General and Flag Officer's Steering Committee, and Battle Staff meetings--the main DRRS governance bodies. In addition, we reviewed key DRRS certification and approval documentation provided by the Human Resources Management Investment Review Board, such as economic viability analyses and the business system certification dashboard and met with Investment Review Board officials to determine the basis for certifying and approving DRRS. To determine the extent to which the features of DRRS have been implemented and are consistent with legislative requirements and DOD guidance, we first examined the language of Section 117 of Title 10 of the United States Code, which directs the Secretary of Defense to establish a comprehensive readiness reporting system. We identified criteria for this system in DOD's directive formally establishing the system.[Footnote 42] We evaluated the system by conducting interviews-- see table 2 below for a list of these organizations--and receiving system demonstrations from members of the readiness community to determine how they used DRRS and how their usage compared with the criteria established for the system. We also conducted content and data analysis of system documents and briefing packages provided by the DIO and Joint Staff. In order to capture the broadest amount of data about the system we conducted a survey of readiness offices at all of the service headquarters, combatant commands, and the National Guard Bureau regarding how DRRS was currently being used and the types and amount of data available in the system.[Footnote 43] In addition, to track the development of DRRS capabilities, we attended Battle Staff meetings and analyzed documentation from meetings of all the DRRS governance bodies. We also searched for and extracted information from DRRS in order to support other GAO ongoing readiness reviews. While our usage of the system was not intended as a formal test of the system, our general observations concerning system functionality and the range of available data were consistent with the observations of most other users, which were noted in our survey. Table 2: Organizations Interviewed during Our Review: Office of the Under Secretary of Defense (Personnel & Readiness), Arlington, Virginia: U.S Army: * Office of the Deputy Chief of Staff (Operations), Arlington, Virginia. * U.S. Army Forces Command, Fort McPherson, Georgia. * U.S. Army Pacific, Fort Shafter, Hawaii. * Installation Management Command, Pacific Region Office, Fort Shafter, Hawaii. U.S. Navy: * Headquarters Navy, Integrated Fleet Readiness, Arlington, Virginia. * Fleet Forces Command, Norfolk, Virginia. * U.S. Pacific Fleet, Pearl Harbor, Hawaii. U.S. Air Force: * Headquarters, U.S. Air Force (Readiness), Arlington, Virginia. * Air Combat Command, Langley Air Force Base, Virginia. * U.S. Pacific Air Forces, Readiness Branch, Hickam Air Force Base, Hawaii. * 13th Air Force, Hickam Air Force Base, Hawaii. U.S. Marine Corps: * Headquarters Marine Corps (Readiness), Arlington, Virginia. * Marine Forces Command, Norfolk, Virginia. * Marine Forces Pacific, Camp Smith, Hawaii. * Combat Logistics Battalion 3, Kaneohe Bay, Hawaii. U.S. Central Command, MacDill Air Force Base, Florida. U.S. Joint Forces Command, Norfolk, Virginia. U.S. Northern Command, Colorado Springs, Colorado. U.S. Pacific Command, Camp Smith, Hawaii. U.S. Special Operations Command, MacDill Air Force Base, Florida. Source: GAO. [End of table] We conducted our work from April 2008 through August 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix II: Detailed Analysis of DRRS Acquisition and Deployment Management and Oversight: Our research and evaluations of information technology programs have shown that the ability to deliver promised system capabilities and benefits on time and within budget largely depends on the extent to which key program management disciplines are employed by an adequately staffed program management office. Among other things, these disciplines include a number of practices associated with effectively developing and managing system requirements, adequately testing system capabilities, and reliably scheduling the work to be performed. They also include proactively managing the program office's human capital needs, and promoting program office accountability through effective program oversight. Department of Defense (DOD) acquisition policies and guidance, along with other relevant guidance, recognize the importance of these management and oversight disciplines.[Footnote 44] As we have previously reported, not employing these and other program management disciplines increases the risk that system acquisitions will not perform as intended and require expensive and time-consuming rework. Defense Readiness Reporting System (DRRS) acquisition and deployment has for years not been effectively managed in accordance with these key program management disciplines that are recognized in DOD and other relevant guidance, and are fundamental to delivering a system that performs as intended on time and within budget. DRRS Requirements Have Not Been Effectively Developed and Managed: Well-defined and well-managed requirements are a cornerstone of effective system development and acquisition. According to recognized guidance, documenting and implementing a disciplined process for developing and managing requirements can help to reduce the risks of producing a system that is not adequately tested, does not meet user needs, and does not perform as intended.[Footnote 45] Effective requirements development and management includes, among other things, (1) effectively eliciting user needs early and continuously in the system life-cycle process, (2) establishing a stable baseline set of requirements and placing this baseline under configuration management, (3) ensuring that system requirements are traceable backward to higher level business or operational requirements (e.g., concept of operations) and forward to system design documents (e.g., software requirements specification) and test plans, and (4) controlling changes to baseline requirements. DRRS requirements have not been effectively developed and managed. Specifically, (1) key users have only recently become engaged in developing requirements, (2) requirements have been experiencing considerable change and are not yet stable, (3) different levels of requirements and related test cases have not been aligned with one another, and (4) changes to requirements have not been effectively controlled. As a result, efforts to develop and deliver initial DRRS capabilities have taken longer than envisioned and these capabilities have not lived up to the readiness communities' expectations. These failures increase the risk of future DRRS capabilities not meeting expectations and ensure that expensive and time-consuming system rework will be necessary. Recent Actions Have Been Taken to Address Limited User Involvement in Developing and Managing Requirements: One of the leading practices associated with effective requirements development is engaging system users early and continuously in the process of defining requirements. As we have previously reported, assessing user needs early in the process increases the probability of success in defining, designing, and delivering a system that meets user needs and performs as intended.[Footnote 46] To the DRRS Implementation Office's (DIO) credit, the October 2008 DRRS Risk Management Plan recognizes this by stating that the success of DRRS depends on participation and support from the broad readiness community, which includes combatant commands, Joint Staff, and the military services. However, until recently, key users were not effectively engaged in DRRS requirements development and management, although reasons vary why they have not. Specifically, DIO officials told us that beginning in 2002, they reached out to all user groups-- combatant commands, Joint Staff, and the military services--in defining requirements. For example, they cited a July 2002 memorandum issued by the Office of the Under Secretary of Defense for Personnel and Readiness (USD P&R) that encouraged the Director of the Joint Chiefs of Staff, Deputy Commanders of the Combat Commands, Service Operations Deputies, and Directors of Defense Agencies to actively support the DRRS effort by ensuring that their organizations are represented at Battle Staff meetings.[Footnote 47] However, these officials told us that the military services and Joint Staff chose not to participate. In contrast, officials from these user groups told us their involvement had been limited by what they characterized as difficulties in submitting requirements through the DRRS governance boards that were in place at that time. For example, an official from the Joint Forces Command said that the Forces Battle Staff governance board did not meet for about a year between 2005 and 2006. Further, the official said that the meetings that were held did not offer users the opportunity to discuss their concerns or influence the requirements process. Similarly, an official from the Marine Corps cited a lack of clear and transparent communication from the DIO as a significant impediment. Notwithstanding this lack of stakeholder involvement in setting requirements, the Office of USD (P&R) developed and issued a DRRS concept of operations in 2004, which DIO officials told us was based on input from the combatant commands, relevant departmental directives, [Footnote 48] and DRRS governance boards (e.g., Battle Staff). In our view, this document provided a high-level overview of proposed DRRS capabilities from which more detailed requirements could be derived. However, the concept of operations was not approved by all key players in the readiness community. Specifically, DIO officials stated that the document had not been approved by the military services and the Joint Staff. According to these officials, the reason for not seeking all stakeholders' approval, and the decision to begin developing more detailed requirements in the absence of an approved concept of operations, was that the 2002 DRRS DOD directive provided a sufficient basis to begin developing and deploying what they anticipated being the initial versions of DRRS.[Footnote 49] In 2008, after 6 years of effort to define DRRS requirements and develop and deploy system capabilities, the Joint Staff, at the direction of the DRRS Executive Committee, conducted an analysis of DRRS capabilities--referred to as the "capabilities gap analysis." To the Joint Staff's credit, this analysis has appropriately focused on soliciting comments from the entire readiness community and on identifying any gaps between the DRRS requirements and the needs of this community. As will be discussed in the next section, this analysis resulted in 530 additional user requirements. The extended period of limited involvement by key DRRS users in defining a concept of operations and related capabilities and requirements has impeded efforts to develop a clear understanding of DRRS expectations, constraints, and limitations, which, in turn, has contributed to significant delays in providing the readiness community with needed system support. While the recent Joint Staff action to engage the entire DRRS user community is a positive step towards overcoming this long-standing problem, it remains to be seen whether this engagement will produce agreement and commitment across the entire readiness user community around DRRS requirements. DRRS Requirements Are Not Stable: As previously noted, establishing an authoritative set of baseline requirements prior to system design and development is necessary to design, develop, and deliver a system that performs as intended and meets users' operational needs.[Footnote 50] In general, a baselined set of requirements are those that are defined to the point that extensive changes are not expected, placed under configuration management, and formally controlled.[Footnote 51] DRRS requirements are currently in a state of flux. Specifically, the fact that 530 new user requirements have recently been identified means that the suite of requirements documentation associated with the system will need to be changed and thus are not stable. To illustrate, program officials told us that, as of late February 2009, these 530 new requirements had not been fully evaluated by the DIO and DRRS governance boards and thus not yet approved. As a result, their impact on the program is not clear. Compounding this instability in the DRRS requirements is the fact that additional changes are envisioned. According to program officials, the changes resulting from the gap analysis and reflected in the latest version of the DRRS concept of operations, which was approved by the DRRS Executive Committee in January 2009, have yet to be reflected in other requirements documents, such as the detailed system requirements. Although defining and developing requirements is inherently an iterative process, having a baseline set of requirements that are stable is a prerequisite to effective and efficient development of an operationally capable and suitable system. Without them, the DIO will not be able to deliver a system that meets user needs on time, and it is unlikely that future development and deployment efforts will produce better results. Alignment among Requirements and Related System Design and Testing Artifacts Has Not Been Demonstrated: One of the leading practices associated with developing and managing requirements is maintaining bidirectional traceability from high-level operational requirements (e.g., concept of operations and functional requirements) through detailed lower-level requirements and design documents (e.g., software requirements specification) to test cases. Such traceability is often accomplished through the use of a requirements traceability matrix, which serves as a crosswalk between different levels of related requirements, design, and testing documentation. The DRRS program management plan recognizes the importance of traceability, stating that requirements are to be documented and linked to acceptance tests, scripts, and criteria. Despite the importance of traceability, DIO officials could not demonstrate that requirements and related system design and testing artifacts are properly aligned. Specifically, we attempted on three separate occasions to verify the traceability of system requirements backward to higher-level requirements and forward to lower-level software specifications and test cases, and each time we found that traceability did not exist. Each attempt is discussed here: * In November 2008, our analysis of the requirements traceability matrix and the software requirements specification showed significant inconsistencies. For example, the traceability matrix did not include 29 requirements that were included in the software requirements specification. As a result, we did not have an authoritative set of requirements to use to generate a random sample of requirements to trace. Program officials attributed the inconsistencies to delays in updating all the documents to reflect the aforementioned capability gap analysis. They also stated that these documents would be updated by December 2008. * In December 2008, we used an updated requirements traceability matrix to generate a randomized sample of 60 software requirements specifications and observed a DIO demonstration of the traceability of this sample. However, DIO officials were unable to demonstrate for us that these specifications could be traced backward to higher-level requirements and forward to test cases. Specifically, attempts to trace the first 21 requirements forward to test cases failed, and DIO officials stated that they could not trace the 60 requirements backward because the associated requirements documents were still being updated. According to the officials, 11 of the 21 could not be traced forward because these were implemented prior to 2006 and the related test information was not maintained by the program office but rather was at the development contractor's site. They added that the remaining 10 either lacked test case information or test results. * In February 2009, we used an updated DIO-provided requirements traceability matrix, a capabilities requirement document, and software requirements specification to generate another randomized sample of 60 detailed specifications. We then observed the development contractor's demonstration of traceability using the contractor's requirements management tool. Because of time constraints, this demonstration focused on 46 of the 60 requirements, and it showed that adequate traceability still did not exist. Specifically, 38 of the 46 could not be traced backward to higher-level requirements or forward to test cases. This means that about 83 percent of the DRRS specifications (95 percent degree of confidence of being between 72 and 91 percent) were not traceable. Of the 38, 14 did not trace because of incomplete traceability documentation; 5 due to inconsistent traceability documentation; 3 due to requirements not being resident in the tracking tool; and 16 due to no actual development work being started. In addition, none of the 46 requirements were traceable to the January 2009 concept of operations. According to contractor officials, this is because the newly developed capability requirements document is considered to be a superset of the concept of operations, and thus traceability to this new document is their focus. However, they were unable to demonstrate traceability to the requirements in either the capability requirements document or the concept of operations. Further, we also found numerous inconsistencies among the capabilities requirements document, software requirements specification, and the requirements traceability matrix. For example, 15 capabilities requirements listed on the traceability matrix were not listed in the capabilities requirements document, but were listed in the updated software requirements specification, dated February 2009. Further, one requirement listed in the traceability matrix was not listed in either of these documents. One possible reason for these inconsistencies is that the traceability matrix was prepared manually, rather than being automatically generated from the tool, which would increase the probability of these and other discrepancies caused by human error. Another reason cited by program officials is that test results that occurred prior to October 2006 had yet to be fully recorded in the contractor's tracking tool. DIO and contractor officials attributed the absence of adequate requirements traceability to the ongoing instability in requirements and magnitude of the effort to update the chain of preexisting and new requirements documentation. They added that they expect traceability to improve as requirements become more stable and the documentation is updated. Regardless, the DIO has and continues to invest in the development of DRRS in the absence of requirements traceability. Without traceable requirements, the DIO does not have a sufficient basis for knowing that the scope of the design, development, and testing efforts will produce a system solution on time and on budget and that will meet users' operational needs and perform as intended. As a result, the risk is significant that expensive and time-consuming system rework will be required. Changes to Requirements Are Not Being Effectively Controlled: Adopting a disciplined process for reviewing and accepting changes to an approved and authoritative baseline set of requirements in light of the estimated costs, benefits, and risk of each proposed change is a recognized best practice. Elements of a disciplined process include: (1) formally documenting a requirements change process; (2) adopting objective criteria for considering proposed changes, such as estimated cost or schedule impact; and (3) rigorously adhering to the documented change control process. Since the inception of the program in 2002, DRRS has been developed and managed without a formally documented and approved process for managing changes to system requirements. Further, while requirements management and change control plans were developed in 2006 by the DRRS software development contractor, according to DIO officials, the plans were not adequate. For example, the plans did not detail how DRRS user requirements were collected or how objective factors, such as cost, impacted development decisions. To address these problems, the Joint Staff developed what it referred to as a conceptual requirements change control process in February 2008, which was to serve as a basis for the DIO to develop more detailed plans that could be implemented. Eleven months later, in January 2009, the DIO drafted more detailed plans--a DRRS requirements management plan and a DRRS requirements configuration management plan, the latter of which the DIO updated in March 2009. Specifically, the draft plans call for new DRRS requirements to be collected using an online tool and reviewed by the DIO to determine whether the requirement constitutes a major change to DRRS. Once approved, the DIO and the contractor are to provide the Battle Staff with a formatted report specifying the anticipated benefit of each new requirement and an initial analysis of the cost and performance impact. The Battle Staff then is to prioritize the requirement based on the DIO's impact analysis. If the issue cannot be resolved by the Battle Staff, it is to be elevated to the senior oversight bodies (i.e., the General Officer's Steering Committee and the DRRS Executive Committee). After a requirement has been approved, the software developer may prepare a more detailed "customer acceptance document" that analyzes the potential cost, schedule, and quality impact to DRRS objectives, which is then to be reviewed by the DIO at subsequent Change Control Board meetings. However, according to the user community and the DIO Director, the revised plans have not been submitted for review and approval to the DRRS community. Specifically, they stated that only a proposed process flow diagram was briefed at the Battle Staff, and according to them, the change control process was still being evaluated. Moreover, the DIO has yet to implement key aspects of its draft plans. For example, the DRRS Chief Engineer stated that until recently, the DIO had continued to accept changes to DRRS requirements that were submitted outside of the designated online tool. In addition, the reports that the Battle Staff are to use in making their requirement change determination do not include the anticipated benefit and estimated cost or schedule impact of new requirements. Rather, these reports only include the estimated number of hours necessary to complete work on a proposed requirement. Moreover, contractor officials and users from the Special Operations Command told us that cost or schedule impacts have rarely been discussed at the Battle Staff or Change Control Board meetings. Our analysis of minutes from change control meetings confirmed this. Furthermore, the DRRS Chief Engineer stated that the customer acceptance documents have only recently been used. Until the DIO effectively controls requirements changes, it increases the risk of needed DRRS capabilities taking longer and costing more to deliver than necessary. DRRS Testing Has Not Been Adequately Performed and Key Test Management Structures and Controls Have Not Been Established: Effective system testing is essential to successfully developing and deploying systems like DRRS. According to DOD and other relevant guidance, system testing should be progressive, meaning that it should consist of a series of test events that first focus on the performance of individual system components, then on the performance of integrated system components, followed by system-level tests that focus on whether the system (or major system increments) are acceptable, interoperable with related systems, and operationally suitable to users. [Footnote 52] For this series of related test events to be conducted effectively, (1) each test event needs to be executed in accordance with well- defined test plans, (2) the results of each test event need to be captured and used to ensure that problems discovered are disclosed and corrected, and (3) all test events need to be governed by a well- defined test management structure. Despite acquiring and partially deploying a subset of DRRS increments, the DIO cannot demonstrate that it has adequately tested any of these system increments, referred to as system releases and subreleases. Specifically, (1) the test events for already acquired, as well as currently deployed and operating, DRRS releases and subreleases were not based on well-defined plans, and test events have not been fully executed in accordance with plans or executed in a verifiable manner, or both; (2) the results of executed test events have not been captured and used to ensure that problems discovered were disclosed to decision makers and ultimately corrected; and (3) the DIO has not established an effective test management structure to include, for example, a clear assignment of test management roles and responsibilities, or a reliable schedule of planned test events. Compounding this absence of test management structures and controls is the fact that the DIO has yet to define how the series of system releases and subreleases relate to its recent restructuring of DRRS increments into a series of 10 modules. Collectively, this means that it is unlikely that already developed and deployed DRRS capabilities can perform as intended and meet user operational needs. Equally doubtful are the chances that the DIO can adequately ensure that yet-to-be developed DRRS capabilities will meet expectations. DIO Has Not Adequately Planned and Executed Key Test Events: Key tests required for already developed and partially fielded DRRS increments either did not have well-defined test plans, or these tests have yet to be conducted. According to program documentation, system releases and subreleases have been subjected to what are described as 30-day test cycles, during which: (1) a Software Test Plan is updated if applicable, (2) test procedures are developed and incorporated in the Software Test Description, (3) a series of developmental tests on each release/subrelease is performed, (4) weekly meetings are held to review software defects identified during testing, (5) final test results are summarized within the Software Test Report and Software Version Description, and (6) the release/subrelease is made available to users. However, the program office has yet to provide us with the developmental test plans and procedures for each release/subrelease that to-date has either been developed or fielded. Instead, it provided us with a Software Test Plan and two Software Test Descriptions that it said applied to two subreleases within release 4.0. However, available information indicates that DRRS subreleases total at least 63, which means that we have yet to receive the test plans and procedures for 61. Further, the test plan that we were provided is generic in nature, meaning that it was not customized to apply specifically to the two subreleases within Release 4.0. Moreover, the plan and procedures lack important elements specified in industry guidance.[Footnote 53] For example, the test plan does not include a schedule of activities to be performed or defined roles and responsibilities for performing them. Also, the test plan does not consistently include test entrance and exit criteria, a test defect management process, and metrics for measuring progress. Moreover, the DIO has yet to demonstrate that it has performed other key developmental and operational test events that are required before the software is fielded for operational use. According to DIO officials, developmental testing concludes only after system integration testing and system acceptance testing, respectively, are performed. Further, following developmental testing, the Joint Interoperability Test Command (JITC), which is a DOD independent test organization, is to conduct both interoperability and operational testing before the system is deployed and put into production (i.e., used operationally). Although increments of DRRS functionality have been put into production, the DIO has not performed system integration testing, system acceptance testing, or operational testing on any DRRS release or subrelease. Further, JITC documentation shows that while an interoperability test of an increment of DRRS functionality known as ESORTS was conducted, this test did not result in an interoperability certification.[Footnote 54] According to JITC and Joint Staff officials, this was because the DIO did not address JITC's identified limitations to the program's Information Support Plan, which identifies essential information-exchange sharing strategies between interdependent systems that are needed for interoperability certification. Without interoperability certification, the ability of the DRRS to exchange accurate and timely readiness data with other critical systems, such as the Joint Operation Planning and Execution System, cannot be ensured. Similarly, while DIO officials stated that acceptance testing has occurred for one increment of DRRS functionality known as SORTSREP, the DIO does not have either a finalized acceptance test plan or documented test results. [Footnote 55] Furthermore, the integrated master schedule (last updated in April 2009) shows that acceptance testing is not to occur until the July/August 2009 time frame, which is about 15 months later than originally envisioned. Moreover, this delay in acceptance testing has in turn delayed interoperability and operational testing by 16 months (May/June 2008 to September/November 2009), according to the latest schedule. Program officials attributed the delays to Marine Corps and Air Force concerns about the quality of SORTSREP. Until the DIO has effectively planned and executed the series of tests needed to demonstrate the readiness of DRRS increments to operate in a production environment, the risk of fielded system increments not performing as intended and requiring expensive rework to correct will be increased, and DOD will continue to experience delays in delivering mission-critical system capabilities to its readiness community. DIO Has Not Adequately Documented and Reported Test Results: Available results of tests performed on already developed and at least partially deployed DRRS releases/subreleases show that the test results have not been effectively captured and analyzed, and have not been fully reported. Moreover, test results for other releases/subreleases do not exist, thus minimizing the value of any testing that has been performed. According to relevant guidance, effective system testing includes recording the results of executing each test procedure and test case as well as capturing, analyzing, correcting, and disclosing to decision makers problems found during testing (test defects). [Footnote 56] It also includes ensuring that test entry and exit criteria are met before beginning and ending, respectively, a given test event. The DIO does not have test results of all developed and tested DRRS releases and subreleases. Specifically, program officials provided us with the Software Test Reports and Software Version Descriptions that, based on program documentation, represent the full set of test results for three subreleases and a partial set of test results for 40 subreleases within releases 1.0, 3.0, and 4.0. However, as noted earlier, DRRS subreleases total at least 63, which means that test reports and results for at least 20 subreleases do not exist. Moreover, the test reports and version descriptions that we received do not consistently include key elements provided for in industry guidance, such as a documented assessment of system capabilities and limitations, entrance/exit criteria status, an assessment as to whether the applicable requirements/thresholds were met, and unresolved defects and applicable resolution plans. [Footnote 57] This information is important because it assists in determining and disclosing to decision makers current system performance and efforts needed to resolve known problems, and provides program officials with a needed basis for ensuring that a system increment is ready to move forward and be used. Without this information, the quality and readiness of a system is not clear. Furthermore, the DIO does not have detailed test defect documentation associated with all executed DRRS test events. According to relevant guidance, defect documentation should, among other things, identify each issue discovered, assign each a priority/criticality level, and provide for each a strategy for resolution or mitigation. [Footnote 58] In lieu of detailed test defect documentation, program officials referred us to the above-mentioned Software Version Descriptions, and stated that additional information is available in an automated tool, known as the ISI BugTracker, that it uses to capture, among other things, defect data. However, these documents do not include the above- cited defect information, and defect data for each test event do not exist in the ISI BugTracker. Compounding the absence and limitations of test results are weaknesses in the program office's process for collecting such results during test execution. According to relevant guidance, test results are to be collected and stored according to defined procedures and placed under appropriate levels of control. [Footnote 59] Furthermore, these test results are to be reviewed against the source data to ensure that they are complete, accurate, and current. For DRRS, the program office is following a partially undocumented, manual process for collecting and storing test results and defects that involves a database and layers of documentation. As explained by program officials and documentation, the DIO initially documents defects and completed test case results manually on paper forms, and once the defect is approved by the test lead, it is input into a database. However, it does not have written procedures governing the entire process, and thus key controls, such as assigned levels of authority for database read/write access, are not clearly defined. Moreover, once the test results and defects are input into the database, traceability back to the original test data for data integrity checks cannot be established because the program office does not retain these original data sets. Program officials acknowledged these internal control weaknesses and stated that they intend to adopt a new test management tool that will allow them to capture in a single database test cases, test results, and test defects. Furthermore, the DIO's process for analyzing and resolving test defects has limitations. According to relevant guidance and the draft SORTSREP Test and Evaluation Master Plan (TEMP), defects should be analyzed and prioritized.[Footnote 60] However, the program office has not established a standard definition for defect priority levels identified during testing. For example, the various release/subrelease test reports (dated through January 2009) prioritize defects on a scale of 1- 3, where a priority 2 means critical but with a viable workaround. In contrast, the SORTSREP TEMP (dated January 2009) prioritizes defects on a scale of 1-5, where a priority 2 means an error that adversely affects the accomplishment of an operational or mission essential function in accordance with official requirements so as to degrade performance and for which no alternative work around solution exists. By not using standard priority definitions for categorizing defects, the program office cannot ensure that it has an accurate and useful understanding of the scope and magnitude of the problems it is facing at any given time, and it will not know if it is addressing the highest priority issues first. In addition, the DIO has not ensured that critical defects are corrected prior to concluding a given test event. According to relevant guidance and the draft SORTSREP TEMP, all critical and high defects should be resolved prior to the conclusion of a test event, and all test results should be reviewed for validity and completeness.[Footnote 61] However, the DRRS release/subrelease test reports show that the DIO concluded five test events even though each had at least 11 open critical defects (priority 1 defects with no workaround). Moreover, these numbers of open critical defects are potentially higher because they do not include defects for which a solution was identified but the solution failed during regression testing and do not include defects that were dismissed because the program official was unable to recreate it. Until the DIO adequately documents and reports the test results, and ensures that severe problems discovered are corrected prior to concluding a given test event, the probability of incomplete test coverage, and insufficient and invalid test results, is increased, thus unnecessarily increasing the risk of DRRS not meeting mission needs or otherwise not performing as intended. DIO Has Not Established an Effective Test Management Structure: The DIO does not have an effective test management structure, to include a well-defined overall test management plan that clearly assigns test management roles and responsibilities, a designated test management lead and a supporting working group, and a reliable schedule of planned test events. According to relevant guidance, these aspects of test management are essential to adequately planning, executing, and reporting a program's series of test events. [Footnote 62] Although the program has been underway for 8 years, it did not have an overarching DRRS TEMP until very recently (February 2009), and this plan is still in draft and has yet to be approved. Further, this draft TEMP does not clearly define DRRS test management roles and responsibilities, such as those of the test manager, and it does not include a reliable schedule of test events that reflect the program's recent restructuring of its software releases/subreleases into 10 modules. According to DIO officials, they recently decided not to approve this overarching TEMP. Instead, they said that they now intend to have individual TEMPs for each of the recently defined 10 modules, and to have supporting test plans for each module's respective developmental and operational test events. According to program documentation, three individual TEMPs are under development (i.e., SORTSREP tool and the Mission Readiness and Readiness Review modules). However, drafts of these TEMPs also do not clearly define test entrance and exit criteria, test funding requirements, an integrated test program schedule, and the respective test management roles and responsibilities. For example, while the draft SORTSREP TEMP identifies the roles and responsibilities of some players, such as the test manager, the personnel or organization that is to be responsible is not always identified. In addition, while the various players in the user community are identified (i.e., military services, combatant commands), their associated roles or responsibilities are not. Furthermore, the DIO has yet to designate a test management lead and establish an effective test management working group. According to relevant guidance, test management responsibility and authority should be assigned to an individual, and this individual should be supported by a working integrated product team that includes program office and operational testing representatives.[Footnote 63] Among other things, the working integrated product team is to develop an overall system test strategy. However, DIO officials told us that the test manager position has been vacant, and this position is now being temporarily filled by the program's chief engineer, who is a contractor. Furthermore, although DRRS system development began prior to 2004, a charter for a test and evaluation working integrated product team was not issued until February 2009. According to DIO officials, the delay in establishing the team has not had any impact because of corresponding delays in finalizing the program's overall test strategy. However, this statement is not consistent with the Defense Acquisition Guidebook, which states that two of the key products of the working integrated product team are the program's test strategy and TEMP. Further, JITC officials stated that the lack of a test manager and an active test and evaluation working integrated product team have reduced the effectiveness of DRRS testing activities. As a result, they stated that they have had to compensate by conducting individual meetings with the user community to discuss and receive documentation to support their operational and interoperability test planning efforts. Moreover, the DIO has yet to establish a reliable schedule of planned test events. For example, the schedule in the TEMPs is not consistent with either the integrated master schedule or the developmental test plans. Specifically, the draft SORTSREP TEMP (last updated in January 2009) identifies SORTSREP developmental testing occurring through January 2009 and ending in early February 2009, while the integrated master schedule (last updated in April 2009) shows SORTSREP development testing as occurring in the July/August 2009 time frame. In addition, while program officials said that development testing for SORTSREP has occurred, the associated development test plans (e.g., system integration and system acceptance test plans) had no established dates for test execution, and are still in draft. As another example, a module referred to as "Mission Readiness" had no established dates for test execution in its TEMP, and while program documentation indicates that this module completed development testing in December 2008, the associated development test plans (e.g., system integration and system acceptance test plans) do not exist[Footnote 64].: In addition, the DIO has yet to define in its draft TEMPs how the development and testing to date of at least 63 subreleases relate to the development and testing of the recently established 10 system modules. According to Joint Staff and JITC officials, they do not know how the releases/subreleases relate to the modules, and attributed this to a lack of an approved description for each module that includes what functionality each is intended to provide. Furthermore, the high-level schedule in the TEMP does not describe what test events for the DRRS releases/subreleases that have already been developed and deployed relate to the development test efforts planned for the respective modules. These problems in linking release/subrelease test events to module test events limit the DIO and JITC in leveraging the testing already completed, which in turn will impact the program's ability to meet cost, schedule, and performance expectations. Collectively, the weaknesses in this program's test management structure increase the chances that the deployed system will not meet certification and operational requirements, and will not perform as intended. DRRS Has Not Been Guided by a Reliable Schedule: The success of any program depends in part on having a reliable schedule that defines, among other things, when work activities will occur, how long they will take, and how they are related to one another. As such, the schedule not only provides a road map for the systematic execution of a program, but also provides the means by which to gauge progress, identify and address potential problems, and promote accountability. From its inception in 2002 until November 2008, the DIO did not have an integrated master schedule. Moreover, the only milestone that we could identify for the program prior to November 2008 was the date that DRRS was to achieve full operational capability, which was originally estimated to occur in fiscal year 2007, but later slipped to fiscal year 2008 and then fiscal year 2011, and is now fiscal year 2014--a 7-year delay. In addition, the DRRS integrated master schedule that was developed in November 2008, and was updated in January 2009 and again in April 2009 to address limitations that we identified and shared with the program office, is still not reliable. Specifically, our research has identified nine practices associated with developing and maintaining a reliable schedule.[Footnote 65] These practices are (1) capturing all key activities, (2) sequencing all key activities, (3) assigning resources to all key activities, (4) integrating all key activities horizontally and vertically, (5) establishing the duration of all key activities, (6) establishing the critical path for all key activities, (7) identifying float between key activities, (8) conducting a schedule risk analysis, and (9) updating the schedule using logic and durations to determine the dates for all key activities.[Footnote 66] However, the program's latest integrated master schedule does not address three of the practices and only partially addresses the remaining six. For example, the schedule does not establish a critical path for all key activities, include a schedule risk analysis, and it is not being updated using logic and durations to determine the dates for all key activities. Further, it does not fully capture, sequence, and establish the duration of all key work activities; fully assign resources to all key work activities; fully integrate all of these activities horizontally and vertically; and fully identify the amount of float-- the time that a predecessor activity can slip before the delay affects successor activities--between these activities. These practices are fundamental to producing a sufficiently reliable schedule baseline that can be used to measure progress and forecast slippages. (See table 3 for the results of our analyses relative to each of the nine practices.) Table 3: DRRS Satisfaction of Nine Schedule-Estimating Key Practices: Practice: Capturing all activities; Explanation: The schedule should reflect all key activities (e.g., steps, events, outcomes, etc.) as defined in the program's work breakdown structure, to include activities to be performed by both the government and its contractors; Practice met? Partially; GAO analysis: The schedule reflects many key activities as defined in the program's work breakdown structure. However, key activities are identified only as milestones rather than in terms of work to be performed and accomplished to achieve the milestones. Thus, the schedule does not account for all work to be performed. As a result, the reliability of the milestones is questionable. Practice: Sequencing all activities; Explanation: The schedule's activities need to be logically sequenced in the order that they are to be carried out. In particular, activities that must finish prior to the start of other activities (i.e., predecessor activities) as well as activities that cannot begin until other activities are completed (i.e., successor activities) should be identified. By doing so, interdependencies among activities that collectively lead to the accomplishment of key events or milestones can be established and used as a basis for guiding work and measuring progress; Practice met? Partially; GAO analysis: The schedule identifies some predecessor and successor activities, but not all. Specifically, out of 290 key activities, 139 do not identify predecessor activities and 121 do not identify successor activities. DIO officials stated that this is because interdependencies are discussed and addressed at various meetings, and thus do not need to be in the schedule. However, recognition of such interdependencies in the schedule is necessary to logically link tasks and events and thereby calculate dates and predict changes in the future. Without proper linkages, tasks that slip early in the schedule do not transmit delays to tasks that should depend on them. By not logically linking all key activities, the schedule does not provide a sufficient basis for understanding the program as a whole, and confidence that the right dates or critical paths are represented is low. This means that the DIO cannot use its schedule to identify disconnects as well as hidden opportunities, and cannot otherwise promote efficiency and accuracy, and control the program by comparing actual to planned progress. Practice: Assigning resources to all activities; Explanation: The schedule should realistically reflect what resources (i.e., labor, material, and overhead) are needed to do the work, whether all required resources will be available when they are needed, and whether any funding or time constraints exist; Practice met?: Partially; GAO analysis: The schedule identifies 15 resources that are assigned to various activities. However, important information is not included, which hampers its usefulness. For example, one benefit of loading resource information is to ensure that resources in short supply will not be overscheduled in any time period. However, the schedule does not define the amount of each resource that is available, but instead assumes only one unit of each resource is available. As a result, 10 of the 15 types of resources (e.g., information assurance and test and evaluation) in the schedule are overscheduled and thus estimated completion dates are not realistic. Further, the schedule does not identify the costs of the resources. Knowing the cost of resources is important, because some resources, such as labor, can cost more if the program takes longer. Practice: Establishing the duration of all activities; Explanation: The schedule should realistically reflect how long each activity will take to execute. In determining the duration of each activity, the same rationale, historical data, and assumptions used for cost estimating should be used for schedule estimating. Further, these durations should be as short as possible, and they should have specific start and end dates. Excessively long periods needed to execute an activity (i.e., greater than 2-3 months) should prompt further decomposition of the activity so that shorter execution durations will result; Practice met? Partially; GAO analysis: The schedule establishes the duration of key activities and includes specific start and end dates. However, the activities are not always of short duration. For example, 23 activities have remaining durations that exceed 100 days (108 to 551 days). By having a schedule summarized at too high a level, the program runs the risk of masking critical work elements within the key activities associated with executing the program and fails to show the risk management approaches being used. Further, it risks masking the schedule's true critical path. Practice: Integrating schedule activities horizontally and vertically; Explanation: The schedule should be horizontally integrated, meaning that it should link the products and outcomes associated with already sequenced activities. These links are commonly referred to as "hand offs" and serve to verify that activities are arranged in the right order to achieve aggregated products or outcomes. The schedule should also be vertically integrated, meaning that traceability exists among varying levels of activities and supporting tasks and subtasks. Such mapping or alignment among levels enables different groups to work to the same master schedule; Practice met? Partially; GAO analysis: The schedule is not horizontally integrated, meaning that the activities are not arranged in the right order to achieve aggregated products or outcomes. This is because, as previously noted, many activities do not identify interdependencies (predecessor and successor activities). As a result, management lacks information on how a slippage in completing one activity will affect others. In contrast, the schedule is vertically integrated, meaning that for those key activities that are identified, traceability exists among varying levels of activities and that lower-level activities are within the constraints of higher-level activities. Practice: Establishing the critical path for all activities; Explanation: The critical path--the longest duration path through the sequenced list of key activities--should be identified using scheduling software. The establishment of a program's critical path is necessary for examining the effects of any activity slipping along this path. High-risk activities along or near the critical path should also be identified and associated time impacts of these risks should be reflected in the schedule; Practice met?: No; GAO analysis: The schedule does not identify a valid critical path because there is no logical link between the program's key activities. Instead, the activities are presented as being independent from one another. Without a critical path, management cannot know the activities critical to the on-time completion of DRRS, or the impact of any changes to activities on the path. Practice: Identifying float between activities; Explanation: The schedule should identify float time so that schedule flexibility can be determined. As a general rule, activities along the critical path typically have the least amount of float time; Practice met? Partially; GAO analysis: The schedule identifies the amount of float allocated to key activities. However, the amount of float associated with 56 of these activities is unusually large (100 - 461 days). Such large amounts of float time can be attributed to the fact that many of the activities do not identify linkages to other activities (predecessor or successor activities). In addition, the amount of float between activities is of questionable accuracy because activity dependencies (predecessor or successor activities) are often not identified. Instead, the start and completion dates for many activities are imposed, rather than based on logic, such as an activity not starting until its predecessor is completed. Further, it is unclear whether activities along the critical path have the least amount of float because, as previously noted, the schedule does not have a valid critical path. Practice: Conducting a schedule risk analysis; Explanation: A schedule risk analysis uses a good critical path method schedule and data about project schedule risks as well as Monte Carlo simulation[A] techniques to predict the level of confidence in meeting a program's completion date, the amount of time contingency needed for a level of confidence, and the identification of high-priority risks. This analysis focuses not only on critical path activities but also on other schedule paths that may become critical. Schedule reserve for contingencies should be calculated by performing a schedule risk analysis. As a general rule, the reserve should be held by the project or program manager and applied as needed to those activities that take longer than scheduled because of the identified risks. Reserves of time should not be apportioned in advance to any specific activity since the risks that will actually occur and the magnitude of their impact is not known in advance; Practice met? No; GAO analysis: The program office did not perform a schedule risk analysis. Without this analysis, the office cannot sufficiently understand the level of confidence in meeting the program's completion date and identifying reserves for contingencies. Practice: Updating the schedule using logic and durations to determine the dates; Explanation: The schedule should use logic and durations in order to reflect realistic start and completion dates for program activities. The schedule should be continually monitored to determine when forecasted completion dates differ from the planned dates, which can be used to determine whether schedule variances will affect downstream work. Maintaining the integrity of the schedule logic is not only necessary to reflect true status, but is also required before conducting a schedule risk analysis. The schedule should avoid logic overrides and artificial constraint dates that are chosen to create a certain result on paper. To ensure that the schedule is properly updated, individuals trained in critical path method scheduling should be responsible for updating the schedule's status; Practice met? No; GAO analysis: The realism of start and completion dates for some activities is questionable because, as previously noted, some activities do not identify logical relationships (i.e., interdependencies with other activities) or are of unusually lengthy duration. In addition, there is no "status date" information in the schedule (i.e., evidence the overall schedule is updated on a regular basis to capture actual start and completion dates). Without this information, the impact of deviations on downstream work cannot be fully understood and proactively addressed. In addition, some dates in the schedule were updated erroneously. For example, the schedule shows 25 activities as having actual start dates in the future, including 6 starting in 2010, even though the updated schedule was developed in April 2009. Likewise, there are 12 activities with actual 100 percent complete finish dates that range from May 2009 to July 2010. Source: GAO analysis of DOD data. [A] A Monte Carlo simulation allows the model's parameters to vary simultaneously according to their associated probability distribution. The result is a set of estimated probabilities of achieving alternative outcomes, given the uncertainty in the underlying parameters. [End of table] The limitations in the program's latest integrated master schedule, coupled with the program's 7-year slippage to date, make it likely that DRRS will incur further delays. Compounding these limitations is the considerable concurrency in the key activities and events in the schedule associated with the 10 recently identified system modules (see figure 2). For example, in 2010 alone, the program office plans to complete development testing on 2 modules and operational testing on 3 modules, while also reaching initial operational capability on 3 modules and full operational capability on 2 modules. By way of comparison, the program office had almost no concurrency across a considerably more modest set of activities and events over the last 5 years, but nevertheless has fallen 7 years behind schedule. As previously reported, such significant overlap and concurrency among major program activities can create contention for limited resources and thus introduce considerable cost, schedule, and performance risks. [Footnote 67] Figure 2: Schedule for Developing and Implementing DRRS Capabilities: [Refer to PDF for image: illustrated table] Modules: Joint Mission Readiness: Begin planning, design, and development: Mid-2004; Complete developmental testing and evaluation: Late 2008; Complete operational testing and evaluation: Late 2009; Complete initial operational capability: Late 2004; Complete final operational capability: Early 2011. Modules: Unit Mission Readiness: Begin planning, design, and development: Mid-2004; Complete developmental testing and evaluation: Mid-2009; Complete operational testing and evaluation: Late 2009; Complete initial operational capability: Late 2009; Complete final operational capability: Mid-2010. Modules: Asset Visibility: Begin planning, design, and development: Early 2007; Complete developmental testing and evaluation: Mid-2010; Complete operational testing and evaluation: Late 2010; Complete initial operational capability: Late 2010; Complete final operational capability: Mid-2011. Modules: Business Intelligence: Begin planning, design, and development: Late 2008; Complete developmental testing and evaluation: Early 2011; Complete operational testing and evaluation: Mid-2011; Complete initial operational capability: Mid-2010; Complete final operational capability: Mid-2011. Modules: Installation Readiness: Begin planning, design, and development: Early 2009; Complete developmental testing and evaluation: Mid-2010; Complete operational testing and evaluation: Mid-2011; Complete initial operational capability: Mid-2009; Complete final operational capability: Mid-2011. Modules: Readiness Reviews: Begin planning, design, and development: Mid-2006; Complete developmental testing and evaluation: Late 2009; Complete operational testing and evaluation: Mid-2010; Complete initial operational capability: Mid-2007; Complete final operational capability: Mid-2010. Modules: Planning/Execution Support: Begin planning, design, and development: Early 2006; Complete developmental testing and evaluation: Late 2012; Complete operational testing and evaluation: Mid-2013; Complete initial operational capability: Mid-2013; Complete final operational capability: Late 2013. Modules: Historical Data: Begin planning, design, and development: Late 2008; Complete developmental testing and evaluation: Mid-2011; Complete operational testing and evaluation: Early 2012; Complete initial operational capability: Early 2012; Complete final operational capability: Late 2012. Modules: System Architecture: Begin planning, design, and development: Mid-2004; Complete final operational capability: Early 2012. Modules: End-User Usability: Begin planning, design, and development: Late 2004; Complete final operational capability: Mid-2009. Source: GAO analysis based on DOD data. [End of figure] In addition, the schedule remains unstable as evidenced by the degree of change it has experienced in just the past few months. For example, the January 2009 schedule had a full operational capability milestone of October 2011. By contrast, the April 2009 schedule has a December 2013 milestone (see figure 3 below). Moreover, some milestones are now to occur much earlier than they were a few months ago. For example, the January 2009 schedule shows initial operational capability for "readiness reviews" to be June 2010. However, the April 2009 schedule shows that this milestone was attained in August 2007. Overall, multiple milestones for four modules were extended by at least 1 year, including two milestones that were extended by more than 2 years. Such change in the schedule in but a few months suggests a large degree of uncertainty, and illustrates the importance of ensuring that the schedule is developed in accordance with best practices. Figure 3: Changes in Estimated Dates for DRRS Capabilities: [Refer to PDF for image: illustration] Modules: Joint Mission Readiness; Complete final operational capability: Late 2009 to Early 2001. Modules: Unit Mission Readiness; Complete operational testing and evaluation: Early 2009 to Mid-2009. Modules: Asset Visibility; Complete developmental testing and evaluation: Late 2009 to Mid-2010; Complete operational testing and evaluation: Early 2010 to Late 2010; Complete initial operational capability: Early 2010 to Late 2010; Complete final operational capability: Late 2011 to Mid-2011. Modules: Business Intelligence; Complete developmental testing and evaluation: Mid-2010 to Early 2011; Complete operational testing and evaluation: Mid-2010 to Mid-2011; Complete initial operational capability: Mid-2010; Complete final operational capability: Mid-2011. Modules: Installation Readiness; Complete developmental testing and evaluation: Mid-2010; Complete operational testing and evaluation: Mid-201; Complete initial operational capability: Mid-2011 to Mid-2009; Complete final operational capability: Mid-2011 to Late 2011. Modules: Readiness Reviews; Complete operational testing and evaluation: Late 2009; Complete initial operational capability: Mid-2010 to Mid-2007; Complete final operational capability: Mid-2010. Modules: Planning/Execution Support; Complete developmental testing and evaluation: Early 2011 to Late 2012; Complete operational testing and evaluation: Mid-2011 to Mid-2013; Complete initial operational capability: Mid-2011 to Mid-2013; Complete final operational capability: Mid-2011 to Late 2013. Modules: Historical Data; Begin planning, design, and development: Mid-2009 to Early 2009; Complete developmental testing and evaluation: Late 2009 to Mid-2011; Complete operational testing and evaluation: Mid-2010 to Early 2012; Complete initial operational capability: Mid-2010 to Early 2012; Complete final operational capability: Mid-2010 to Late 2012. Modules: System Architecture; Complete final operational capability: Mid-2011 to Early 2012. Modules: End-User Usability; Complete final operational capability: Late 2008 to Mid-2009. Source: GAO analysis based on DOD data. [End of figure] DIO Lacks Adequate Staffing and an Effective Approach to Meeting its Human Capital Needs: As we have previously reported, effective human capital management is an essential ingredient to achieving successful program outcomes. [Footnote 68] Among other things, effective human capital management involves a number of actions to proactively understand and address any shortfalls in meeting a program's current and future workforce needs. These include an assessment of the core competencies and essential knowledge, skills, and abilities needed to perform key program management functions, an inventory of the program's existing workforce capabilities, and an analysis of the gap between the assessed needs and the existing capabilities. Moreover, they include explicitly defined strategies and actions for filling identified gaps, such as strategies for hiring new staff, training existing staff, and contracting for support services. The DIO is responsible for performing a number of fundamental DRRS program management functions. For example, it is responsible for acquisition planning, performance management, requirements development and management, test management, contractor tracking and oversight, quality management, and configuration management. To effectively perform such functions, program offices, such as the DIO, need to have not only well-defined policies and procedures and support tools for each of these functions, but also sufficient human capital to implement the processes and use the tools throughout the program's life cycle. Without sufficient human capital, it is unlikely that a program office can effectively perform its basic program management functions, which in turn increases the risk that the program will not deliver promised system capabilities and benefits on time and on budget. The DIO does not currently have adequate staff to fulfill its system acquisition and deployment responsibilities. In particular, the DIO is staffed with a single full-time government employee--the DIO Director. All other key program office functions are staffed by either contractor staff or staff temporarily detailed, on an as-needed basis, from other DOD organizations (referred to as "matrixed" staff). As a result, program management positions that the DIO itself has identified as critical to the program's success, such as configuration manager and security manager, are being staffed by contractors. Moreover, these contractor staff report to program management positions also staffed by contractors. Other key positions, such as those for performing acquisition management, requirements development and management, and performance management, have not even been established within the DIO. Furthermore, key positions, such as test manager, are vacant. These human capital limitations were acknowledged by the DRRS Executive Committee in November 2008. According to DIO and contractor officials, they recognize that additional program management staffing is needed. They also stated that while DRRS has been endorsed by USD (P&R) leadership and received funding support, past requests for additional staff have not been approved by USD (P&R) due to other competing demands for staffing. Further, DIO officials stated that the requests for staff were not based on a strategic gap analysis of its workforce needs and existing capabilities. Specifically, the program has not assessed its human capital needs and the gap between these needs and its onboard workforce capabilities. Until the program office adopts a strategic, proactive approach to managing its human capital needs, it is unlikely that it will have an adequate basis for obtaining the people it needs to effectively and efficiently manage DRRS. [End of section] Appendix III: Comments from the Department of Defense: Office Of The Under Secretary Of Defense: Personnel And Readiness: 4000 Defense Pentagon: Washington, DC 20301-4000: September 3, 2009: Ms. Sharon Pickup: Director Defense Capabilities and Management: U.S. Government Accountability Office: Washington, D.C. 22548: Dear Ms. Pickup: This is the Department of Defense (DoD) response to the GAO draft report, GAO-09-518, "Military Readiness: DoD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System," dated July 1, 2009 (GAO Code 351217). Thank you for the opportunity to review this draft report. However, in our view, the report is flawed in its assessment of the Defense Readiness Reporting System (DRRS). DRRS is a true net-centric application that provides broad and detailed visibility on readiness issues. Achieving this data sharing across the DoD enterprise was groundbreaking work, fraught with barriers and obstacles - many of which have now been overcome. We were disappointed to learn that the report did not address the cultural impediments that are the root cause of many of the issues it cites, and the cause of so many previous congressional concerns on readiness reporting. Instead, the report focused on past issues and problems in the software development and acquisition processes that have now been remedied. We believe that this focus, coupled with inaccurate and misleading factual information included in the report, has lead the GAO to develop an incomplete picture of the program. Attached, please find the Department's detailed comments in response to the GAO's specific recommendation. Sincerely, Signed by: William J. Carr: Deputy Under Secretary of Defense (Military Personnel Policy): Performing the Duties of the Under Secretary of Defense (Personnel and Readiness): Attachment: As stated: [End of letter] GAO Draft Report - Dated July 1, 2009: GAO Code 351217/GAO-09-518: "Military Readiness: DoD Needs to Strengthen Management and Oversight of the Defense Readiness Reporting System" Department Of Defense Comments To The Recommendations: Recommendation 1: The GAO recommends that the Secretary of Defense direct the Deputy Secretary of Defense, as the Chair of the Defense Business Systems Management Committee, to reconsider the committee's recent approval of the Defense Readiness Reporting System (DRRS) planned investment for fiscal years 2009 and 2010, and convene the Defense Business Systems Management Committee to review the program's past performance and the DRRS Implementation Office's capability to manage and deliver DRRS going forward. DOD Response: Non-Concur. The IRB and DBSMC certifications were granted in compliance with the established certification process, which is designed to not be overly redundant with the acquisition process. Consistent with the Department's concept of tiered accountability, oversight of the specific issues identified by the GAO in this report are the responsibility of component responsible for the system. In this case, USD (P&R) chartered a DRRS Executive Committee (DEXCOM) to provide for governance of the effort in the readiness community, The DEXCOM has and will continue to provide appropriate governance for this effort. Additionally, the USD (P&R) will ensure that the program is compliant with all acquisition requirements prior to submission for further certifications. Recommendation 2: The GAO recommends that the Secretary of Defense direct the Deputy Secretary of Defense to have the Director of the Business Transformation Agency, using the appropriate team of functional and technical experts and the established risk assessment methodology, conduct a program risk assessment of DRRS, and to use the findings in GAO's report and the risk assessment to decide how to redirect the program's structure, approach, funding, management, and oversight. In this regard, GAO recommends that the Secretary of Defense direct the Deputy Secretary of Defense to solicit the advice and recommendations of the DRRS Executive Committee. DOD Response: Concur. The Department accepts the GAO's recommendation that a program risk assessment would be constructive. The Business Transformation Agency (BTA) will work with the DRRS Executive Committee (DEXCOM) to determine the scope of the assessment and will subsequently work with the DEXCOM to analyze its results. This assessment will be complete by the middle of Fiscal Year 2010. Recommendation 3: The GAO recommends that the Secretary of Defense through the appropriate chain of command, take steps to ensure that DRRS requirements are effectively developed and managed with appropriate input from the Services, Joint Staff, and Combatant Commanders, including: (a) establishing an authoritative set of baseline requirements prior to further system design and development; (b) ensuring that the different levels of requirements and their associated design specifications and test cases, are aligned with one another, and; (c) developing and instituting a disciplined process for reviewing and accepting changes to the baseline requirements in light of estimated costs, benefits, and risk. DOD Response: Non-Concur. The DRRS program has an authoritative set of baseline requirements already established. The current DRRS governance process, which includes membership from all CoComs, Services, Combat Support Agencies (CSAs), Joint Staff, and OSD oversees the DRRS requirements management process. These requirements are reviewed bi- weekly as part of the DRRS configuration control process. Recommendation 4: The GAO recommends that the Secretary of Defense through the appropriate chain of command, take steps to ensure that DRRS testing is effectively managed, including: (a) developing lest plans and procedures for each system increment test event that include a schedule of planned test activities, defined roles and responsibilities, test entrance and exit criteria, test defect management processes, and metrics for' measuring test progress, (b) ensuring that all key lest events are conducted on all DRRS increments, (c) capturing, analyzing, reporting, and resolving all test results and test defects of all developed and tested DRRS increments; and, (d) establishing an effective test management structure that includes assigned test management roles and responsibilities, a designated test management lead and a supporting working group, and a reliable schedule of test events. DoD Response: Non-Concur. DRRS testing is already in place and performing effectively. The DIO goes through a rigorous testing regimen that includes documenting test plans with user test cases for each incremental release. The DRRS program utilizes System Integration Testing (SIT), System Acceptance Testing (SAT), Interoperability Testing (IOT) and Operational Testing (OT) with System Integration Test Plans and System Acceptance Test Plans. There are designated testers independent of the developers that validate the user test cases and functionality prior to a deployment decision. Finally, the D10 conducts a release review with the developers and the testers to determine if the increment is ready for release and meets all the exit criteria. For interoperability testing with systems external to DRRS, the DIO has a designated test director and the Joint Interoperability Test Command (JITC) is the designated Interoperability Test Activity and the Operational Test Activity. JITC is responsible for developing the Interoperability Test Plan and the Operational Test Plan. Recommendation 5: The GAO recommends t oat the Secretary of Defense through the appropriate chain of command, take steps to ensure that the DRRS integrated master schedule is reliable, including ensuring that the schedule: (a) captures all activities from the work breakdown structure, including the work to be performed and the resources to he used; (b) identifies the logical sequencing of all activities, including defining predecessor and successor activities; (c) reflects whether all required resources will he available when needed and their cost; (d) ensures that all activities and their duration are not summarized at a level that could mask critical elements; (e) achieves horizontal integration in the schedule by ensuring that all external interfaces (hand-offs) are established and interdependencies among activities are defined; (f) identifies float between activities by ensuing that the linkages among all activities are defined; (g) defines a critical path that runs continuously to the program's finish date; (h) incorporates the results of a schedule risk analysis to determine the level of confidence in meeting the program's activities and completion date, and; (i) includes the actual start and completion dotes of work activities per formed so that the impact of deviations on downstream work can he proactively addressed. DOD Response: Non Concur. A process is already in place to ensure the DRRS integrated master schedule is current, reliable, and meets all of the criteria outlined in the recommendation, With the approval of the DEXCOM charter and the renewed/refined Battle Staff and General Officer Steering Committee (GOSC), many of the program issues regarding Plan of Action and Milestones (POA&M) software development, integration, and testing in the GAO report are being address or have been corrected. Recommendation 6: The GAO recommends t nit the Secretary of Defense through the appropriate chain of command, take steps to ensure that the DRRS program office is staffed on the basis of a human capital strategy that is grounded in an assessment of the core competencies and essential knowledge, skills, and abilities needed to perform key DRRS program management functions, an inventory of the program office's existing workforce capabilities, and an analysis of the gap between the assessed needs and the existing capabilities. DOD Response: Partially concur. The DIO was intentionally kept small in order to maximize flexibility and responsiveness to customer requirements as well as to implement commercial best practices for agile software development. However, due to increase demands on DRRS through additional requirements and capabilities requested by users, additional billets, both military and civilian are needed to expedite the implementation of DRRS. To that end, the USD (P& R) is planning on adding full time civilian acquisition support for the DIO and the conversion of contractor to civilian billets during 2010/2011 timeframe as part of the Department's in-sourcing initiative. Recommendation 7: The GAO recommends brat the Secretary of Defense through the appropriate chain of command, take steps to ensure that DRRS is developed and implemented in a manner that does not increase the reporting burden on units and addresses the timeliness. precision, and objectivity of metrics that are reported to Congress. DOD Response: Non-Concur. One of the primary tenets for DRRS development has always been to reduce the reporting requirements on the war fighter. DRRS is already using state of the art technology in order to ensure that data which already exists is available within the DRRS enterprise in near-real time for the war fighters. The MRS governance structure that is currently in place ensures that DRRS development does not deviate from these core principles. Recommendation 8: The GAO recommends that the Secretary of Defense direct the Deputy Secretary of Defense to ensure that both. he Human Resources Management Investment Review Board and the DRRS Executive Committee conduct frequent oversight activities of the DRRS program. and report any significant issues to the Deputy Secretary of Defense. DOD Response: Concur. The Department is committed to ensuring that the proper level of oversight of the DRRS program is adhered to. As an Acquisition Category Ill level program, DRRS will be managed from an acquisition perspective by the Component Acquisition Executive (CAE) within USD(P&R). The USD (P&R) CAE is already working with the program to ensure that they become fully compliant with all acquisition requirements. This is consistent with tiered accountability. The CAE will certify to the IIRM IRB and the DCMO of compliance prior to submission of future certification requests. Finally, the current DEXCOM governance process will continue to provide sustained functional oversight of the DRRS program. Issues that arise in any of these areas will be elevated for review as appropriate. [End of section] Appendix IV: GAO Contacts and Staff Acknowledgments: GAO Contacts: Sharon L. Pickup, (202) 512-9619 or pickups@gao.gov Randolph C. Hite, (202) 512-6256 or hiter@gao.gov: Staff Acknowledgments: In addition to the contacts named above, key contributors to this report were Michael Ferren (Assistant Director), Neelaxi Lakhmani (Assistant Director), April Baugh, Mathew Butler, Richard J. Hagerman, Nicole Harms, James Houtz, John Lee, Stephen Pruitt, Terry Richardson, Karen Richey, Karl Seifert, and Kristy Williams. [End of section] Footnotes: [1] Pub. L. No. 105-261, §373 (1998). Codified at 10 U.S.C. §117. [2] Department of Defense Directive 7730.65, Department of Defense Readiness Reporting System (DRRS) (June 3, 2002). [3] ESORTS is an automated readiness reporting tool designed to collect capability and resource data, while DRRS is the broader system that, according to the 2002 DRRS Directive, measures and reports on the readiness of military forces and the supporting infrastructure to meet missions and goals assigned by the Secretary of Defense. However, ESORTS is now viewed as a tool within DRRS. [4] Secretary of Defense Memorandum, Policy Implementation to Establish Commander, USJFCOM (CDRUSJFCOM), as the Primary Joint Force Provider (JFP) (June 25, 2004). The U.S. military organizes its global presence into a series of geographic and functional combatant commands. The geographic combatant commands--U.S. Central Command, U.S. European Command, U.S. Northern Command, U.S. Pacific Command, U.S. Southern Command, and U.S. Africa Command --have authority over all U.S. military forces operating within a specified area of operation and are directly responsible for the performance of missions assigned to the command. The functional combatant commands--U.S. Joint Forces Command, U.S. Special Operations Command, U.S. Strategic Command, and U.S. Transportation Command--possess worldwide functional responsibilities, such as joint training, force provision, and global command and control. [5] Chairman of the Joint Chiefs of Staff Instruction 3401.02A, Global Status of Resources and Training System (GSORTS) (Feb. 27, 2004). [6] All combat, combat support, and combat service support units that have the potential to support an: operations, contingency, or single integrated operational plan; or a contingency operation are required to report. [7] A mission is a task or a series of tasks with a purpose. Mission essential tasks are those tasks that are absolutely necessary, indispensable, or critical to mission success. DRRS' capability data are based on commanders' assessments of their organizations' capabilities to carry out their missions and mission essential tasks. [8] The Strom Thurmond National Defense Authorization Act for Fiscal Year 1999, Pub. L. No. 105-261 §373 (1998), codified at 10 U.S.C. §117. Section 373 initially directed the Secretary to establish and implement the readiness reporting system no later than January 15, 2000. An amendment in the National Defense Authorization Act for Fiscal Year 2000, Pub. L. No. 106-65, §361 changed the date from January 15, 2000, to April 1, 2000. [9] Secretary of Defense Memorandum, Policy Implementation to Establish Commander, USJFCOM (CDRUSJFCOM), as the Primary Joint Force Provider (JFP), (June 25, 2004). The memorandum's primary purpose was to direct the Commander of the U.S. Joint Forces Command to assume the duties of DOD's primary joint force provider. [10] Under Secretary of Defense for Personnel and Readiness Memorandum, Department of Defense Readiness Reporting System (DRRS) Interim Implementation Guidance (Nov. 2, 2004). USD (P&R) issued three subsequent memorandums between August 10, 2005, and August 23, 2006, which expanded and clarified reporting requirements. [11] The four Investment Review Boards are for (1) Financial Management, established by the Deputy Under Secretary of Defense for Financial Management; (2) Weapon Systems Lifecycle Management and Materiel Supply and Services Management; (3) Real Property and Installations Lifecycle Management, both established by the Under Secretary of Defense for Acquisition, Technology and Logistics; and (4) Human Resources Management, established by USD (P&R). [12] The purpose of a risk assessment is to reduce systemic risk and support informed decision making by focusing on delivering business capabilities rapidly and at a reduced cost, and identifying program vulnerabilities and assisting in developing mitigation solutions. The assessment is generally performed prior to major acquisition decisions, although assessments may be requested for other reasons. [13] See for example, Department of Defense Instruction 5000.02, Operation of the Defense Acquisition System (Dec. 8, 2008); Defense Acquisition University, Test and Evaluation Management Guide, 5th ed. (Fort Belvoir, Va.: January 2005); Institute of Electrical and Electronics Engineers, Standard 1012-2004 for Software Verification and Validation (New York, N.Y.: June 8, 2005); GAO, GAO Cost Estimating and Assessment Guide: Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington, D.C.: March 2009); and GAO, A Model of Strategic Human Capital Management, GAO-02-373SP (Washington, D.C.: Mar. 15, 2002). [14] The users of readiness data include the Secretary of Defense, the military services, the Chairman of the Joint Chiefs of Staff, the combatant commands, defense agencies, and field activities. [15] GAO, Military Readiness: New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed, [hyperlink, http://www.gao.gov/products/GAO-03-456] (Washington, D.C.: Mar. 28, 2003). [16] GAO, GAO Cost Estimating and Assessment Guide: Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington, D.C.: March 2009). [17] Float is the amount of time an activity can slip before affecting the critical path. [18] A Major Automated Information System is a program or initiative that is so designated by the Assistant Secretary of Defense (Networks and Information Integration)/Chief Information Officer or that is estimated to require program costs in any single year in excess of $32 million, total program costs in excess of $126 million, or total life- cycle costs in excess of $378 million in fiscal year 2000 constant dollars. [19] As previously noted, the Investment Review Board is responsible for all business system investments in DOD that involve more than $1 million in obligations. [20] The primary GSORTS metric--the "C" rating--is a mission-focused metric that is based not only on a unit's resources but also on the unit's capabilities to execute training to standards, in a specified environment. [21] The GSORTS metric that captures information concerning a unit's capability to execute these other assigned or directed missions is the percent effective metric. This percent effective rating is a subjective assessment that is based on a commander's professional military judgment. It is based on a number of considerations but not strictly tied to resources or training performance. [22] The software to collect and display this enhanced capability data has been in place for several years and the DIO has stated that DRRS reached its initial operating capability when the system was able to collect and display this information. [23] The SIPRNet is operated by the Defense Information Systems Agency and is DOD's largest interoperable command and control data network. In addition to supporting the input of data to DRRS, the SIPRNet supports the Global Command and Control System, the Defense Message System, collaborative planning, and numerous other classified warfighter applications. [24] The memorandum--Under Secretary of Defense for Personnel and Readiness Memorandum, Department of Defense Readiness Reporting System (DRRS) Interim Implementation Guidance, (Aug. 10, 2005)--directed units report mission essential task capabilities in DRRS through the automated ESORTS reporting tool. [25] When missions or mission essential tasks are not standardized, capability searches may not yield desired results because similar units are measuring their capabilities against different task conditions and standards. [26] Under Secretary of Defense for Personnel and Readiness Memorandum, Status of Defense Readiness Reporting System (DRRS) Implementation (Sept. 29, 2005). [27] Current reporting guidelines require the services to report both in the Chairman of the Joint Chiefs of Staff's GSORTS and in OSD's DRRS. [28] Department of the Air Force Memorandum, Defense Readiness Reporting System (DRRS) Core Unit Mission Essential Task List (METL) Implementation Guidance (Mar. 3, 2009). [29] Marine Corps Administrative Message 0307//09, Update to Interim Defense Readiness Reporting System (DRRS) Policy and Procedures for Marine Corps Units and Installations (May 12, 2009). [30] The law also lists annual requirements for training establishments, defense installations, and facilities to report their capabilities, and a number of other monthly, quarterly, and annual requirements. [31] While certain data, such as personnel data, are captured in multiple data systems, the services or OSD designate a single data system as the authoritative database for each type of information. [32] [hyperlink, http://www.gao.gov/products/GAO-03-456]. [33] Chairman of the Joint Chiefs of Staff Instruction 3401.02A, Global Status of Resources and Training System (GSORTS) (Feb. 27, 2004), permits commanders to subjectively upgrade or downgrade their units' overall assessment, although not their individual resource levels. However, these changes are easy to identify and commanders are required to provide justifications for any changes. A recent GAO review of Army data found that commanders' subjective changes constituted less than 10 percent of the total assessments. [34] The memorandum's primary purpose was to provide policy implementation establishing the Commander of the U.S. Joint Forces Command to assume the duties as DOD's primary joint force provider. [35] These historical data would include training data, personnel data, and equipment supply and condition data as well as data concerning unit capabilities to perform the missions they were organized and designed to perform. [36] Ronald W. Reagan National Defense Authorization Act for Fiscal Year 2005, Pub. L. No. 108-375, § 332 (2004) (codified at 10 U.S.C. §§ 186 and 2222). [37] Department of Defense Instruction Number 5000.02, Operation of the Defense Acquisition System, Enclosure 11 (December 8, 2008). [38] The Capability Maturity Model® Integration for Development, developed by the Software Engineering Institute of Carnegie Mellon University, defines key practices that are recognized hallmarks for successful organizations that, if effectively implemented, can greatly increase the chances of successfully developing and acquiring software and systems. Carnegie Mellon Software Engineering Institute, Capability Maturity Model® Integration for Development, version 1.2 (Pittsburgh, Pa.: August 2006). [39] See for example, Department of Defense Instruction 5000.02, Operation of the Defense Acquisition System (Dec. 8, 2008); Defense Acquisition University, Test and Evaluation Management Guide, 5th ed. (Fort Belvoir, Va.: January 2005); Institute of Electrical and Electronic Engineers, Standard 1012-2004 for Software Verification and Validation (New York, N.Y.: June 8, 2005); and Software Engineering Institute, Capability Maturity Model Integration for Acquisition version 1.2 (Pittsburgh, Pa.: November 2007). [40] GAO, GAO Cost Estimating and Assessment Guide: Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington D.C.: March 2009). [41] GAO, A Model of Strategic Human Capital Management, [hyperlink, http://www.gao.gov/products/GAO-02-373SP] (Washington, D.C.: Mar. 15, 2002). [42] Department of Defense Directive 7730.65, Department of Defense Readiness Reporting System (DRRS) (June 3, 2002). [43] The reporting officer for U.S. Pacific Command indicated that he felt that the GAO's survey did not "accurately and fully assess the usefulness and functionality of DRRS" and provided additional written information to support his survey answers. After consultation with GAO's office of Applied Research and Methods we determined the survey did assess DRRS functionality. In order to capture U.S. Pacific Command's concerns, we incorporated the information it provided into the report sections describing current DRRS usage. [44] See for example, Department of Defense Instruction 5000.02, Operation of the Defense Acquisition System (Dec. 8, 2008); Defense Acquisition University, Test and Evaluation Management Guide, 5th ed. (Fort Belvoir, Va.: January 2005), Institute of Electrical and Electronic Engineers, Standard 1012-2004 for Software Verification and Validation (New York, N.Y.: June 8, 2005); GAO, GAO Cost Estimating and Assessment Guide: Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington D.C.: March 2009); and GAO, A Model of Strategic Human Capital Management, [hyperlink, http://www.gao.gov/products/GAO-02-373SP] (Washington, D.C.: Mar. 15, 2002). [45] The Capability Maturity Model® Integration for Development, developed by the Software Engineering Institute of Carnegie Mellon University, defines key practices that are recognized hallmarks for successful organizations that, if effectively implemented, can greatly increase the chances of successfully developing and acquiring software and systems. Carnegie Mellon Software Engineering Institute, Capability Maturity Model® Integration for Development, version 1.2 (Pittsburgh, Pa.: August 2006). [46] GAO, Secure Border Initiative: DHS Needs to Address Significant Risks in Delivering Key Technology Investment, [hyperlink, http://www.gao.gov/products/GAO-08-1086] (Washington, D.C.: Sept. 22, 2008). [47] Office of the Under Secretary of Defense for Personnel and Readiness Memorandum, Implementing the Defense Readiness Reporting System (July 1, 2002). [48] Department of Defense Directive 7730.65, Department of Defense Readiness Reporting System (DRRS) (June 3, 2002). Additional requirements for DRRS have been established in DOD directives and instructions that govern information systems including information assurance, net-centric architecture, net-centric data strategy, and information system interoperability and supportability. [49] Department of Defense Directive 7730.65, Department of Defense Readiness Reporting System (DRRS) (June 3, 2002). [50] [hyperlink, http://www.gao.gov/products/GAO-08-1086]. [51] The Capability Maturity Model® Integration for Development, developed by the Software Engineering Institute of Carnegie Mellon University, defines key practices that are recognized hallmarks for successful organizations that, if effectively implemented, can greatly increase the chances of successfully developing and acquiring software and systems. Carnegie Mellon Software Engineering Institute, Capability Maturity Model® Integration for Development, Version 1.2 (Pittsburgh, Pa.: August 2006). [52] See for example, Department of Defense Instruction 5000.02, Operation of the Defense Acquisition System; Defense Acquisition University, Test and Evaluation Management Guide; Institute of Electrical and Electronics Engineers, Standard 1012-2004 for Software Verification and Validation; and Software Engineering Institute, Capability Maturity Model Integration for Acquisition version 1.2 (Pittsburgh, Pa.: November 2007). [53] See for example, Institute of Electrical and Electronics Engineers, Standard 829-2008 Standard for Software and System Test Documentation (New York, N.Y.: July 18, 2008) and Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2. [54] According to the Concept of Operations, ESORTS is to provide current readiness status for operational forces and defense support organizations in terms of their ability to perform their mission essential tasks. [55] According to the draft SORTSREP Test and Evaluation Master Plan, SORTSREP is a tool to capture and input military departments' readiness data requirements into a readiness reporting database. [56] See for example, Defense Acquisition University, Defense Acquisition Guidebook (Arlington, Va.: November 2006); Institute of Electrical and Electronics Engineers, Standard 1012-2004 for Software Verification and Validation; Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2. [57] See for example, Institute of Electrical and Electronics Engineers, Standard 829-2008 Standard for Software and System Test Documentation and Software Engineering Institute, Capability Maturity Model Integration for Acquisition version 1.2. [58] See for example, Institute for Electrical and Electronics Engineers, Standard 829-2008 Standard for Software and System Documentation. [59] See for example, Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2. [60] See for example, GAO, Office of Personnel Management: Improvements Needed to Ensure Successful Retirement Systems Modernization, [hyperlink, http://www.gao.gov/products/GAO-08-345] (Washington, D.C.: January 2008); Institute of Electrical and Electronics Engineers, Standard 1012-2004 for Software Verification and Validation; and Institute of Electrical and Electronics Engineers, Standard 1012-1986 for Software Verification and Validation Plans (New York, N.Y.: Nov. 14, 1986). [61] See for example, [hyperlink, http://www.gao.gov/products/GAO-08-345]; Institute of Electrical and Electronics Engineers, Standard 1012-2004 for Software Verification and Validation; and Institute of Electrical and Electronics Engineers, Standard 1012-1986 for Software Verification and Validation Plans. [62] See for example, Department of Defense Instruction 5000.02, Operation of the Defense Acquisition System; Defense Acquisition University, Defense Acquisition Guidebook; Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2; Institute for Electrical and Electronics Engineers, Standard 829-2008 Software and System Test Documentation (New York, N.Y.: July 18, 2008). [63] See for example, Defense Acquisition University, Defense Acquisition Guidebook; Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2; Institute for Electrical and Electronics Engineers, Standard 829-2008 Software and System Test Documentation. [64] According to the Mission Readiness TEMP, Mission Readiness is to create a common standard metric for assessing readiness by capability that can be uniformly applied throughout the department. [65] GAO, GAO Cost Estimating and Assessment Guide: Best Practices for Developing and Managing Capital Program Costs, [hyperlink, http://www.gao.gov/products/GAO-09-3SP] (Washington, D.C.: March 2009). [66] Float is the amount of time an activity can slip before affecting the critical path. [67] GAO, Information Technology: Improvements for Acquisition of Customs Trade Processing System Continue, but Further Efforts Needed to Avoid More Cost and Schedule Shortfalls, GAO-08-46 (Washington, D.C.: Oct. 25, 2007) and Secure Border Initiative: SBInet Planning and Management Improvements Needed to Control Risks, [hyperlink, http://www.gao.gov/products/GAO-07-504T] (Washington, D.C.: Feb. 27, 2007). [68] GAO, A Model of Strategic Human Capital Management, [hyperlink, http://www.gao.gov/products/GAO-02-373SP] (Washington, D.C.: Mar. 15, 2002). [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO‘s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO‘s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.