Military Readiness

New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed Gao ID: GAO-03-456 March 28, 2003

The Department of Defense's (DOD) readiness assessment system was designed to assess the ability of units and joint forces to fight and meet the demands of the national security strategy. In 1998, GAO concluded that the readiness reports provided to Congress were vague and ineffective as oversight tools. Since that time, Congress added reporting requirements to enhance its oversight of military readiness. Therefore, the Chairman asked GAO to examine (1) the progress DOD made in resolving issues raised in the 1998 GAO report on both the unit-level readiness reporting system and the lack of specificity in DOD's Quarterly Readiness Reports to the Congress, (2) the extent to which DOD has complied with legislative reporting requirements enacted since 1997, and (3) DOD's plans to improve readiness reporting.

Since 1998, DOD has made some progress in improving readiness reporting--particularly at the unit level--but some issues remain. For example, DOD uses readiness measures that vary 10 percentage points or more to determine readiness ratings and often does not report the precise measurements outside DOD. DOD included more information in its Quarterly Readiness Reports to the Congress. But quality issues remain--in that the reports do not specifically describe readiness problems, their effects on readiness, or remedial actions to correct problems. Nor do the reports contain information about funding programmed to address specific remedial actions. Although current law does not specifically require this information, Congress could use it for its oversight role. DOD complied with most, though not all, of the legislative readiness reporting requirements enacted by Congress in the National Defense Authorization Acts for Fiscal Years 1998-2002. For example, DOD (1) is now listing the individual units that have reported low readiness and reporting on the readiness of prepositioned equipment, as required by the fiscal year 1998 Act; (2) is reporting on 11 of 19 readiness indicators that commanders identified as important and that Congress required to be added to the quarterly reports in the fiscal year 1998 Act, but is not reporting on the other 8 readiness indicators; and (3) has not yet implemented a new comprehensive readiness reporting system as required in the fiscal year 1999 Act. As a result, Congress is not receiving all the information mandated by law. DOD issued a directive in June 2002 to establish a new comprehensive readiness reporting system that DOD officials said they plan to use to comply with the reporting requirements specified by Congress. The new system is intended to implement many of the recommendations included in a congressionally directed independent study for establishing such a system. However, the extent to which the new system will actually address the current system's shortcomings is unknown, because the new system is currently only a concept, and full capability is not scheduled until 2007. As of January 2003, DOD had not developed an implementation plan containing measurable performance goals, identification of resources, performance indicators, and an evaluation plan to assess progress in developing the new reporting system. Without such a plan, neither DOD nor the Congress will be able to fully assess whether the new system's development is on schedule and achieving desired results.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-03-456, Military Readiness: New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed This is the accessible text file for GAO report number GAO-03-456 entitled 'Military Readiness: New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed' which was released on March 28, 2003. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products‘ accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. Report to the Chairman, Subcommittee on Readiness, Committee on Armed Services, House of Representatives: United States General Accounting Office: GAO: March 2003: Military Readiness: New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed: GAO Highlights: Highlights of GAO-03-456, a report to the Chairman, Subcommittee on Readiness, Committee on Armed Services, House of Representatives Why GAO Did This Study: The Department of Defense‘s (DOD) readiness assessment system was designed to assess the ability of units and joint forces to fight and meet the demands of the national security strategy. In 1998, GAO concluded that the readiness reports provided to Congress were vague and ineffective as oversight tools. Since that time, Congress added reporting requirements to enhance its oversight of military readiness. Therefore, the Chairman asked GAO to examine (1) the progress DOD made in resolving issues raised in the 1998 GAO report on both the unit- level readiness reporting system and the lack of specificity in DOD‘s Quarterly Readiness Reports to the Congress, (2) the extent to which DOD has complied with legislative reporting requirements enacted since 1997, and (3) DOD‘s plans to improve readiness reporting. What GAO Found: Since 1998, DOD has made some progress in improving readiness reporting ”particularly at the unit level”but some issues remain. For example, DOD uses readiness measures that vary 10 percentage points or more to determine readiness ratings and often does not report the precise measurements outside DOD. DOD included more information in its Quarterly Readiness Reports to the Congress. But quality issues remain”in that the reports do not specifically describe readiness problems, their effects on readiness, or remedial actions to correct problems. Nor do the reports contain information about funding programmed to address specific remedial actions. Although current law does not specifically require this information, Congress could use it for its oversight role. DOD complied with most, though not all, of the legislative readiness reporting requirements enacted by Congress in the National Defense Authorization Acts for Fiscal Years 1998-2002. For example, DOD * is now listing the individual units that have reported low readiness and reporting on the readiness of prepositioned equipment, as required by the fiscal year 1998 Act; * is reporting on 11 of 19 readiness indicators that commanders identified as important and that Congress required to be added to the quarterly reports in the fiscal year 1998 Act, but is not reporting on the other 8 readiness indicators; and * has not yet implemented a new comprehensive readiness reporting system as required in the fiscal year 1999 Act. As a result, Congress is not receiving all the information mandated by law. DOD issued a directive in June 2002 to establish a new comprehensive readiness reporting system that DOD officials said they plan to use to comply with the reporting requirements specified by Congress. The new system is intended to implement many of the recommendations included in a congressionally directed independent study for establishing such a system. However, the extent to which the new system will actually address the current system‘s shortcomings is unknown, because the new system is currently only a concept, and full capability is not scheduled until 2007. As of January 2003, DOD had not developed an implementation plan containing measurable performance goals, identification of resources, performance indicators, and an evaluation plan to assess progress in developing the new reporting system. Without such a plan, neither DOD nor the Congress will be able to fully assess whether the new system‘s development is on schedule and achieving desired results. What GAO Recommends: GAO made recommendations to improve readiness reporting and to develop an implementation plan to allow DOD and the Congress to gauge progress in developing DOD‘s new readiness reporting system. DOD did not agree with our recommendations. After reviewing its comments, we modified one recommendation but retained the others as originally stated. www.gao.gov/cgi-bin/getrpt?GAO-03-456. To view the full report, including the scope and methodology, click on the link above. For more information, contact Neal P. Curtin at (757) 552-8100 or curtinn@gao.gov. March 2003 GAO-03-456: Contents: This is a work of the U.S. Government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. It may contain copyrighted graphics, images or other materials. Permission from the copyright holder may be necessary should you wish to reproduce copyrighted materials separately from GAO‘s product. Letter: Results in Brief: Background: Some Progress in Readiness Reporting, but Some Issues Remain: DOD Has Not Fully Complied with All Legislative Requirements: Lack of Implementation Plan Could Hinder Development of New Readiness Reporting System: Conclusions: Recommendations for Executive Action: Agency Comments and our Response: Appendix I: Scope and Methodology: Appendix II: Comments from the Department of Defense: Figures: Figure 1: DOD‘s Readiness Assessment Process, as of January 2003: Figure 2: Timeline of Recommended, Required, and Planned Readiness Reporting Changes Since 1994: Abbreviations: ACTD: Advanced Concept Technology Demonstration: GSORTS: Global Status of Resources and Training System: OUSD P&R: Office of the Under Secretary of Defense for Personnel and Readiness: United States General Accounting Office: Washington, DC 20548: March 28, 2003: The Honorable Joel Hefley Chairman Subcommittee on Readiness Committee on Armed Services House of Representatives: The Department of Defense‘s (DOD) readiness assessment system was designed to assess the ability of units and joint forces to fight and meet the demands of the national security strategy. For more than a decade ending in 1998, various audit and oversight organizations questioned the thoroughness and reliability of DOD reports on the readiness of U.S. forces. Since 1998, Congress has added reporting requirements to enhance its oversight of military readiness. In doing so, Congress expressed concern over contradictions between assessments of military unit readiness in reports and observations made by military personnel in the field.[Footnote 1] DOD provides Congress a quarterly report that contains readiness information from several sources: the unit-level readiness assessment system; results of scenario-based assessments; and summaries of information briefed to senior DOD officials. We reviewed DOD‘s readiness assessment and reporting system in 1998 and concluded that the readiness reports provided Congress offered a vague description of readiness problems and remedial actions and therefore were not effective as oversight tools.[Footnote 2] Considering the concerns raised in our 1998 report, reporting requirements added by Congress since 1998, and the new national security strategy, you asked us to provide an updated assessment of DOD‘s readiness reporting. As agreed with your office, we examined (1) the progress DOD has made in resolving issues raised in our prior report on both the unit level readiness reporting system and the lack of specificity in the Department‘s Quarterly Readiness Reports to the Congress, (2) the extent to which DOD has complied with legislative reporting requirements enacted since 1997, and (3) DOD‘s plans to improve readiness reporting. In conducting this analysis, we compared current reported readiness data with legislative requirements and with data reported in 1998. We also identified DOD initiatives for improving readiness reporting. We conducted our review from June 2002 through January 2003 in accordance with generally accepted government auditing standards. (For a complete description of our methodology, see app. I.): Results in Brief: Since our 1998 report identifying limitations in readiness reporting, DOD has made some progress in improving readiness reporting, such as adding information on equipment cannibalization rates in its reports to Congress. Some issues, however, remain. Although DOD has improved its unit-level readiness reporting system, it still uses readiness measures that vary 10 percentage points or more to determine readiness ratings and often does not widely report the precise measurements outside DOD. Since 1998, DOD has included more information in its Quarterly Readiness Reports to the Congress, such as an annex presenting equipment cannibalization rates. However, some degradation in these reports has occurred. For example, DOD eliminated some previously provided information--such as the joint force readiness assessments-- and the reports still contain very broad statements of readiness issues and the remedial actions taken or planned to address readiness problems. Even though some report annexes contain data, the data are not adequately explained or related to the broad statements of readiness issues mentioned in the reports. Also, the reports do not contain information about funding that is programmed to address specific remedial actions. Although this information is not required by law, we believe it would be useful for Congress to understand the significance of the information in these reports for use in its oversight role. DOD has complied with most, though not all, of the legislative readiness reporting requirements enacted by Congress in the National Defense Authorization Acts for Fiscal Years 1998-2002. For example, as directed by the National Defense Authorization Act for Fiscal Year 1998, DOD is now reporting on the readiness of prepositioned equipment and DOD is listing individual units that have reported low readiness.[Footnote 3] For some provisions of the National Defense Authorization Acts for the period we examined, DOD provided some of the information specified in the acts. For example, DOD reports on some but not all of the readiness indicators Congress, in the fiscal year 1998 act, required be added to the quarterly reports.[Footnote 4] For a few requirements, DOD has yet to comply. For example, DOD has not implemented a new readiness reporting system, which the National Defense Authorization Act for Fiscal Year 1999 required to be implemented by April 1, 2000.[Footnote 5] As a result, Congress is not receiving all the information mandated by law. DOD‘s main effort to improve readiness reporting is to develop and implement a new comprehensive readiness reporting system in response to the aforementioned legislation. However, the system‘s completion is still several years away. DOD issued a directive in June 2002 to establish this new system.[Footnote 6] Officials from the Office of the Under Secretary of Defense for Personnel and Readiness (OUSD P&R) responsible for developing the system stated they plan to use the new system to comply with the reporting requirements contained in the National Defense Authorization Acts and that the new system is intended to implement many of the recommendations included in a congressionally directed independent study for establishing a comprehensive readiness reporting system. However, the extent to which the new system will actually address the current system‘s shortcomings is unknown. This is because the new system is currently only a concept and full capability is not scheduled until 2007--some 7 years after a new comprehensive system was legislatively mandated to be implemented.[Footnote 7] As of January 2003 DOD had not developed an implementation plan to assess progress in developing the new reporting system. We believe that an implementation plan containing measurable performance goals, identification of resources, performance indicators, and an evaluation plan could help DOD as well as the Congress assess whether the new system‘s development is on schedule and achieving desired results. We are making recommendations to improve readiness reporting and to develop an implementation plan to allow DOD and the Congress to gauge DOD‘s progress in developing its new readiness reporting system. In commenting on a draft of this report, DOD did not agree with our recommendation to improve readiness reporting, saying that the quarterly reports are comprehensive and spending more time on the report would be counterproductive. Although the quarterly reports contain voluminous data, the reader cannot assess its significance since the data are explained inadequately. However, between 1998 and mid-2001, DOD did include an unclassified summary of issues for each service in the quarterly reports addressing several topics, including personnel and equipment. Since DOD has highlighted key issues in past quarterly reports, we believe that improving these reports by again including a summary that highlights key readiness issues would be beneficial. In doing so, DOD could focus on the most critical issues that are of greatest concern to the services. Therefore, we have modified our recommendation that DOD improve the quality of readiness reporting to focus on issues deemed to be critical by the Secretary and the services, including analyes and planned remedial actions for each issue. DOD also did not agree with our recommendations to develop an implementation plan for the new readiness system and to provide Congress annual updates on the new system‘s development. DOD said that it is developing better tools for assessing readiness and had established milestones and expected outcomes. Thus, DOD believes that further planning and providing an annual update to Congress is unnecessary. Considering that Congress expressed concern about DOD‘s lack of progress in developing a comprehensive system and that DOD does not plan for the new system to be fully capable until 2007, we retained these two recommendations. A detailed discussion of DOD‘s comments and our response is contained in the body of this report. Background: DOD‘s readiness assessment and reporting system was designed to assess and report on military readiness at three levels--(1) the unit level; (2) the joint force level; and (3) the aggregate, or strategic, level. Unit-level readiness is assessed with the Global Status of Resources and Training System (GSORTS), which is an automated system that assesses the extent to which military units possess the required resources and training to undertake their wartime missions. To address joint readiness, the Chairman of the Joint Chiefs of Staff established the Joint Monthly Readiness Review (now called the Joint Quarterly Readiness Review or JQRR), that compiles readiness assessments from the combatant commands, the combat support agencies, and the military services. The Joint Staff and the services use these assessments to brief DOD‘s leadership on the Senior Readiness Oversight Council--an executive-level forum for monitoring emerging readiness issues at the strategic level. The briefings to the council are intended to present a view of readiness at the aggregate force level. From these briefings to the council, DOD prepares a legislatively mandated quarterly readiness report to Congress.[Footnote 8] Figure 1 provides an overview of DOD‘s readiness assessment process. Figure 1: DOD‘s Readiness Assessment Process, as of January 2003: [See PDF for image] [End of figure] We have issued several reports containing recommendations for improving readiness reporting. In 1994, we recommended DOD develop a more comprehensive readiness system to include 26 specific readiness indicators.[Footnote 9] In 1998, we reported on shortcomings in DOD‘s readiness assessment system. At that time, we stated GSORTS‘ limitations included lack of precision in measurements, late reporting, subjective input, and lack of standardization. Secondly, we reported that while the Quarterly Readiness Reports to the Congress accurately reflected briefs to the Senior Readiness Oversight Council, they lacked specific details on deficiencies and remedial actions and thus did not meet the requirements of 10 U.S.C. 482 (b). DOD concurred with our recommendation that the Secretary of Defense take steps to better fulfill the legislative reporting requirements under 10 U.S.C. 482 by providing (1) supporting data on key readiness deficiencies and (2) specific information on planned remedial actions. Finally, we reported that deficiencies identified as a result of the Joint Monthly Readiness Reviews remained open because the solutions require funding over the long term. In 2002, we issued a classified report on DOD‘s process for tracking the status of deficiencies identified in the Joint Monthly Readiness Reviews. We made recommendations to improve DOD‘s deficiency status reporting system and for DOD to develop funding estimates for correcting critical readiness deficiencies. In its comments, DOD generally agreed with the report‘s findings and recommendations. Some Progress in Readiness Reporting, but Some Issues Remain: Although DOD has made progress in resolving readiness reporting issues raised in our 1998 report, we found that some of the same issues still exist today. For example, DOD has added information to its Quarterly Readiness Reports to the Congress (hereafter referred to as the quarterly reports). However, we found that the reports still contain vague descriptions of readiness problems and remedial actions. Even though some report annexes contain detailed data, the data as presented are not ’user friendly“--it is largely unevaluated and is not linked to readiness issues mentioned in the report plus the report text does not explain how the data relates to units‘ readiness. Thus, as we reported in 1998, these reports do not specifically describe readiness problems or remedial actions as required under 10 U.S.C. 482 (b). We believe that this kind of information would be useful for Congress to understand the significance of the information in these reports for use in its oversight role. Some Improvements in Unit-Level Reporting, but Some Issues Unchanged Since 1998: DOD has improved some aspects of its unit-level reporting system, the Global Status of Resources and Training System (GSORTS). For example, in 1998 GSORTS‘ data were maintained in multiple databases and data were not synchronized. As of September 2002, the data are reported to a central site, and there is one database of record. Also in 1998, U.S. Army GSORTS review procedures delayed submission of Army data, and all the services‘ data entry was manual. As of September 2002, Army reporting procedures require reporting consistent with GSORTS‘ requirements, and all the services have automated data entry, which reduces errors. In 1998, combat units only reported on readiness for wartime missions. As of September 2002, combat units report on assigned mission readiness in addition to wartime mission readiness. Conversely, DOD has not resolved some issues we raised in 1998. For example, readiness ratings are still reported in broad bands and actual percentages of required resources are not externally reported. These issues remain because the manual specifying readiness reporting rules has not changed in these areas.[Footnote 10] The manual‘s definition of readiness levels for personnel has not changed since our 1998 report-- it still defines readiness levels in bands of 10 percentage points or more and does not require external reporting of actual percentages. For example, the highest personnel rating can range from 90 percent to 100 percent, and there is no requirement to report the actual percentage outside of DOD. We have also reported that GSORTS does not always reflect training and equipment deficiencies. For example, we reported in April and June 2002 that readiness data do not reflect the effect of training range restrictions on unit readiness.[Footnote 11] We have also reported that GSORTS does not include whether a unit‘s chemical/ biological equipment is usable.[Footnote 12] In commenting on our analysis, the OUSD P&R office responsible for readiness reporting stated that it recognized the imprecision of the current measurements. According to that office, an effort to develop the planned new readiness reporting system, which is discussed later in this report, includes working with the DOD components to enhance and expand readiness reporting. Some Information Added to Quarterly Reports Since 1998, but Other Information Eliminated: Since our 1998 report, the quarterly reports improved in some areas, but degraded in others. Although some information was added, we found that some of the same quality issues remain--namely, that the reports do not specifically describe readiness problems, their effects on readiness, or remedial actions. DOD has added information to the quarterly reports in response to legislative direction. For example, DOD added information on the services‘ cannibalization rates.[Footnote 13] Also, DOD added annual reports on infrastructure and institutional training readiness.[Footnote 14] However, some information was eliminated from the quarterly reports. For example, the law requires results of joint readiness reviews to be reported to Congress.[Footnote 15] DOD included these results until the July-September 2001 Quarterly Readiness Report to the Congress. Since that report, four quarterly reports have been issued without the joint force assessments. Defense officials responsible for readiness reporting said that the joint readiness reviews were not included because the scenarios were based on the former national security strategy of two major wars. The officials stated they plan to include results from the joint readiness reviews in future reports. In commenting on our analysis the OUSD P&R office responsible for readiness reporting stated that it continues to seek better ways to provide concise, quality information. Quarterly Reports Do Not Adequately Explain Readiness Issues: As we reported in 1998, we found that the quarterly reports still contain broad statements of readiness issues and remedial actions, which are not supported by detailed examples and are not related to data in the reports‘ annexes. Among other things, the law requires the quarterly reports to specifically describe each readiness problem and deficiency as well as planned remedial actions.[Footnote 16] The reports did not specifically describe the nature of each readiness problem or discuss the effects of each on unit readiness. Also, the reports included only broad statements of remedial actions that lacked details on timelines, objectives, or funding requirements. For example, one report said that the Air Force continued to experience shortages in critical job skills that affected the service‘s ability to train. The report did not refer the reader to data in annexes showing training readiness ratings; it did not state which skills were short, which training was not accomplished, or whether this shortage had or was expected to affect units‘ readiness ratings. Further, the report did not explain the remedial actions taken or planned to reverse the skill shortage, how long it would take to solve this problem, or what funding was programmed to implement remedial actions. Defense readiness officials agreed, stating that information in the quarterly reports is summarized to the point that there are no details on readiness deficiencies, remedial actions, or funding programmed to implement remedial actions. We believe the Congress needs this type of information to understand the significance of the information reported. Although some of the quarterly report annexes contain voluminous data, the data are not adequately explained or related to units‘ readiness. The law does not mandate specific explanations of these ’readiness indicators,“ but we believe it is essential for Congress to understand the significance of the information in these reports for use in its oversight role. For example, DOD is required to report on the maintenance backlog.[Footnote 17] Although the report provides the quantity of the backlog, it does not explain the effect the backlog had on readiness. Specifically, the report did not explain whether units‘ readiness were affected because maintenance was not accomplished when needed. In addition, DOD is required to report on training commitments and deployments.[Footnote 18] The Expanded Quarterly Readiness Report to Congress Implementation Plan dated February 1998 stated that ’either an excessive or a reduced level of commitment could be an indicator of potential readiness problems.“ However, OUSD P&R did not define what kind of ’readiness problems“ this data may indicate would occur as a result of ’excessive or reduced“ levels of training and deployments, such as degraded equipment or training. The data reported are the amount of training away from home station and the amount of deployments. However, these data are not explained or related to a unit‘s equipment or training ratings. Further, criteria have not been established to distinguish between acceptable and unacceptable levels of the training and deployment data reported. As a result, the reader does not know whether the data reported indicate a problem or the extent of the problem. In commenting on our analyses, OUSD P&R acknowledged ’the Department would be better served by providing more information as to how various data relates to readiness.“: Generally, the quarterly reports also do not contain information on funding programmed to implement specific remedial actions. For example, one quarterly report included the statement that budgets were revised ’to address readiness and capabilities issues,“ but no examples were provided. Also, the report lacked statements explaining how this ’budget revision“ would improve readiness. Although not required by law, we believe it would prove useful for Congress to understand how DOD addresses specific readiness problems. In commenting on our analysis, OUSD P&R officials stated that they would continue to work with the services to provide more fidelity with the information presented in the quarterly report annexes. However, they also said that detailed examples require significant staff effort throughout DOD and that the added time for more detailed analysis could render the report a historical document. They further said that complete information would certainly be desired and agreed it is important for the Congress to understand the significance of the information in the quarterly reports for use in its oversight role. DOD Has Not Fully Complied with All Legislative Requirements: DOD has complied with most, but not all, of the readiness reporting requirements added by Congress in the National Defense Authorization Acts for Fiscal Years 1998 through 2002.[Footnote 19] Congress added readiness reporting requirements out of concern over contradictions between assessment of military unit readiness in official readiness reports and observations made by military personnel in the field.[Footnote 20] In a review of these acts, we identified both recurring readiness reporting requirements that were added to existing law and one-time reporting requirements related to military readiness. We compared current readiness reporting to the requirements in these acts to make an overall judgment on the extent of compliance. We did not develop a total count of the number of reporting requirements because the acts included a series of sections and subsections that could be totaled in various ways. Because DOD is not reporting on all the requirements added over the past several years, the Congress is not receiving all the information mandated by law. DOD Is Complying with Most of the New Requirements: Our analysis showed that DOD has complied with most of the requirements added in the National Defense Authorization Acts for Fiscal Years 1998- 2002. For example, DOD took the following actions in response to legislative requirements: * DOD is now reporting on the readiness of prepositioned equipment and is listing individual units that have reported low readiness as required by the National Defense Authorization Act for Fiscal Year 1998.[Footnote 21] * DOD is reporting on infrastructure and institutional training readiness as required by the National Defense Authorization Act for Fiscal Year 1999.[Footnote 22] * DOD contracted for an independent study of requirements for a comprehensive readiness reporting system and submitted the study report to the Congress as required by the National Defense Authorization Act for Fiscal Year 2000.[Footnote 23] * DOD has added quarterly information on the military services‘ cannibalization rates as required by the National Defense Authorization Act for Fiscal Year 2001.[Footnote 24] DOD Is Not Reporting on All Items Required: DOD is reporting on some, though not all, of the items Congress required be added to the quarterly readiness reports. For example, the National Defense Authorization Act for Fiscal Year 1998 required 19 specific items be reported that are consistent with our previously cited 1994 report on readiness reporting.[Footnote 25] The 1994 report included a list of 26 readiness indicators that DOD commanders said were important for a more comprehensive assessment of readiness. A 1994 DOD-funded study by the Logistics Management Institute found that 19 of the 26 indicators could help DOD monitor critical aspects of readiness. The 19 items listed in the National Defense Authorization Act for Fiscal Year 1998 are very similar to those identified in the 1994 Logistics Management Institute study. DOD is reporting on 11 of the 19 items and is not reporting on the other 8. The eight items are (1) historical personnel strength data and trends, (2) personnel status, (3) borrowed manpower, (4) personnel morale, (5) operations tempo, (6) training funding, (7) deployed equipment, and (8) condition of nonpacing equipment[Footnote 26] as required in the Act.[Footnote 27] In an implementation plan setting forth how it planned to comply with reporting on the 19 items, which was also required by the National Defense Authorization Act for Fiscal Year 1998,[Footnote 28] DOD stated that it would not report on these eight indicators for the following reasons: * Deployed equipment was considered part of the status of prepositioned equipment indicator. * Historical personnel strength data and trends were available from the annual Defense Manpower Requirements Report. * Training funding and operations tempo were believed to be represented adequately in the budget requests as flying hours, steaming days, or vehicle miles and were not considered good measures of readiness output. * Personnel strength status was considered to be part of the personnel rating, but DOD agreed to investigate other ways to evaluate the effect of service personnel working outside the specialty and grade for which they were qualified. * Borrowed manpower data was only captured in a limited sector of the deployable force and may not be meaningful until a better method is developed to capture the data. * Personnel morale had no existing data sources. * The condition of nonpacing equipment had no reasonable measurement to use as an indicator. Notwithstanding the reasoning that DOD stated, these eight indicators continue to be required by law, and we saw no indication in our work that DOD is working to develop data for them. Also, DOD is not complying with some of the requirements in the National Defense Authorization Act for Fiscal Year 1999. Examples are as follows: * The act required DOD to establish and implement a comprehensive readiness reporting system by April 2000.[Footnote 29] As of January 2003, DOD had not implemented a new system, and officials said it is not expected to be fully capable until 2007 or 7 years later than required. * The act also required DOD to develop implementing regulations for the new readiness reporting system.[Footnote 30] DOD had not developed implementing regulations as of January 2003. * The act required DOD to issue regulations for reporting changes in the readiness of training or defense infrastructure establishments within 72 hours.[Footnote 31] Although DOD has provided some guidance, officials stated they have not issued regulations because no mechanism exists for institutional training or defense infrastructure establishments to report changes and because these entities are not part of an established readiness reporting system. In commenting on our analyses, DOD officials acknowledged ’the Department is not in full compliance“ and stated that they plan to achieve compliance with the new readiness reporting system under development. OUSD P&R officials said that the shortfalls in reporting are unwieldy under the current system; OUSD P&R intends to correct these shortfalls when the new system is functional. However, as noted above, DOD does not plan to implement its new system until 2007. As of January 2003, DOD also had not targeted incremental improvements in readiness reporting during the period in which the new system is being developed. Until then, Congress will receive less readiness information than it mandated by law. Lack of Implementation Plan Could Hinder Development of New Readiness Reporting System: DOD issued a directive in June 2002 to establish a new readiness reporting system. The Undersecretary of Defense for Personnel and Readiness is to oversee the system to ensure the accuracy, completeness, and timeliness of its information and data, its responsiveness, and its effective and efficient use of modern practices and technologies. Officials in the OUSD P&R readiness office responsible for developing the new system said that they plan to use the new system to comply with the requirements in the National Defense Authorization Acts and to address many of the recommendations contained in a congressionally directed independent study. However, as of January 2003, there are few details of what the new system would include. Although the new system may have the potential to improve readiness reporting, as of January 2003, it is only a concept without detailed plans to guide development and monitor implementation. As a result, the extent to which the new system will address existing shortcomings is unknown. DOD Directive Establishes a New Readiness Reporting System: The National Defense Authorization Act for Fiscal Year 1999 required DOD to establish a comprehensive readiness reporting system.[Footnote 32] In doing so, the Congress expressed concern about DOD‘s lack of progress in developing a more comprehensive readiness measurement system reflective of operational realities. The Congress also noted that past assessments have suffered from DOD‘s inability to create and implement objective and consistent readiness reporting criteria capable of providing a clear picture to senior officials and the Congress.[Footnote 33] Subsequently, the August 2001 Defense Planning Guidance for Fiscal Years 2003-2007 called for the development of a strategic plan for transforming DOD readiness reporting. In June 2002, DOD issued a directive establishing the Department of Defense Readiness Reporting System.[Footnote 34] The system will measure and report on the readiness of military forces and the supporting infrastructure to meet missions and goals assigned by the Secretary of Defense. All DOD components will align their readiness reporting processes in accordance with the directive. The directive assigns oversight and implementing responsibility to the Undersecretary of Defense for Personnel and Readiness. The Undersecretary is responsible for developing, fielding, maintaining, and funding the new system and scenario assessment tools. The Undersecretary--in collaboration with the Joint Chiefs of Staff, Services, Defense Agencies, and Combatant Commanders--is to issue implementing instructions. The Chairman of the Joint Chiefs of Staff, the Service Secretaries, the commanders of the combatant commands, and the heads of other DOD components are each assigned responsibilities related to readiness reporting. OUSD P&R established a timetable to implement the new readiness reporting system. OUSD P&R plans to achieve initial capability in 2004 and reach full capability in 2007. OUSD P&R officials involved in developing the system said that they have been briefing the concept for the new reporting system since October 2002. As of January 2003 these officials stated that they are continuing what they have termed the ’concept demonstration“ phase, which began in October 2002. This phase consists of briefing various offices within DOD, the Joint Staff, and the services to build consensus and refine the new system‘s concept. These officials also said that the new system will incorporate many, but not all, of the recommendations contained in a legislatively mandated independent study of readiness reporting, which concluded that improvements were needed to meet legislatively mandated readiness reporting requirements and included numerous recommendations for what a new system should include.[Footnote 35] For example, the study recommended that (1) DOD report on all elements essential to readiness, such as depots, combat support agencies, and Defense agencies; (2) reporting should be in terms of mission essential tasks; and (3) DOD should measure the capability to carry out the full range of National Security Strategy requirements--not just a unit‘s wartime mission. An Implementation Plan Would Help Gauge Program Progress: We believe that successfully developing and implementing a large-scale effort, such as DOD‘s new readiness reporting system, requires an implementation plan that includes measurable performance goals, identification of resources, performance indicators, and an evaluation plan. As discussed earlier, full implementation of DOD‘s new readiness reporting system is several years away, and much remains to be done. In January 2003 the OUSD P&R office responsible for developing the new system said that the new readiness reporting system is a large endeavor that requires buy-in from many users and that the development of the system will be challenging. This office also wrote that it had just been given approval to develop the new readiness reporting system, was targeting development of an implementing instruction in the March 2003 time frame, and had not developed an implementation plan to assess progress in developing and implementing the new reporting system. The directive establishing the new reporting system requires the Undersecretary of Defense for Personnel and Readiness, in collaboration with others, to issue implementing instructions for the new system. DOD has experienced delays in implementing smaller readiness improvements than envisioned in the new readiness reporting system. One such effort involved development of an interface to query the existing readiness data base (GSORTS). In a July 2002 report, the DOD Inspector General reported that the planned implementation of this interface slipped 44 months, or just over 3.5 years.[Footnote 36] Also, history has shown it takes DOD time to make changes in the readiness reporting system. As illustrated in figure 2, DOD began reporting on specific readiness indicators 4 years after it agreed with GAO recommendations to include them in readiness reporting (see fig. 2). Figure 2: Timeline of Recommended, Required, and Planned Readiness Reporting Changes Since 1994: [See PDF for image] [End of figure] Other DOD development efforts recognize the need for effective planning to guide development. For example, DOD is working to transform military training as directed by the Defense Planning Guidance for Fiscal Years 2003-07. A March 2002 Strategic Plan for Transforming DOD Training developed by a different office within OUSD P&R discusses a training transformation road map with major tasks subdivided into near-, mid-, and long-term actions. The plan includes a list of near-term actions to be completed by October 2003 and definition of mid-and long-term actions in a comprehensive implementation plan that will identify specific tasks, responsibilities, timelines, resources, and methods to assess completion and measure success. The May 2002 Defense Planning Guidance update for fiscal years 2004-2009 directs OUSD P&R, working with other DOD components, to develop a comprehensive program to implement the strategic training transformation plan and provide it to the Deputy Secretary of Defense by April 1, 2003. Since the directive for creating a new readiness reporting system established broad policy with no specifics and since DOD has not developed an implementation plan, the extent to which the new system will address the current system‘s shortcomings will remain unknown until the new system is fully capable in 2007. Until then, readiness reporting will continue to be based on the existing system. Commenting on its plans for the new system, OUSD P&R said that it is in the process of creating an Advanced Concept Technology Demonstration (ACTD) structure for the new system and will produce all necessary planning documents required within the established ACTD process. However, this process is intended to provide decision makers an opportunity to understand the potential of a new concept before an acquisition decision. We do not believe the ACTD process will necessarily result in an implementation plan to effectively monitor development and assess whether the new system is being implemented on schedule and achieving desired results. DOD‘s ACTD guidelines state the principal management tool for ACTDs is a management plan, which provides a top-level description of the objectives, critical events, schedule, funding, and measures of evaluation for the project. We reported in December 2002 that these guidelines contain advice and suggestions as opposed to formal directives and regulations.[Footnote 37] DOD‘s guidelines state that the ACTD should plan exercises or demonstrations to provide an adequate basis for utility assessment. We also reported in December 2002 that DOD lacks specific criteria to evaluate demonstration results, which may cause acquisition decisions to be based on too little knowledge. Therefore, we still believe an implementation plan is necessary since the ACTD process does not require a detailed implementation plan and does not always include specific criteria to evaluate effectiveness. Conclusions: While DOD has made some improvements in readiness reporting since 1998, some of the same issues remain unresolved today. Although DOD is providing Congress more data than in 1998, the voluminous data are neither evaluated nor explained. The quarterly reports do not link the effects of ’readiness issues“ or deficiencies to changes in readiness at the unit level. Also, as in 1998, the reports contain vague descriptions of remedial actions not linked to specific deficiencies. Finally, the quarterly reports do not discuss funding that is programmed to implement specific remedial actions. As a result, the information available to Congress is not as effective as it could be as an oversight tool. Even though DOD directed development of a new readiness reporting system, it has not yet developed an implementation plan identifying objective and measurable performance goals, the resources and personnel needed to achieve the goals, performance indicators, and an evaluation plan to compare program results with goals, and milestones to guide overall development of the new readiness system. Even though the new system may have the potential to improve readiness reporting, without an implementation plan little assurance exists that the new system will actually improve readiness assessments by the time full capability is planned in 2007. Without such a plan, it will also remain difficult to gauge progress toward meeting the 2007 target date. This concern is reinforced in light of the (1) years-long delays in implementing other readiness reporting improvements and (2) the deficiencies in existing reporting that OUSD P&R plans to rectify with the new system. Furthermore, without an implementation plan neither senior DOD leadership nor the Congress will be able to determine if the resources spent on this system are achieving their desired results. Recommendations for Executive Action: To improve the information available to Congress for its use in its oversight role, we recommend that the Secretary of Defense direct the OUSD P&R to improve the quality of information contained in the quarterly reports. Specifically, we recommend that DOD‘s reports explain (in the unclassified section) the most critical readiness issues that are of greatest concern to the department and the services. For each issue, we recommend that DOD‘s reports include: * an analysis of the readiness deficiencies, including a clear explanation of how the issue affects units‘ readiness; * a statement of the specific remedial actions planned or implemented; and: * clear statements of the funding programmed to implement each remedial action. To be able to assess progress in developing the new readiness system, we recommend that the Secretary of Defense direct the OUSD P&R to develop an implementation plan that identifies: * performance goals that are objective, quantifiable, and measurable; * the cost and personnel resources needed to achieve the goals, including an identification of the new system‘s development and implementation costs in the President‘s Budget beginning in fiscal year 2005 and Future Years Defense Plan; * performance indicators to measure outcomes; * an evaluation plan to compare program results with established goals; and: * milestones to guide development to the planned 2007 full capability date. To assist Congress in its oversight role, we recommend that the Secretary of Defense give annual updates to the Congress on the new readiness reporting system‘s development to include: * performance measures, * progress toward milestones, * comparison of progress with established goals, and: * remedial actions, if needed, to maintain the implementation schedule. Agency Comments and our Response: In written comments on a draft of this report, which are reprinted in appendix II, the Department of Defense did not agree with our recommendations. In response to our recommendation that DOD improve the quality of information contained in its quarterly readiness reports, DOD said that the Quarterly Readiness Report to the Congress is one of the most comprehensive and detailed reports submitted to the Congress that discusses serious readiness issues and ways in which these issues are being addressed. DOD further stated that the department presents briefings on specific readiness issues to the Congress and that spending more time and resources expanding the existing written report would be counterproductive. We recognize that the Quarterly Readiness Reports to the Congress contain voluminous data. However, as discussed in this report, we found that the quarterly reports‘ annexes are large and mostly consist of charts or other data that are not adequately explained and are not related to units‘ readiness. In some cases, criteria have not been established to enable the reader to distinguish between acceptable and unacceptable levels of the data reported. As a result, the reader cannot assess the significance of the data because it is not at all clear whether the data reported indicate a problem or the extent of the problem. Considering that the quarterly reports contain inadequately explained data and that much of the information is not ’user friendly,“ we continue to believe the quality of information in the quarterly reports can be improved. In fact, we reviewed all the quarterly reports provided to Congress since 1998 and found that through the January-June 2001 report[Footnote 38] the reports did include an unclassified summary of readiness issues for each service addressing four topics-- personnel, equipment, training, and enablers (critical units or capabilities, such as specialized aircraft, essential to support operations). However, the reports did not include supporting data or a discussion of remedial actions. Since that time, these summaries have been eliminated from the quarterly reports. For example, the unclassified narrative of the last two reports available at the time we performed our work--January-March 2002 and April-June 2002--were less than two pages long and neither discussed readiness issues nor ways in which these issues are being addressed. One report discussed the new readiness reporting system, and the other discussed a review of seven environmental laws. Given that DOD has highlighted key issues in the past, we believe that improving the quarterly reports would be beneficial if DOD were to focus on the most critical readiness issues that are of greatest concern to the services and includes supporting data and a discussion of remedial actions. Therefore, we have modified our recommendation that DOD improve the quality of readiness reporting to focus on readiness issues deemed to be critical by the Secretary and the military services and to provide more detailed data and analyses of those issues and the remedial actions planned for each one. DOD did not agree with our recommendations that it (1) develop an implementation plan with, among other things, performance goals that are objective, quantifiable, and measurable and (2) provide annual updates to the Congress on the new readiness reporting system‘s development. DOD said that it had undertaken an initiative to develop better tools for assessing readiness and that it intended to apprise Congress on its efforts to develop tools for readiness assessment. DOD further stated that the effort to improve readiness reporting is in its infancy, but that it has established milestones, cost estimates, functional responsibilities, and expected outcomes. DOD believes that further planning and a prescriptive annual update to the Congress is unnecessary. We agree that the new readiness reporting system may have the potential to improve readiness reporting. However, as discussed in this report, the directive establishing the new system contains very broad, high- level statements of overall functional responsibilities and outcomes, but no details on how these will be accomplished. Further, DOD has established two milestones--initial capability in 2004 and full capability in 2007. DOD does not have a road map explaining the steps needed to achieve full capability by 2007, which is seven years after Congress mandated a new system be in place. In addition, as discussed earlier in this report, DOD has experienced delays in implementing much smaller readiness improvements. While DOD has undertaken an initiative to develop better tools for assessing readiness and intends to routinely and fully apprise the Congress on its development efforts, tools are the mechanics for evaluating readiness data. As such, tools are not the same thing as the comprehensive readiness reporting system mandated by Congress that DOD has said will include new metrics and will evaluate entities within DOD that currently do not report readiness. Considering that Congress expressed concern about DOD‘s lack of progress in developing a comprehensive system, that developing and implementing DOD‘s planned new system is scheduled to take 4 more years, and that delays have been experienced in earlier efforts to make small improvements in readiness reporting, we continue to believe that it is important for DOD to develop an implementation plan to gauge progress in developing and implementing the new readiness reporting system and to provide annual updates to the Congress. Such a plan would be consistent with DOD‘s approach to other major initiatives such as transforming training. We have therefore retained these two recommendations. DOD also provided technical corrections and we have modified the report where appropriate. We are sending copies of this report to the Ranking Minority Member, Subcommittee on Readiness, House Committee on Armed Services; the Chairman and Ranking Minority Member, Subcommittee on Readiness and Management Support, Senate Committee on Armed Services; other interested congressional committees; Secretary of Defense; and the Director, Office of Management and Budget. We will also make copies available to others on request. In addition, the report will be available at no charge on the GAO Web site at http://www.gao.gov. If you or your staff have any questions, please call me on (757) 552- 8111 or by E-mail at curtinn@gao.gov. Major contributors to this report were Steven Sternlieb, Brenda Waterfield, James Lewis, Dawn Godfrey, and Herbert Dunn. Sincerely yours, Neal Curtin Director Defense Capabilities and Management: Signed by Neal Curtin: [End of section] Appendix I: Scope and Methodology: To assess the progress the Department of Defense (DOD) has made in resolving issues raised in our prior report concerning both the unit level readiness reporting system and the lack of specificity in DOD‘s Quarterly Readiness Reports to the Congress, we met with DOD officials and reviewed regulations and quarterly reports. Specifically, we met with officials of the Office of the Undersecretary of Defense for Personnel and Readiness (OUSD P&R) responsible for readiness reporting, the Joint Staff, and the military services to discuss their individual progress in each of these areas. To assess progress regarding unit level readiness reporting, we reviewed the Chairman of the Joint Chiefs of Staff manual governing this system and the related service implementing instructions to determine if these documents had changed since our 1998 report or if the manual and service instructions continued to allow reporting in the same manner as reflected in our earlier report. Through a comparison of the current and prior documents, discussions with pertinent officials, and our analysis, we determined whether the readiness reporting issues we raised in 1998 had been resolved. We also reviewed the content of quarterly reports to assess their quality and usefulness, and assess whether the problems we reported in 1998 had been rectified. We discussed our analysis with OUSD P&R officials and provided them with our analyses in order that they could fully consider and comment on our methodology and conclusions. We did not assess the accuracy of reported readiness data. To determine the extent to which DOD has complied with legislative reporting requirements enacted since our prior report, we compared a complete listing of these requirements to DOD‘s readiness reporting. First, we identified the legislatively mandated readiness reporting requirements enacted since our 1998 report. To accomplish this, we reviewed the National Defense Authorization Acts for Fiscal Years 1998- 2002 to list the one-time and recurring reporting requirements related to military readiness. We also requested congressional staff and OUSD P&R to review the list, and officials from both offices agreed it was accurate. We did not develop a total count of the number of reporting requirements because the acts included a series of sections and subsections that could be totaled in various ways. Once we obtained concurrence that this listing was complete and accurate, we compared this list to current readiness reporting to make an overall judgment on the extent of compliance. To assess how DOD plans to improve readiness reporting, we reviewed the June 2002 DOD directive establishing a new readiness reporting system and a progress update briefing on the new system. We also obtained readiness briefings from each of the services, OUSD P&R, and Joint Staff officials. We performed several electronic searches of the Deputy Under Secretary of Defense (Readiness) electronic Web site to determine the status of readiness reporting. To assess how smoothly other readiness improvements progressed, we reviewed DOD audit reports. We discussed our findings with OUSD P&R officials and worked proactively with them in conducting our analyses. Specifically, we provided them drafts of our analyses for their comments and corrections. We conducted our review from June 2002 through January 2003 in accordance with generally accepted government auditing standards. [End of section] Appendix II: Comments from the Department of Defense: OFFICE OF THE UNDER SECRETARY OF DEFENSE: 4000 DEFENSE PENTAGON WASHINGTON, D.C. 20301-4000: MAR 21 2003: PERSONNEL AND READINESS: Mr. Neal P. Curtin: Director, Defense Capabilities and Management U. S. General Accounting Office: Washington, D.C. 20548: Dear Mr. Curtin: This is the Department of Defense response to the General Accounting Office draft report, MILITARY READINESS: New Reporting System Is Intended to Address Longstanding Problems, but Better Planning Is Needed,“ (GAO-03-456/GAO Code 350222). We have reviewed the draft report and do not concur with the recommendations. We also note a number of errors and will forward our technical corrections separately. We appreciate the opportunity to review and comment on this report. Sincerely, Joseph J. Angello, Jr. Director Readiness Programming and Assessment: Signed by Joseph J. Angello, Jr.: Attachment: --GAO DRAFT REPORT - DATED FEBRUARY 18, 2003 GAO CODE 350222/GAO-03- 456: ’MILITARY READINESS: New Reporting System Is Intended to Address Longstanding Problems, but Better Planning Is Needed“: DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS: RECOMMENDATION 1: To improve the information available to Congress for its use in its oversight role, the GAO recommended that the Secretary of Defense direct the OUSD P&R to improve the quality of information contained in the quarterly reports. Specifically, the GAO recommended that DoD‘s reports include: * analyses of readiness deficiencies, including a clear explanation of how the deficiency affects units‘ readiness; * a statement of the specific remedial actions planned or implemented for each deficiency; and: * clear statements of the funding programmed to implement each remedial action. (p. 22/GAO Draft Report): DOD RESPONSE: We cannot agree with this recommendation. The Quarterly Readiness Report to Congress is one of the most comprehensive and detailed reports submitted to the Congress on a routine basis. The latest report was 289 pages, covering numerous readiness indicators, pressing readiness issues, meeting summaries, and unit readiness trends. Our reports routinely discuss serious readiness issues within the Department of Defense and ways in which these issues are being addressed. This recommendation calls for the report to capture readiness ’deficiencies,“ how these deficiencies affect unit readiness, and the resource implications to fix them. Given the Department has roughly 10,000 units reporting readiness information, this tasking is clearly infeasible. The Department does, however, routinely present detailed and timely briefings on specific readiness issues and concerns to the Congress. For example, our recent quarterly reports included issues and actions associated with the Global War on Terror and potential follow-on operations; preparedness for Chemical Biological defense; and the management of low density/high demand assets. Spending more time and resources expanding the existing written report would be counterproductive, and ultimately provide less timely and meaningful information to Congress. RECOMMENDATION 2: To be able to assess progress in developing the new readiness system, the GAO recommended that the Secretary of Defense direct the OUSD P&R to develop an implementation plan that identifies: * performance goals that are objective, quantifiable, and measurable; * the cost and personnel resources needed to achieve the goals, including an identification of the new system‘s development and implementation costs in the President‘s Budget beginning in fiscal year 2005 and Future Years Defense Plan; * performance indicators to measure outcomes; * an evaluation plan to compare program results with established goals; and: * milestones to guide development to the planned 2007 full operational capability date. (p. 22/GAO Draft Report): DOD RESPONSE: We cannot concur with this recommendation. The Department has undertaken an initiative to develop better tools for assessing the readiness of our military forces and supporting infrastructure to meet the threats facing our nation. This effort is in its infancy, but we have established milestones, cost estimates, functional responsibilities, and expected outcomes. We also have detailed plans for specific information technology requirements and progress indicators. More detailed and voluminous planning on top of the current planning is unproductive and unnecessary for this effort. RECOMMENDATION 3: To assist Congress in its oversight role, the GAO recommended that the Secretary of Defense provide annual updates to the Congress on the new readiness reporting system‘s development to include: * performance measures, * progress toward milestones, * comparison of progress with established goals, and: * remedial actions, if needed, to maintain the implementation schedule. (p. 23/GAO Draft Report): DOD RESPONSE: We disagree with this recommendation. The Department intends to routinely and fully apprise the Congress on our effort to develop tools supporting readiness assessment. A prescriptive ’annual“ update is neither necessary nor warranted for this effort. [End of section] FOOTNOTES [1] Public Law 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281 (1998); H.R. Conf. Rep. No. 105-736, at 644 (1998). [2] U.S. General Accounting Office, Military Readiness: Reports to Congress Provide Few Details on Deficiencies and Solutions, GAO/ NSIAD-98-68 (Washington, D.C.: Mar. 30, 1998). [3] 10 U.S.C. sec. 482(d)(7)(B) and (e)(added by section 322 of Pub.L. 105-85, Nov. 18, 1997). [4] 10 U.S.C. sec. 482(d)(added by section 322 of Pub. L. 105-85, Nov. 8, 1997). [5] 10 U.S.C. sec. 117 (added by Pub.L. 105-261, sec.373(a)(1), Oct. 17, 1998 and as amended by Pub.L. 106-65, sec. 361(d)(2), Oct. 5, 1999.) [6] Department of Defense Readiness Reporting System (DRRS), DOD Directive 7730.65, June 3, 2002. [7] 10 U.S.C. sec. 117 note. [8] DOD is required under 10 U.S.C. sec. 482 to submit a quarterly readiness report to Congress. Under 10 U.S.C. sec. 482(b), each report is to specifically describe (1) each readiness problem and deficiency identified; (2) planned remedial actions; and (3) the key indicators and other relevant information related to each identified problem and deficiency. The quarterly reports provided to Congress are to be based on readiness assessments provided during that quarter to any DOD body that has responsibility for readiness oversight and whose membership includes at least one civilian officer in the Office of the Secretary of Defense at the level of Assistant Secretary or higher; by senior civilian and military officers of the military departments and commanders of the unified and specified commands; and as part of any regularly established process of periodic readiness reviews for the Department of Defense as a whole. [9] U.S. General Accounting Office, Military Readiness: DOD Needs to Develop a More Comprehensive Measurement System, GAO/NSIAD-95-29 (Washington, D.C.: Oct. 27, 1994). [10] Chairman of the Joint Chiefs of Staff Manual 3150.02, Global Status of Resources and Training System (GSORTS), (Washington, D.C.: Apr. 2000). [11] U.S. General Accounting Office, Military Training: DOD Lacks a Comprehensive Plan to Manage Encroachment on Training Ranges, GAO-02-614 (Washington, D.C.: June 11, 2002); and U.S. General Accounting Office, Military Training: Limitations Exist Overseas but Are Not Reflected in Readiness Reporting, GAO-02-525, (Washington, D.C.: Apr. 30, 2002). [12] U.S. General Accounting Office, Chemical and Biological Defense: Units Better Equipped, but Training and Readiness Reporting Problems Remain, GAO-01-27, (Washington, D.C.: Nov. 14, 2000); and U.S. General Accounting Office, Chemical and Biological Defense: Emphasis Remains Insufficient to Resolve Continuing Problems, GAO/NSIAD-96-103, (Washington, D.C.: Mar. 29, 1996). [13] 10 U.S.C. sec. 117(c)(7). [14] 10 U.S.C. sec. 117(c)(2) and (3). [15] 10 U.S.C. sec. 117(d) and (e). Also as explained in footnote 8, 10 U.S.C. sec. 482(c), requires that the quarterly reports be based on readiness assessments provided during that quarter as part of certain senior-level readiness assessments for DOD as a whole. [16] 10 U.S.C. sec. 482(b). [17] 10 U.S.C. sec. 482(d)(6)(A). [18] 10 U.S.C. sec. 482(d)(4)(D). [19] Our review of the National Defense Authorization Act for Fiscal Year 2002 (Pub.L. 107-107, Dec. 28, 2001) disclosed no new readiness reporting requirements. [20] Pub.L. 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281 (1998); H.R. Conf. Rep. No. 105-736, at 644 (1998). [21] See note 3. [22] 10 U.S.C. sec. 117(c)(2), (3), and (5). [23] Pub.L. 106-65, sec. 361(a) and (c), Oct. 5, 1999. [24] 10 U.S.C. sec. 117(c)(7). [25] 10 U.S.C. sec. 482(d)(added by Pub.L. 105-85, sec. 322(a)(1), Nov. 18, 1997). The Act included seven ’readiness indicators,“ each of which included from one to five specific reportable items, for a total of 19 specific items. [26] Nonpacing equipment is equipment not reported in GSORTS but nevertheless necessary for mission accomplishment. [27] 10 U.S.C. sec. 482(d)(1)(A) and (B), 2(B), 3(A), 4(B) and (C), 5(A) and (E). [28] Pub.L. 105-85, sec. 322(b), Nov. 18, 1997. [29] Pub.L. 105-261, sec. 373(b) and (c), Oct. 17, 1998 as amended by Pub.L. 106-65, sec. 361(d)(2), Oct. 5, 1999. [30] 10 U.S.C. sec. 117(f). [31] 10 U.S.C. sec. 117 (b)(2)and(f). [32] 10 U.S.C. sec. 117(a). [33] Public Law 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281 (1998); H.R. Conf. Rep. No. 105-736, at 644 (1998). [34] See note 6 above. [35] This study was directed by the National Defense Authorization Act for Fiscal Year 2000. Pub.L. 106-65 sec. 361(a), Oct. 5, 1999. [36] Department of Defense Inspector General, Information Technology: Global Command and Control System Readiness Assessment System Output Tool, D-2002-133 (Washington, D.C: July 24, 2002). [37] U.S. General Accounting Office, Defense Acquisitions: Factors Affecting Outcomes of Advanced Concept Technology Demonstrations GAO-03-52 (Washington, D.C.: Dec. 2, 2002). [38] The January-June 2001 report covered two quarters. GAO‘s Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO‘s commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO‘s Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as ’Today‘s Reports,“ on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select ’Subscribe to daily E-mail alert for newly released products“ under the GAO Reports heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.