Military Readiness
New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed
Gao ID: GAO-03-456 March 28, 2003
The Department of Defense's (DOD) readiness assessment system was designed to assess the ability of units and joint forces to fight and meet the demands of the national security strategy. In 1998, GAO concluded that the readiness reports provided to Congress were vague and ineffective as oversight tools. Since that time, Congress added reporting requirements to enhance its oversight of military readiness. Therefore, the Chairman asked GAO to examine (1) the progress DOD made in resolving issues raised in the 1998 GAO report on both the unit-level readiness reporting system and the lack of specificity in DOD's Quarterly Readiness Reports to the Congress, (2) the extent to which DOD has complied with legislative reporting requirements enacted since 1997, and (3) DOD's plans to improve readiness reporting.
Since 1998, DOD has made some progress in improving readiness reporting--particularly at the unit level--but some issues remain. For example, DOD uses readiness measures that vary 10 percentage points or more to determine readiness ratings and often does not report the precise measurements outside DOD. DOD included more information in its Quarterly Readiness Reports to the Congress. But quality issues remain--in that the reports do not specifically describe readiness problems, their effects on readiness, or remedial actions to correct problems. Nor do the reports contain information about funding programmed to address specific remedial actions. Although current law does not specifically require this information, Congress could use it for its oversight role. DOD complied with most, though not all, of the legislative readiness reporting requirements enacted by Congress in the National Defense Authorization Acts for Fiscal Years 1998-2002. For example, DOD (1) is now listing the individual units that have reported low readiness and reporting on the readiness of prepositioned equipment, as required by the fiscal year 1998 Act; (2) is reporting on 11 of 19 readiness indicators that commanders identified as important and that Congress required to be added to the quarterly reports in the fiscal year 1998 Act, but is not reporting on the other 8 readiness indicators; and (3) has not yet implemented a new comprehensive readiness reporting system as required in the fiscal year 1999 Act. As a result, Congress is not receiving all the information mandated by law. DOD issued a directive in June 2002 to establish a new comprehensive readiness reporting system that DOD officials said they plan to use to comply with the reporting requirements specified by Congress. The new system is intended to implement many of the recommendations included in a congressionally directed independent study for establishing such a system. However, the extent to which the new system will actually address the current system's shortcomings is unknown, because the new system is currently only a concept, and full capability is not scheduled until 2007. As of January 2003, DOD had not developed an implementation plan containing measurable performance goals, identification of resources, performance indicators, and an evaluation plan to assess progress in developing the new reporting system. Without such a plan, neither DOD nor the Congress will be able to fully assess whether the new system's development is on schedule and achieving desired results.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-03-456, Military Readiness: New Reporting System Is Intended to Address Long-Standing Problems, but Better Planning Is Needed
This is the accessible text file for GAO report number GAO-03-456
entitled 'Military Readiness: New Reporting System Is Intended to
Address Long-Standing Problems, but Better Planning Is Needed' which
was released on March 28, 2003.
This text file was formatted by the U.S. General Accounting Office
(GAO) to be accessible to users with visual impairments, as part of a
longer term project to improve GAO products‘ accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
Report to the Chairman, Subcommittee on Readiness, Committee on Armed
Services, House of Representatives:
United States General Accounting Office:
GAO:
March 2003:
Military Readiness:
New Reporting System Is Intended to Address Long-Standing Problems, but
Better Planning Is Needed:
GAO Highlights:
Highlights of GAO-03-456, a report to the Chairman, Subcommittee on
Readiness, Committee on Armed Services, House of Representatives
Why GAO Did This Study:
The Department of Defense‘s (DOD) readiness assessment system was
designed to assess the ability of units and joint forces to fight and
meet the demands of the national security strategy. In 1998, GAO
concluded that the readiness reports provided to Congress were vague
and ineffective as oversight tools. Since that time, Congress added
reporting requirements to enhance its oversight of military readiness.
Therefore, the Chairman asked GAO to examine (1) the progress DOD made
in resolving issues raised in the 1998 GAO report on both the unit-
level readiness reporting system and the lack of specificity in DOD‘s
Quarterly Readiness Reports to the Congress, (2) the extent to which
DOD has complied with legislative reporting requirements enacted since
1997, and (3) DOD‘s plans to improve readiness reporting.
What GAO Found:
Since 1998, DOD has made some progress in improving readiness reporting
”particularly at the unit level”but some issues remain. For example,
DOD uses readiness measures that vary 10 percentage points or more to
determine readiness ratings and often does not report the precise
measurements outside DOD. DOD included more information in its
Quarterly Readiness Reports to the Congress. But quality issues
remain”in that the reports do not specifically describe readiness
problems, their effects on readiness, or remedial actions to correct
problems. Nor do the reports contain information about funding
programmed to address specific remedial actions. Although current law
does not specifically require this information, Congress could use it
for its oversight role. DOD complied with most, though not all, of the
legislative readiness reporting requirements enacted by Congress in
the National Defense Authorization Acts for Fiscal Years 1998-2002.
For example, DOD
* is now listing the individual units that have reported low readiness
and reporting on the readiness of prepositioned equipment, as required
by the fiscal year 1998 Act;
* is reporting on 11 of 19 readiness indicators that commanders
identified as important and that Congress required to be added to the
quarterly reports in the fiscal year 1998 Act, but is not reporting on
the other 8 readiness indicators; and
* has not yet implemented a new comprehensive readiness reporting
system as required in the fiscal year 1999 Act.
As a result, Congress is not receiving all the information mandated by
law.
DOD issued a directive in June 2002 to establish a new comprehensive
readiness reporting system that DOD officials said they plan to use to
comply with the reporting requirements specified by Congress. The new
system is intended to implement many of the recommendations included
in a congressionally directed independent study for establishing such
a system. However, the extent to which the new system will actually
address the current system‘s shortcomings is unknown, because the new
system is currently only a concept, and full capability is not
scheduled until 2007.
As of January 2003, DOD had not developed an implementation plan
containing measurable performance goals, identification of resources,
performance indicators, and an evaluation plan to assess progress in
developing the new reporting system. Without such a plan, neither DOD
nor the Congress will be able to fully assess whether the new system‘s
development is on schedule and achieving desired results.
What GAO Recommends:
GAO made recommendations to improve readiness reporting and to develop
an implementation plan to allow DOD and the Congress to gauge progress
in developing DOD‘s new readiness reporting system. DOD did not agree
with our recommendations. After reviewing its comments, we modified
one recommendation but retained the others as originally stated.
www.gao.gov/cgi-bin/getrpt?GAO-03-456.
To view the full report, including the scope and methodology, click on
the link above. For more information, contact Neal P. Curtin at (757)
552-8100 or curtinn@gao.gov.
March 2003
GAO-03-456:
Contents:
This is a work of the U.S. Government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. It may contain
copyrighted graphics, images or other materials. Permission from the
copyright holder may be necessary should you wish to reproduce
copyrighted materials separately from GAO‘s product.
Letter:
Results in Brief:
Background:
Some Progress in Readiness Reporting, but Some Issues Remain:
DOD Has Not Fully Complied with All Legislative Requirements:
Lack of Implementation Plan Could Hinder Development of New Readiness
Reporting System:
Conclusions:
Recommendations for Executive Action:
Agency Comments and our Response:
Appendix I: Scope and Methodology:
Appendix II: Comments from the Department of Defense:
Figures:
Figure 1: DOD‘s Readiness Assessment Process, as of January 2003:
Figure 2: Timeline of Recommended, Required, and Planned Readiness
Reporting Changes Since 1994:
Abbreviations:
ACTD: Advanced Concept Technology Demonstration:
GSORTS: Global Status of Resources and Training System:
OUSD P&R: Office of the Under Secretary of Defense for Personnel and
Readiness:
United States General Accounting Office:
Washington, DC 20548:
March 28, 2003:
The Honorable Joel Hefley
Chairman
Subcommittee on Readiness
Committee on Armed Services
House of Representatives:
The Department of Defense‘s (DOD) readiness assessment system was
designed to assess the ability of units and joint forces to fight and
meet the demands of the national security strategy. For more than a
decade ending in 1998, various audit and oversight organizations
questioned the thoroughness and reliability of DOD reports on the
readiness of U.S. forces. Since 1998, Congress has added reporting
requirements to enhance its oversight of military readiness. In doing
so, Congress expressed concern over contradictions between assessments
of military unit readiness in reports and observations made by military
personnel in the field.[Footnote 1]
DOD provides Congress a quarterly report that contains readiness
information from several sources: the unit-level readiness assessment
system; results of scenario-based assessments; and summaries of
information briefed to senior DOD officials. We reviewed DOD‘s
readiness assessment and reporting system in 1998 and concluded that
the readiness reports provided Congress offered a vague description of
readiness problems and remedial actions and therefore were not
effective as oversight tools.[Footnote 2]
Considering the concerns raised in our 1998 report, reporting
requirements added by Congress since 1998, and the new national
security strategy, you asked us to provide an updated assessment of
DOD‘s readiness reporting. As agreed with your office, we examined (1)
the progress DOD has made in resolving issues raised in our prior
report on both the unit level readiness reporting system and the lack
of specificity in the Department‘s Quarterly Readiness Reports to the
Congress, (2) the extent to which DOD has complied with legislative
reporting requirements enacted since 1997, and (3) DOD‘s plans to
improve readiness reporting. In conducting this analysis, we compared
current reported readiness data with legislative requirements and with
data reported in 1998. We also identified DOD initiatives for improving
readiness reporting. We conducted our review from June 2002 through
January 2003 in accordance with generally accepted government auditing
standards. (For a complete description of our methodology, see app.
I.):
Results in Brief:
Since our 1998 report identifying limitations in readiness reporting,
DOD has made some progress in improving readiness reporting, such as
adding information on equipment cannibalization rates in its reports to
Congress. Some issues, however, remain. Although DOD has improved its
unit-level readiness reporting system, it still uses readiness measures
that vary 10 percentage points or more to determine readiness ratings
and often does not widely report the precise measurements outside DOD.
Since 1998, DOD has included more information in its Quarterly
Readiness Reports to the Congress, such as an annex presenting
equipment cannibalization rates. However, some degradation in these
reports has occurred. For example, DOD eliminated some previously
provided information--such as the joint force readiness assessments--
and the reports still contain very broad statements of readiness issues
and the remedial actions taken or planned to address readiness
problems. Even though some report annexes contain data, the data are
not adequately explained or related to the broad statements of
readiness issues mentioned in the reports. Also, the reports do not
contain information about funding that is programmed to address
specific remedial actions. Although this information is not required by
law, we believe it would be useful for Congress to understand the
significance of the information in these reports for use in its
oversight role.
DOD has complied with most, though not all, of the legislative
readiness reporting requirements enacted by Congress in the National
Defense Authorization Acts for Fiscal Years 1998-2002. For example, as
directed by the National Defense Authorization Act for Fiscal Year
1998, DOD is now reporting on the readiness of prepositioned equipment
and DOD is listing individual units that have reported low
readiness.[Footnote 3] For some provisions of the National Defense
Authorization Acts for the period we examined, DOD provided some of the
information specified in the acts. For example, DOD reports on some but
not all of the readiness indicators Congress, in the fiscal year 1998
act, required be added to the quarterly reports.[Footnote 4] For a few
requirements, DOD has yet to comply. For example, DOD has not
implemented a new readiness reporting system, which the National
Defense Authorization Act for Fiscal Year 1999 required to be
implemented by April 1, 2000.[Footnote 5] As a result, Congress is not
receiving all the information mandated by law.
DOD‘s main effort to improve readiness reporting is to develop and
implement a new comprehensive readiness reporting system in response to
the aforementioned legislation. However, the system‘s completion is
still several years away. DOD issued a directive in June 2002 to
establish this new system.[Footnote 6] Officials from the Office of the
Under Secretary of Defense for Personnel and Readiness (OUSD P&R)
responsible for developing the system stated they plan to use the new
system to comply with the reporting requirements contained in the
National Defense Authorization Acts and that the new system is intended
to implement many of the recommendations included in a congressionally
directed independent study for establishing a comprehensive readiness
reporting system. However, the extent to which the new system will
actually address the current system‘s shortcomings is unknown. This is
because the new system is currently only a concept and full capability
is not scheduled until 2007--some 7 years after a new comprehensive
system was legislatively mandated to be implemented.[Footnote 7] As of
January 2003 DOD had not developed an implementation plan to assess
progress in developing the new reporting system. We believe that an
implementation plan containing measurable performance goals,
identification of resources, performance indicators, and an evaluation
plan could help DOD as well as the Congress assess whether the new
system‘s development is on schedule and achieving desired results.
We are making recommendations to improve readiness reporting and to
develop an implementation plan to allow DOD and the Congress to gauge
DOD‘s progress in developing its new readiness reporting system.
In commenting on a draft of this report, DOD did not agree with our
recommendation to improve readiness reporting, saying that the
quarterly reports are comprehensive and spending more time on the
report would be counterproductive. Although the quarterly reports
contain voluminous data, the reader cannot assess its significance
since the data are explained inadequately. However, between 1998 and
mid-2001, DOD did include an unclassified summary of issues for each
service in the quarterly reports addressing several topics, including
personnel and equipment. Since DOD has highlighted key issues in past
quarterly reports, we believe that improving these reports by again
including a summary that highlights key readiness issues would be
beneficial. In doing so, DOD could focus on the most critical issues
that are of greatest concern to the services. Therefore, we have
modified our recommendation that DOD improve the quality of readiness
reporting to focus on issues deemed to be critical by the Secretary and
the services, including analyes and planned remedial actions for each
issue.
DOD also did not agree with our recommendations to develop an
implementation plan for the new readiness system and to provide
Congress annual updates on the new system‘s development. DOD said that
it is developing better tools for assessing readiness and had
established milestones and expected outcomes. Thus, DOD believes that
further planning and providing an annual update to Congress is
unnecessary. Considering that Congress expressed concern about DOD‘s
lack of progress in developing a comprehensive system and that DOD does
not plan for the new system to be fully capable until 2007, we retained
these two recommendations. A detailed discussion of DOD‘s comments and
our response is contained in the body of this report.
Background:
DOD‘s readiness assessment and reporting system was designed to assess
and report on military readiness at three levels--(1) the unit level;
(2) the joint force level; and (3) the aggregate, or strategic, level.
Unit-level readiness is assessed with the Global Status of Resources
and Training System (GSORTS), which is an automated system that
assesses the extent to which military units possess the required
resources and training to undertake their wartime missions. To address
joint readiness, the Chairman of the Joint Chiefs of Staff established
the Joint Monthly Readiness Review (now called the Joint Quarterly
Readiness Review or JQRR), that compiles readiness assessments from the
combatant commands, the combat support agencies, and the military
services. The Joint Staff and the services use these assessments to
brief DOD‘s leadership on the Senior Readiness Oversight Council--an
executive-level forum for monitoring emerging readiness issues at the
strategic level. The briefings to the council are intended to present a
view of readiness at the aggregate force level. From these briefings to
the council, DOD prepares a legislatively mandated quarterly readiness
report to Congress.[Footnote 8] Figure 1 provides an overview of DOD‘s
readiness assessment process.
Figure 1: DOD‘s Readiness Assessment Process, as of January 2003:
[See PDF for image]
[End of figure]
We have issued several reports containing recommendations for improving
readiness reporting. In 1994, we recommended DOD develop a more
comprehensive readiness system to include 26 specific readiness
indicators.[Footnote 9] In 1998, we reported on shortcomings in DOD‘s
readiness assessment system. At that time, we stated GSORTS‘
limitations included lack of precision in measurements, late reporting,
subjective input, and lack of standardization. Secondly, we reported
that while the Quarterly Readiness Reports to the Congress accurately
reflected briefs to the Senior Readiness Oversight Council, they lacked
specific details on deficiencies and remedial actions and thus did not
meet the requirements of 10 U.S.C. 482 (b). DOD concurred with our
recommendation that the Secretary of Defense take steps to better
fulfill the legislative reporting requirements under 10 U.S.C. 482 by
providing (1) supporting data on key readiness deficiencies and (2)
specific information on planned remedial actions. Finally, we reported
that deficiencies identified as a result of the Joint Monthly Readiness
Reviews remained open because the solutions require funding over the
long term. In 2002, we issued a classified report on DOD‘s process for
tracking the status of deficiencies identified in the Joint Monthly
Readiness Reviews. We made recommendations to improve DOD‘s deficiency
status reporting system and for DOD to develop funding estimates for
correcting critical readiness deficiencies. In its comments, DOD
generally agreed with the report‘s findings and recommendations.
Some Progress in Readiness Reporting, but Some Issues Remain:
Although DOD has made progress in resolving readiness reporting issues
raised in our 1998 report, we found that some of the same issues still
exist today. For example, DOD has added information to its Quarterly
Readiness Reports to the Congress (hereafter referred to as the
quarterly reports). However, we found that the reports still contain
vague descriptions of readiness problems and remedial actions. Even
though some report annexes contain detailed data, the data as presented
are not ’user friendly“--it is largely unevaluated and is not linked to
readiness issues mentioned in the report plus the report text does not
explain how the data relates to units‘ readiness. Thus, as we reported
in 1998, these reports do not specifically describe readiness problems
or remedial actions as required under 10 U.S.C. 482 (b). We believe
that this kind of information would be useful for Congress to
understand the significance of the information in these reports for use
in its oversight role.
Some Improvements in Unit-Level Reporting, but Some Issues Unchanged
Since 1998:
DOD has improved some aspects of its unit-level reporting system, the
Global Status of Resources and Training System (GSORTS). For example,
in 1998 GSORTS‘ data were maintained in multiple databases and data
were not synchronized. As of September 2002, the data are reported to a
central site, and there is one database of record. Also in 1998, U.S.
Army GSORTS review procedures delayed submission of Army data, and all
the services‘ data entry was manual. As of September 2002, Army
reporting procedures require reporting consistent with GSORTS‘
requirements, and all the services have automated data entry, which
reduces errors. In 1998, combat units only reported on readiness for
wartime missions. As of September 2002, combat units report on assigned
mission readiness in addition to wartime mission readiness.
Conversely, DOD has not resolved some issues we raised in 1998. For
example, readiness ratings are still reported in broad bands and actual
percentages of required resources are not externally reported. These
issues remain because the manual specifying readiness reporting rules
has not changed in these areas.[Footnote 10] The manual‘s definition of
readiness levels for personnel has not changed since our 1998 report--
it still defines readiness levels in bands of 10 percentage points or
more and does not require external reporting of actual percentages. For
example, the highest personnel rating can range from 90 percent to 100
percent, and there is no requirement to report the actual percentage
outside of DOD. We have also reported that GSORTS does not always
reflect training and equipment deficiencies. For example, we reported
in April and June 2002 that readiness data do not reflect the effect of
training range restrictions on unit readiness.[Footnote 11] We have
also reported that GSORTS does not include whether a unit‘s chemical/
biological equipment is usable.[Footnote 12]
In commenting on our analysis, the OUSD P&R office responsible for
readiness reporting stated that it recognized the imprecision of the
current measurements. According to that office, an effort to develop
the planned new readiness reporting system, which is discussed later in
this report, includes working with the DOD components to enhance and
expand readiness reporting.
Some Information Added to Quarterly Reports Since 1998, but Other
Information Eliminated:
Since our 1998 report, the quarterly reports improved in some areas,
but degraded in others. Although some information was added, we found
that some of the same quality issues remain--namely, that the reports
do not specifically describe readiness problems, their effects on
readiness, or remedial actions.
DOD has added information to the quarterly reports in response to
legislative direction. For example, DOD added information on the
services‘ cannibalization rates.[Footnote 13] Also, DOD added annual
reports on infrastructure and institutional training
readiness.[Footnote 14] However, some information was eliminated from
the quarterly reports. For example, the law requires results of joint
readiness reviews to be reported to Congress.[Footnote 15] DOD included
these results until the July-September 2001 Quarterly Readiness Report
to the Congress. Since that report, four quarterly reports have been
issued without the joint force assessments. Defense officials
responsible for readiness reporting said that the joint readiness
reviews were not included because the scenarios were based on the
former national security strategy of two major wars. The officials
stated they plan to include results from the joint readiness reviews in
future reports.
In commenting on our analysis the OUSD P&R office responsible for
readiness reporting stated that it continues to seek better ways to
provide concise, quality information.
Quarterly Reports Do Not Adequately Explain Readiness Issues:
As we reported in 1998, we found that the quarterly reports still
contain broad statements of readiness issues and remedial actions,
which are not supported by detailed examples and are not related to
data in the reports‘ annexes. Among other things, the law requires the
quarterly reports to specifically describe each readiness problem and
deficiency as well as planned remedial actions.[Footnote 16] The
reports did not specifically describe the nature of each readiness
problem or discuss the effects of each on unit readiness. Also, the
reports included only broad statements of remedial actions that lacked
details on timelines, objectives, or funding requirements. For example,
one report said that the Air Force continued to experience shortages in
critical job skills that affected the service‘s ability to train. The
report did not refer the reader to data in annexes showing training
readiness ratings; it did not state which skills were short, which
training was not accomplished, or whether this shortage had or was
expected to affect units‘ readiness ratings. Further, the report did
not explain the remedial actions taken or planned to reverse the skill
shortage, how long it would take to solve this problem, or what funding
was programmed to implement remedial actions. Defense readiness
officials agreed, stating that information in the quarterly reports is
summarized to the point that there are no details on readiness
deficiencies, remedial actions, or funding programmed to implement
remedial actions. We believe the Congress needs this type of
information to understand the significance of the information reported.
Although some of the quarterly report annexes contain voluminous data,
the data are not adequately explained or related to units‘ readiness.
The law does not mandate specific explanations of these ’readiness
indicators,“ but we believe it is essential for Congress to understand
the significance of the information in these reports for use in its
oversight role. For example, DOD is required to report on the
maintenance backlog.[Footnote 17] Although the report provides the
quantity of the backlog, it does not explain the effect the backlog had
on readiness. Specifically, the report did not explain whether units‘
readiness were affected because maintenance was not accomplished when
needed. In addition, DOD is required to report on training commitments
and deployments.[Footnote 18] The Expanded Quarterly Readiness Report
to Congress Implementation Plan dated February 1998 stated that ’either
an excessive or a reduced level of commitment could be an indicator of
potential readiness problems.“ However, OUSD P&R did not define what
kind of ’readiness problems“ this data may indicate would occur as a
result of ’excessive or reduced“ levels of training and deployments,
such as degraded equipment or training. The data reported are the
amount of training away from home station and the amount of
deployments. However, these data are not explained or related to a
unit‘s equipment or training ratings. Further, criteria have not been
established to distinguish between acceptable and unacceptable levels
of the training and deployment data reported. As a result, the reader
does not know whether the data reported indicate a problem or the
extent of the problem. In commenting on our analyses, OUSD P&R
acknowledged ’the Department would be better served by providing more
information as to how various data relates to readiness.“:
Generally, the quarterly reports also do not contain information on
funding programmed to implement specific remedial actions. For example,
one quarterly report included the statement that budgets were revised
’to address readiness and capabilities issues,“ but no examples were
provided. Also, the report lacked statements explaining how this
’budget revision“ would improve readiness. Although not required by
law, we believe it would prove useful for Congress to understand how
DOD addresses specific readiness problems.
In commenting on our analysis, OUSD P&R officials stated that they
would continue to work with the services to provide more fidelity with
the information presented in the quarterly report annexes. However,
they also said that detailed examples require significant staff effort
throughout DOD and that the added time for more detailed analysis could
render the report a historical document. They further said that
complete information would certainly be desired and agreed it is
important for the Congress to understand the significance of the
information in the quarterly reports for use in its oversight role.
DOD Has Not Fully Complied with All Legislative Requirements:
DOD has complied with most, but not all, of the readiness reporting
requirements added by Congress in the National Defense Authorization
Acts for Fiscal Years 1998 through 2002.[Footnote 19] Congress added
readiness reporting requirements out of concern over contradictions
between assessment of military unit readiness in official readiness
reports and observations made by military personnel in the
field.[Footnote 20] In a review of these acts, we identified both
recurring readiness reporting requirements that were added to existing
law and one-time reporting requirements related to military readiness.
We compared current readiness reporting to the requirements in these
acts to make an overall judgment on the extent of compliance. We did
not develop a total count of the number of reporting requirements
because the acts included a series of sections and subsections that
could be totaled in various ways. Because DOD is not reporting on all
the requirements added over the past several years, the Congress is not
receiving all the information mandated by law.
DOD Is Complying with Most of the New Requirements:
Our analysis showed that DOD has complied with most of the requirements
added in the National Defense Authorization Acts for Fiscal Years 1998-
2002. For example, DOD took the following actions in response to
legislative requirements:
* DOD is now reporting on the readiness of prepositioned equipment and
is listing individual units that have reported low readiness as
required by the National Defense Authorization Act for Fiscal Year
1998.[Footnote 21]
* DOD is reporting on infrastructure and institutional training
readiness as required by the National Defense Authorization Act for
Fiscal Year 1999.[Footnote 22]
* DOD contracted for an independent study of requirements for a
comprehensive readiness reporting system and submitted the study report
to the Congress as required by the National Defense Authorization Act
for Fiscal Year 2000.[Footnote 23]
* DOD has added quarterly information on the military services‘
cannibalization rates as required by the National Defense Authorization
Act for Fiscal Year 2001.[Footnote 24]
DOD Is Not Reporting on All Items Required:
DOD is reporting on some, though not all, of the items Congress
required be added to the quarterly readiness reports. For example, the
National Defense Authorization Act for Fiscal Year 1998 required 19
specific items be reported that are consistent with our previously
cited 1994 report on readiness reporting.[Footnote 25] The 1994 report
included a list of 26 readiness indicators that DOD commanders said
were important for a more comprehensive assessment of readiness. A 1994
DOD-funded study by the Logistics Management Institute found that 19 of
the 26 indicators could help DOD monitor critical aspects of readiness.
The 19 items listed in the National Defense Authorization Act for
Fiscal Year 1998 are very similar to those identified in the 1994
Logistics Management Institute study. DOD is reporting on 11 of the 19
items and is not reporting on the other 8. The eight items are (1)
historical personnel strength data and trends, (2) personnel status,
(3) borrowed manpower, (4) personnel morale, (5) operations tempo, (6)
training funding, (7) deployed equipment, and (8) condition of
nonpacing equipment[Footnote 26] as required in the Act.[Footnote 27]
In an implementation plan setting forth how it planned to comply with
reporting on the 19 items, which was also required by the National
Defense Authorization Act for Fiscal Year 1998,[Footnote 28] DOD stated
that it would not report on these eight indicators for the following
reasons:
* Deployed equipment was considered part of the status of prepositioned
equipment indicator.
* Historical personnel strength data and trends were available from the
annual Defense Manpower Requirements Report.
* Training funding and operations tempo were believed to be represented
adequately in the budget requests as flying hours, steaming days, or
vehicle miles and were not considered good measures of readiness
output.
* Personnel strength status was considered to be part of the personnel
rating, but DOD agreed to investigate other ways to evaluate the effect
of service personnel working outside the specialty and grade for which
they were qualified.
* Borrowed manpower data was only captured in a limited sector of the
deployable force and may not be meaningful until a better method is
developed to capture the data.
* Personnel morale had no existing data sources.
* The condition of nonpacing equipment had no reasonable measurement to
use as an indicator.
Notwithstanding the reasoning that DOD stated, these eight indicators
continue to be required by law, and we saw no indication in our work
that DOD is working to develop data for them.
Also, DOD is not complying with some of the requirements in the
National Defense Authorization Act for Fiscal Year 1999. Examples are
as follows:
* The act required DOD to establish and implement a comprehensive
readiness reporting system by April 2000.[Footnote 29] As of January
2003, DOD had not implemented a new system, and officials said it is
not expected to be fully capable until 2007 or 7 years later than
required.
* The act also required DOD to develop implementing regulations for the
new readiness reporting system.[Footnote 30] DOD had not developed
implementing regulations as of January 2003.
* The act required DOD to issue regulations for reporting changes in
the readiness of training or defense infrastructure establishments
within 72 hours.[Footnote 31] Although DOD has provided some guidance,
officials stated they have not issued regulations because no mechanism
exists for institutional training or defense infrastructure
establishments to report changes and because these entities are not
part of an established readiness reporting system.
In commenting on our analyses, DOD officials acknowledged ’the
Department is not in full compliance“ and stated that they plan to
achieve compliance with the new readiness reporting system under
development. OUSD P&R officials said that the shortfalls in reporting
are unwieldy under the current system; OUSD P&R intends to correct
these shortfalls when the new system is functional. However, as noted
above, DOD does not plan to implement its new system until 2007. As of
January 2003, DOD also had not targeted incremental improvements in
readiness reporting during the period in which the new system is being
developed. Until then, Congress will receive less readiness information
than it mandated by law.
Lack of Implementation Plan Could Hinder Development of New Readiness
Reporting System:
DOD issued a directive in June 2002 to establish a new readiness
reporting system. The Undersecretary of Defense for Personnel and
Readiness is to oversee the system to ensure the accuracy,
completeness, and timeliness of its information and data, its
responsiveness, and its effective and efficient use of modern practices
and technologies. Officials in the OUSD P&R readiness office
responsible for developing the new system said that they plan to use
the new system to comply with the requirements in the National Defense
Authorization Acts and to address many of the recommendations contained
in a congressionally directed independent study. However, as of January
2003, there are few details of what the new system would include.
Although the new system may have the potential to improve readiness
reporting, as of January 2003, it is only a concept without detailed
plans to guide development and monitor implementation. As a result, the
extent to which the new system will address existing shortcomings is
unknown.
DOD Directive Establishes a New Readiness Reporting System:
The National Defense Authorization Act for Fiscal Year 1999 required
DOD to establish a comprehensive readiness reporting system.[Footnote
32] In doing so, the Congress expressed concern about DOD‘s lack of
progress in developing a more comprehensive readiness measurement
system reflective of operational realities. The Congress also noted
that past assessments have suffered from DOD‘s inability to create and
implement objective and consistent readiness reporting criteria capable
of providing a clear picture to senior officials and the
Congress.[Footnote 33] Subsequently, the August 2001 Defense Planning
Guidance for Fiscal Years 2003-2007 called for the development of a
strategic plan for transforming DOD readiness reporting.
In June 2002, DOD issued a directive establishing the Department of
Defense Readiness Reporting System.[Footnote 34] The system will
measure and report on the readiness of military forces and the
supporting infrastructure to meet missions and goals assigned by the
Secretary of Defense. All DOD components will align their readiness
reporting processes in accordance with the directive.
The directive assigns oversight and implementing responsibility to the
Undersecretary of Defense for Personnel and Readiness. The
Undersecretary is responsible for developing, fielding, maintaining,
and funding the new system and scenario assessment tools. The
Undersecretary--in collaboration with the Joint Chiefs of Staff,
Services, Defense Agencies, and Combatant Commanders--is to issue
implementing instructions. The Chairman of the Joint Chiefs of Staff,
the Service Secretaries, the commanders of the combatant commands, and
the heads of other DOD components are each assigned responsibilities
related to readiness reporting.
OUSD P&R established a timetable to implement the new readiness
reporting system. OUSD P&R plans to achieve initial capability in 2004
and reach full capability in 2007. OUSD P&R officials involved in
developing the system said that they have been briefing the concept for
the new reporting system since October 2002. As of January 2003 these
officials stated that they are continuing what they have termed the
’concept demonstration“ phase, which began in October 2002. This phase
consists of briefing various offices within DOD, the Joint Staff, and
the services to build consensus and refine the new system‘s concept.
These officials also said that the new system will incorporate many,
but not all, of the recommendations contained in a legislatively
mandated independent study of readiness reporting, which concluded that
improvements were needed to meet legislatively mandated readiness
reporting requirements and included numerous recommendations for what a
new system should include.[Footnote 35] For example, the study
recommended that (1) DOD report on all elements essential to readiness,
such as depots, combat support agencies, and Defense agencies; (2)
reporting should be in terms of mission essential tasks; and (3) DOD
should measure the capability to carry out the full range of National
Security Strategy requirements--not just a unit‘s wartime mission.
An Implementation Plan Would Help Gauge Program Progress:
We believe that successfully developing and implementing a large-scale
effort, such as DOD‘s new readiness reporting system, requires an
implementation plan that includes measurable performance goals,
identification of resources, performance indicators, and an evaluation
plan. As discussed earlier, full implementation of DOD‘s new readiness
reporting system is several years away, and much remains to be done. In
January 2003 the OUSD P&R office responsible for developing the new
system said that the new readiness reporting system is a large endeavor
that requires buy-in from many users and that the development of the
system will be challenging. This office also wrote that it had just
been given approval to develop the new readiness reporting system, was
targeting development of an implementing instruction in the March 2003
time frame, and had not developed an implementation plan to assess
progress in developing and implementing the new reporting system. The
directive establishing the new reporting system requires the
Undersecretary of Defense for Personnel and Readiness, in collaboration
with others, to issue implementing instructions for the new system.
DOD has experienced delays in implementing smaller readiness
improvements than envisioned in the new readiness reporting system. One
such effort involved development of an interface to query the existing
readiness data base (GSORTS). In a July 2002 report, the DOD Inspector
General reported that the planned implementation of this interface
slipped 44 months, or just over 3.5 years.[Footnote 36] Also, history
has shown it takes DOD time to make changes in the readiness reporting
system. As illustrated in figure 2, DOD began reporting on specific
readiness indicators 4 years after it agreed with GAO recommendations
to include them in readiness reporting (see fig. 2).
Figure 2: Timeline of Recommended, Required, and Planned Readiness
Reporting Changes Since 1994:
[See PDF for image]
[End of figure]
Other DOD development efforts recognize the need for effective planning
to guide development. For example, DOD is working to transform military
training as directed by the Defense Planning Guidance for Fiscal Years
2003-07. A March 2002 Strategic Plan for Transforming DOD Training
developed by a different office within OUSD P&R discusses a training
transformation road map with major tasks subdivided into near-, mid-,
and long-term actions. The plan includes a list of near-term actions to
be completed by October 2003 and definition of mid-and long-term
actions in a comprehensive implementation plan that will identify
specific tasks, responsibilities, timelines, resources, and methods to
assess completion and measure success. The May 2002 Defense Planning
Guidance update for fiscal years 2004-2009 directs OUSD P&R, working
with other DOD components, to develop a comprehensive program to
implement the strategic training transformation plan and provide it to
the Deputy Secretary of Defense by April 1, 2003.
Since the directive for creating a new readiness reporting system
established broad policy with no specifics and since DOD has not
developed an implementation plan, the extent to which the new system
will address the current system‘s shortcomings will remain unknown
until the new system is fully capable in 2007. Until then, readiness
reporting will continue to be based on the existing system.
Commenting on its plans for the new system, OUSD P&R said that it is in
the process of creating an Advanced Concept Technology Demonstration
(ACTD) structure for the new system and will produce all necessary
planning documents required within the established ACTD process.
However, this process is intended to provide decision makers an
opportunity to understand the potential of a new concept before an
acquisition decision. We do not believe the ACTD process will
necessarily result in an implementation plan to effectively monitor
development and assess whether the new system is being implemented on
schedule and achieving desired results.
DOD‘s ACTD guidelines state the principal management tool for ACTDs is
a management plan, which provides a top-level description of the
objectives, critical events, schedule, funding, and measures of
evaluation for the project. We reported in December 2002 that these
guidelines contain advice and suggestions as opposed to formal
directives and regulations.[Footnote 37] DOD‘s guidelines state that
the ACTD should plan exercises or demonstrations to provide an adequate
basis for utility assessment. We also reported in December 2002 that
DOD lacks specific criteria to evaluate demonstration results, which
may cause acquisition decisions to be based on too little knowledge.
Therefore, we still believe an implementation plan is necessary since
the ACTD process does not require a detailed implementation plan and
does not always include specific criteria to evaluate effectiveness.
Conclusions:
While DOD has made some improvements in readiness reporting since 1998,
some of the same issues remain unresolved today. Although DOD is
providing Congress more data than in 1998, the voluminous data are
neither evaluated nor explained. The quarterly reports do not link the
effects of ’readiness issues“ or deficiencies to changes in readiness
at the unit level. Also, as in 1998, the reports contain vague
descriptions of remedial actions not linked to specific deficiencies.
Finally, the quarterly reports do not discuss funding that is
programmed to implement specific remedial actions. As a result, the
information available to Congress is not as effective as it could be as
an oversight tool.
Even though DOD directed development of a new readiness reporting
system, it has not yet developed an implementation plan identifying
objective and measurable performance goals, the resources and personnel
needed to achieve the goals, performance indicators, and an evaluation
plan to compare program results with goals, and milestones to guide
overall development of the new readiness system. Even though the new
system may have the potential to improve readiness reporting, without
an implementation plan little assurance exists that the new system will
actually improve readiness assessments by the time full capability is
planned in 2007. Without such a plan, it will also remain difficult to
gauge progress toward meeting the 2007 target date. This concern is
reinforced in light of the (1) years-long delays in implementing other
readiness reporting improvements and (2) the deficiencies in existing
reporting that OUSD P&R plans to rectify with the new system.
Furthermore, without an implementation plan neither senior DOD
leadership nor the Congress will be able to determine if the resources
spent on this system are achieving their desired results.
Recommendations for Executive Action:
To improve the information available to Congress for its use in its
oversight role, we recommend that the Secretary of Defense direct the
OUSD P&R to improve the quality of information contained in the
quarterly reports. Specifically, we recommend that DOD‘s reports
explain (in the unclassified section) the most critical readiness
issues that are of greatest concern to the department and the services.
For each issue, we recommend that DOD‘s reports include:
* an analysis of the readiness deficiencies, including a clear
explanation of how the issue affects units‘ readiness;
* a statement of the specific remedial actions planned or implemented;
and:
* clear statements of the funding programmed to implement each remedial
action.
To be able to assess progress in developing the new readiness system,
we recommend that the Secretary of Defense direct the OUSD P&R to
develop an implementation plan that identifies:
* performance goals that are objective, quantifiable, and measurable;
* the cost and personnel resources needed to achieve the goals,
including an identification of the new system‘s development and
implementation costs in the President‘s Budget beginning in fiscal year
2005 and Future Years Defense Plan;
* performance indicators to measure outcomes;
* an evaluation plan to compare program results with established goals;
and:
* milestones to guide development to the planned 2007 full capability
date.
To assist Congress in its oversight role, we recommend that the
Secretary of Defense give annual updates to the Congress on the new
readiness reporting system‘s development to include:
* performance measures,
* progress toward milestones,
* comparison of progress with established goals, and:
* remedial actions, if needed, to maintain the implementation schedule.
Agency Comments and our Response:
In written comments on a draft of this report, which are reprinted in
appendix II, the Department of Defense did not agree with our
recommendations.
In response to our recommendation that DOD improve the quality of
information contained in its quarterly readiness reports, DOD said that
the Quarterly Readiness Report to the Congress is one of the most
comprehensive and detailed reports submitted to the Congress that
discusses serious readiness issues and ways in which these issues are
being addressed. DOD further stated that the department presents
briefings on specific readiness issues to the Congress and that
spending more time and resources expanding the existing written report
would be counterproductive.
We recognize that the Quarterly Readiness Reports to the Congress
contain voluminous data. However, as discussed in this report, we found
that the quarterly reports‘ annexes are large and mostly consist of
charts or other data that are not adequately explained and are not
related to units‘ readiness. In some cases, criteria have not been
established to enable the reader to distinguish between acceptable and
unacceptable levels of the data reported. As a result, the reader
cannot assess the significance of the data because it is not at all
clear whether the data reported indicate a problem or the extent of the
problem. Considering that the quarterly reports contain inadequately
explained data and that much of the information is not ’user friendly,“
we continue to believe the quality of information in the quarterly
reports can be improved. In fact, we reviewed all the quarterly reports
provided to Congress since 1998 and found that through the January-June
2001 report[Footnote 38] the reports did include an unclassified
summary of readiness issues for each service addressing four topics--
personnel, equipment, training, and enablers (critical units or
capabilities, such as specialized aircraft, essential to support
operations). However, the reports did not include supporting data or a
discussion of remedial actions. Since that time, these summaries have
been eliminated from the quarterly reports. For example, the
unclassified narrative of the last two reports available at the time we
performed our work--January-March 2002 and April-June 2002--were less
than two pages long and neither discussed readiness issues nor ways in
which these issues are being addressed. One report discussed the new
readiness reporting system, and the other discussed a review of seven
environmental laws. Given that DOD has highlighted key issues in the
past, we believe that improving the quarterly reports would be
beneficial if DOD were to focus on the most critical readiness issues
that are of greatest concern to the services and includes supporting
data and a discussion of remedial actions. Therefore, we have modified
our recommendation that DOD improve the quality of readiness reporting
to focus on readiness issues deemed to be critical by the Secretary and
the military services and to provide more detailed data and analyses of
those issues and the remedial actions planned for each one.
DOD did not agree with our recommendations that it (1) develop an
implementation plan with, among other things, performance goals that
are objective, quantifiable, and measurable and (2) provide annual
updates to the Congress on the new readiness reporting system‘s
development. DOD said that it had undertaken an initiative to develop
better tools for assessing readiness and that it intended to apprise
Congress on its efforts to develop tools for readiness assessment. DOD
further stated that the effort to improve readiness reporting is in its
infancy, but that it has established milestones, cost estimates,
functional responsibilities, and expected outcomes. DOD believes that
further planning and a prescriptive annual update to the Congress is
unnecessary.
We agree that the new readiness reporting system may have the potential
to improve readiness reporting. However, as discussed in this report,
the directive establishing the new system contains very broad, high-
level statements of overall functional responsibilities and outcomes,
but no details on how these will be accomplished. Further, DOD has
established two milestones--initial capability in 2004 and full
capability in 2007. DOD does not have a road map explaining the steps
needed to achieve full capability by 2007, which is seven years after
Congress mandated a new system be in place. In addition, as discussed
earlier in this report, DOD has experienced delays in implementing much
smaller readiness improvements. While DOD has undertaken an initiative
to develop better tools for assessing readiness and intends to
routinely and fully apprise the Congress on its development efforts,
tools are the mechanics for evaluating readiness data. As such, tools
are not the same thing as the comprehensive readiness reporting system
mandated by Congress that DOD has said will include new metrics and
will evaluate entities within DOD that currently do not report
readiness. Considering that Congress expressed concern about DOD‘s lack
of progress in developing a comprehensive system, that developing and
implementing DOD‘s planned new system is scheduled to take 4 more
years, and that delays have been experienced in earlier efforts to make
small improvements in readiness reporting, we continue to believe that
it is important for DOD to develop an implementation plan to gauge
progress in developing and implementing the new readiness reporting
system and to provide annual updates to the Congress. Such a plan would
be consistent with DOD‘s approach to other major initiatives such as
transforming training. We have therefore retained these two
recommendations.
DOD also provided technical corrections and we have modified the report
where appropriate.
We are sending copies of this report to the Ranking Minority Member,
Subcommittee on Readiness, House Committee on Armed Services; the
Chairman and Ranking Minority Member, Subcommittee on Readiness and
Management Support, Senate Committee on Armed Services; other
interested congressional committees; Secretary of Defense; and the
Director, Office of Management and Budget. We will also make copies
available to others on request. In addition, the report will be
available at no charge on the GAO Web site at http://www.gao.gov.
If you or your staff have any questions, please call me on (757) 552-
8111 or by E-mail at curtinn@gao.gov. Major contributors to this report
were Steven Sternlieb, Brenda Waterfield, James Lewis, Dawn Godfrey,
and Herbert Dunn.
Sincerely yours,
Neal Curtin
Director
Defense Capabilities and Management:
Signed by Neal Curtin:
[End of section]
Appendix I: Scope and Methodology:
To assess the progress the Department of Defense (DOD) has made in
resolving issues raised in our prior report concerning both the unit
level readiness reporting system and the lack of specificity in DOD‘s
Quarterly Readiness Reports to the Congress, we met with DOD officials
and reviewed regulations and quarterly reports. Specifically, we met
with officials of the Office of the Undersecretary of Defense for
Personnel and Readiness (OUSD P&R) responsible for readiness reporting,
the Joint Staff, and the military services to discuss their individual
progress in each of these areas. To assess progress regarding unit
level readiness reporting, we reviewed the Chairman of the Joint Chiefs
of Staff manual governing this system and the related service
implementing instructions to determine if these documents had changed
since our 1998 report or if the manual and service instructions
continued to allow reporting in the same manner as reflected in our
earlier report. Through a comparison of the current and prior
documents, discussions with pertinent officials, and our analysis, we
determined whether the readiness reporting issues we raised in 1998 had
been resolved. We also reviewed the content of quarterly reports to
assess their quality and usefulness, and assess whether the problems we
reported in 1998 had been rectified. We discussed our analysis with
OUSD P&R officials and provided them with our analyses in order that
they could fully consider and comment on our methodology and
conclusions. We did not assess the accuracy of reported readiness data.
To determine the extent to which DOD has complied with legislative
reporting requirements enacted since our prior report, we compared a
complete listing of these requirements to DOD‘s readiness reporting.
First, we identified the legislatively mandated readiness reporting
requirements enacted since our 1998 report. To accomplish this, we
reviewed the National Defense Authorization Acts for Fiscal Years 1998-
2002 to list the one-time and recurring reporting requirements related
to military readiness. We also requested congressional staff and OUSD
P&R to review the list, and officials from both offices agreed it was
accurate. We did not develop a total count of the number of reporting
requirements because the acts included a series of sections and
subsections that could be totaled in various ways. Once we obtained
concurrence that this listing was complete and accurate, we compared
this list to current readiness reporting to make an overall judgment on
the extent of compliance.
To assess how DOD plans to improve readiness reporting, we reviewed the
June 2002 DOD directive establishing a new readiness reporting system
and a progress update briefing on the new system. We also obtained
readiness briefings from each of the services, OUSD P&R, and Joint
Staff officials. We performed several electronic searches of the Deputy
Under Secretary of Defense (Readiness) electronic Web site to determine
the status of readiness reporting. To assess how smoothly other
readiness improvements progressed, we reviewed DOD audit reports. We
discussed our findings with OUSD P&R officials and worked proactively
with them in conducting our analyses. Specifically, we provided them
drafts of our analyses for their comments and corrections.
We conducted our review from June 2002 through January 2003 in
accordance with generally accepted government auditing standards.
[End of section]
Appendix II: Comments from the Department of Defense:
OFFICE OF THE UNDER SECRETARY OF DEFENSE:
4000 DEFENSE PENTAGON WASHINGTON, D.C. 20301-4000:
MAR 21 2003:
PERSONNEL AND READINESS:
Mr. Neal P. Curtin:
Director, Defense Capabilities and Management U. S. General Accounting
Office:
Washington, D.C. 20548:
Dear Mr. Curtin:
This is the Department of Defense response to the General Accounting
Office draft report, MILITARY READINESS: New Reporting System Is
Intended to Address Longstanding Problems, but Better Planning Is
Needed,“ (GAO-03-456/GAO Code 350222).
We have reviewed the draft report and do not concur with the
recommendations. We also note a number of errors and will forward our
technical corrections separately.
We appreciate the opportunity to review and comment on this report.
Sincerely,
Joseph J. Angello, Jr. Director
Readiness Programming and Assessment:
Signed by Joseph J. Angello, Jr.:
Attachment:
--GAO DRAFT REPORT - DATED FEBRUARY 18, 2003 GAO CODE 350222/GAO-03-
456:
’MILITARY READINESS: New Reporting System Is Intended to Address
Longstanding Problems, but Better Planning Is Needed“:
DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS:
RECOMMENDATION 1: To improve the information available to Congress for
its use in its oversight role, the GAO recommended that the Secretary
of Defense direct the OUSD P&R to improve the quality of information
contained in the quarterly reports. Specifically, the GAO recommended
that DoD‘s reports include:
* analyses of readiness deficiencies, including a clear explanation of
how the deficiency affects units‘ readiness;
* a statement of the specific remedial actions planned or implemented
for each deficiency; and:
* clear statements of the funding programmed to implement each remedial
action. (p. 22/GAO Draft Report):
DOD RESPONSE: We cannot agree with this recommendation. The Quarterly
Readiness Report to Congress is one of the most comprehensive and
detailed reports submitted to the Congress on a routine basis. The
latest report was 289 pages, covering numerous readiness indicators,
pressing readiness issues, meeting summaries, and unit readiness
trends. Our reports routinely discuss serious readiness issues within
the Department of Defense and ways in which these issues are being
addressed. This recommendation calls for the report to capture
readiness ’deficiencies,“ how these deficiencies affect unit readiness,
and the resource implications to fix them. Given the Department has
roughly 10,000 units reporting readiness information, this tasking is
clearly infeasible. The Department does, however, routinely present
detailed and timely briefings on specific readiness issues and concerns
to the Congress. For example, our recent quarterly reports included
issues and actions associated with the Global War on Terror and
potential follow-on operations; preparedness for Chemical Biological
defense; and the management of low density/high demand assets. Spending
more time and resources expanding the existing written report would be
counterproductive, and ultimately provide less timely and meaningful
information to Congress.
RECOMMENDATION 2: To be able to assess progress in developing the new
readiness system, the GAO recommended that the Secretary of Defense
direct the OUSD P&R to develop an implementation plan that identifies:
* performance goals that are objective, quantifiable, and measurable;
* the cost and personnel resources needed to achieve the goals,
including an identification of the new system‘s development and
implementation costs in the President‘s Budget beginning in fiscal year
2005 and Future Years Defense Plan;
* performance indicators to measure outcomes;
* an evaluation plan to compare program results with established goals;
and:
* milestones to guide development to the planned 2007 full operational
capability date. (p. 22/GAO Draft Report):
DOD RESPONSE: We cannot concur with this recommendation. The Department
has undertaken an initiative to develop better tools for assessing the
readiness of our military forces and supporting infrastructure to meet
the threats facing our nation. This effort is in its infancy, but we
have established milestones, cost estimates, functional
responsibilities, and expected outcomes. We also have detailed plans
for specific information technology requirements and progress
indicators. More detailed and voluminous planning on top of the current
planning is unproductive and unnecessary for this effort.
RECOMMENDATION 3: To assist Congress in its oversight role, the GAO
recommended that the Secretary of Defense provide annual updates to the
Congress on the new readiness reporting system‘s development to
include:
* performance measures,
* progress toward milestones,
* comparison of progress with established goals, and:
* remedial actions, if needed, to maintain the implementation schedule.
(p. 23/GAO Draft Report):
DOD RESPONSE: We disagree with this recommendation. The Department
intends to routinely and fully apprise the Congress on our effort to
develop tools supporting readiness assessment. A prescriptive ’annual“
update is neither necessary nor warranted for this effort.
[End of section]
FOOTNOTES
[1] Public Law 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281
(1998); H.R. Conf. Rep. No. 105-736, at 644 (1998).
[2] U.S. General Accounting Office, Military Readiness: Reports to
Congress Provide Few Details on Deficiencies and Solutions, GAO/
NSIAD-98-68 (Washington, D.C.: Mar. 30, 1998).
[3] 10 U.S.C. sec. 482(d)(7)(B) and (e)(added by section 322 of Pub.L.
105-85, Nov. 18, 1997).
[4] 10 U.S.C. sec. 482(d)(added by section 322 of Pub. L. 105-85, Nov.
8, 1997).
[5] 10 U.S.C. sec. 117 (added by Pub.L. 105-261, sec.373(a)(1), Oct.
17, 1998 and as amended by Pub.L. 106-65, sec. 361(d)(2), Oct. 5,
1999.)
[6] Department of Defense Readiness Reporting System (DRRS), DOD
Directive 7730.65, June 3, 2002.
[7] 10 U.S.C. sec. 117 note.
[8] DOD is required under 10 U.S.C. sec. 482 to submit a quarterly
readiness report to Congress. Under 10 U.S.C. sec. 482(b), each report
is to specifically describe (1) each readiness problem and deficiency
identified; (2) planned remedial actions; and (3) the key indicators
and other relevant information related to each identified problem and
deficiency. The quarterly reports provided to Congress are to be based
on readiness assessments provided during that quarter to any DOD body
that has responsibility for readiness oversight and whose membership
includes at least one civilian officer in the Office of the Secretary
of Defense at the level of Assistant Secretary or higher; by senior
civilian and military officers of the military departments and
commanders of the unified and specified commands; and as part of any
regularly established process of periodic readiness reviews for the
Department of Defense as a whole.
[9] U.S. General Accounting Office, Military Readiness: DOD Needs to
Develop a More Comprehensive Measurement System, GAO/NSIAD-95-29
(Washington, D.C.: Oct. 27, 1994).
[10] Chairman of the Joint Chiefs of Staff Manual 3150.02, Global
Status of Resources and Training System (GSORTS), (Washington, D.C.:
Apr. 2000).
[11] U.S. General Accounting Office, Military Training: DOD Lacks a
Comprehensive Plan to Manage Encroachment on Training Ranges,
GAO-02-614 (Washington, D.C.: June 11, 2002); and U.S. General
Accounting Office, Military Training: Limitations Exist Overseas but
Are Not Reflected in Readiness Reporting, GAO-02-525, (Washington,
D.C.: Apr. 30, 2002).
[12] U.S. General Accounting Office, Chemical and Biological Defense:
Units Better Equipped, but Training and Readiness Reporting Problems
Remain, GAO-01-27, (Washington, D.C.: Nov. 14, 2000); and U.S. General
Accounting Office, Chemical and Biological Defense: Emphasis Remains
Insufficient to Resolve Continuing Problems, GAO/NSIAD-96-103,
(Washington, D.C.: Mar. 29, 1996).
[13] 10 U.S.C. sec. 117(c)(7).
[14] 10 U.S.C. sec. 117(c)(2) and (3).
[15] 10 U.S.C. sec. 117(d) and (e). Also as explained in footnote 8, 10
U.S.C. sec. 482(c), requires that the quarterly reports be based on
readiness assessments provided during that quarter as part of certain
senior-level readiness assessments for DOD as a whole.
[16] 10 U.S.C. sec. 482(b).
[17] 10 U.S.C. sec. 482(d)(6)(A).
[18] 10 U.S.C. sec. 482(d)(4)(D).
[19] Our review of the National Defense Authorization Act for Fiscal
Year 2002 (Pub.L. 107-107, Dec. 28, 2001) disclosed no new readiness
reporting requirements.
[20] Pub.L. 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281
(1998); H.R. Conf. Rep. No. 105-736, at 644 (1998).
[21] See note 3.
[22] 10 U.S.C. sec. 117(c)(2), (3), and (5).
[23] Pub.L. 106-65, sec. 361(a) and (c), Oct. 5, 1999.
[24] 10 U.S.C. sec. 117(c)(7).
[25] 10 U.S.C. sec. 482(d)(added by Pub.L. 105-85, sec. 322(a)(1), Nov.
18, 1997). The Act included seven ’readiness indicators,“ each of which
included from one to five specific reportable items, for a total of 19
specific items.
[26] Nonpacing equipment is equipment not reported in GSORTS but
nevertheless necessary for mission accomplishment.
[27] 10 U.S.C. sec. 482(d)(1)(A) and (B), 2(B), 3(A), 4(B) and (C),
5(A) and (E).
[28] Pub.L. 105-85, sec. 322(b), Nov. 18, 1997.
[29] Pub.L. 105-261, sec. 373(b) and (c), Oct. 17, 1998 as amended by
Pub.L. 106-65, sec. 361(d)(2), Oct. 5, 1999.
[30] 10 U.S.C. sec. 117(f).
[31] 10 U.S.C. sec. 117 (b)(2)and(f).
[32] 10 U.S.C. sec. 117(a).
[33] Public Law 105-261, Oct. 17, 1998; H.R. Rep. No. 105-532, at 281
(1998); H.R. Conf. Rep. No. 105-736, at 644 (1998).
[34] See note 6 above.
[35] This study was directed by the National Defense Authorization Act
for Fiscal Year 2000. Pub.L. 106-65 sec. 361(a), Oct. 5, 1999.
[36] Department of Defense Inspector General, Information Technology:
Global Command and Control System Readiness Assessment System Output
Tool, D-2002-133 (Washington, D.C: July 24, 2002).
[37] U.S. General Accounting Office, Defense Acquisitions: Factors
Affecting Outcomes of Advanced Concept Technology Demonstrations
GAO-03-52 (Washington, D.C.: Dec. 2, 2002).
[38] The January-June 2001 report covered two quarters.
GAO‘s Mission:
The General Accounting Office, the investigative arm of Congress,
exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO‘s commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO‘s Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as ’Today‘s Reports,“ on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select ’Subscribe to daily E-mail alert for newly
released products“ under the GAO Reports heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. General Accounting Office
441 G Street NW,
Room LM Washington,
D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S.
General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C.
20548: