Military Readiness
Navy's Fleet Response Plan Would Benefit from a Comprehensive Management Approach and Rigorous Testing
Gao ID: GAO-06-84 November 22, 2005
The Navy has been transforming itself to better meet 21st century needs. Since 2000, the Congress has appropriated about $50 billion annually for the Navy to operate and maintain its forces and support around 376,000 military personnel. In recognizing that the Navy faces affordability issues in sustaining readiness within its historical share of the defense budget, the Chief of Naval Operations announced a concept called the Fleet Response Plan to enhance its deployment readiness status. The Fleet Response Plan is designed to more rapidly prepare and sustain readiness in ships and squadrons. GAO evaluated the extent to which the Navy has (1) employed a sound management approach in implementing the Fleet Response Plan and (2) tested and evaluated the effectiveness of the plan and shared results to improve implementation.
In establishing the Fleet Response Plan, the Navy has embraced a major change in the way it manages its forces. However, the Navy's management approach in implementing the Fleet Response Plan has not fully incorporated sound management practices needed to guide and assess implementation. These practices include (1) establishing a coherent mission and strategic goals, including resource commitments; (2) setting implementation goals and a timeline; and (3) establishing a communication strategy. While the Navy has taken a number of positive actions to implement the plan, it has not provided readiness goals for units other than carrier strike groups; resource and maintenance goals; performance measures and timelines; or a communications strategy. Sound management practices were not fully developed because senior leaders wanted to quickly implement the plan in response to changes in the security environment. However, without an overall management plan containing all of these elements, it may be difficult for the Navy to determine whether its efforts to improve the fleet's readiness are achieving the desired results, adequately measuring overall progress, or identifying what resources are needed to implement the Fleet Response Plan. The Navy has not fully tested and evaluated the Fleet Response Plan or developed lessons learned to identify the effectiveness of its implementation and success over time. Systematic testing and evaluation of new concepts is an established practice to gain insight into how systems and capabilities will perform in actual operations. However, instead of methodically conducting realistic tests to evaluate the Fleet Response Plan, the Navy has tried to demonstrate the viability of the plan by relying on loosely linked events that were not part of an overall test and evaluation strategy. This approach could impair the Navy's ability to validate the plan and evaluate its success over time. In addition, the Navy has not used its lessons learned system to share the results of its Fleet Response Plan events or as an analytical tool to evaluate the progress of the plan and improve implementation, which limits the Navy's ability to identify and correct weaknesses across the fleet.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-06-84, Military Readiness: Navy's Fleet Response Plan Would Benefit from a Comprehensive Management Approach and Rigorous Testing
This is the accessible text file for GAO report number GAO-06-84
entitled 'Military Readiness: Navy's Fleet Response Plan Would Benefit
from a Comprehensive Management Approach and Rigorous Testing' which
was released on November 22, 2005.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to Congressional Committees:
United States Government Accountability Office:
GAO:
November 2005:
Military Readiness:
Navy's Fleet Response Plan Would Benefit from a Comprehensive
Management Approach and Rigorous Testing:
GAO-06-84:
GAO Highlights:
Highlights of GAO-06-84, a report to congressional committees:
Why GAO Did This Study:
The Navy has been transforming itself to better meet 21st century
needs. Since 2000, the Congress has appropriated about $50 billion
annually for the Navy to operate and maintain its forces and support
around 376,000 military personnel. In recognizing that the Navy faces
affordability issues in sustaining readiness within its historical
share of the defense budget, the Chief of Naval Operations announced a
concept called the Fleet Response Plan to enhance its deployment
readiness status. The Fleet Response Plan is designed to more rapidly
prepare and sustain readiness in ships and squadrons.
GAO evaluated the extent to which the Navy has (1) employed a sound
management approach in implementing the Fleet Response Plan and (2)
tested and evaluated the effectiveness of the plan and shared results
to improve implementation.
What GAO Found:
In establishing the Fleet Response Plan, the Navy has embraced a major
change in the way it manages its forces. However, the Navy‘s management
approach in implementing the Fleet Response Plan has not fully
incorporated sound management practices needed to guide and assess
implementation. These practices include (1) establishing a coherent
mission and strategic goals, including resource commitments; (2)
setting implementation goals and a timeline; and (3) establishing a
communication strategy. While the Navy has taken a number of positive
actions to implement the plan, it has not provided readiness goals for
units other than carrier strike groups; resource and maintenance goals;
performance measures and timelines; or a communications strategy. Sound
management practices were not fully developed because senior leaders
wanted to quickly implement the plan in response to changes in the
security environment. However, without an overall management plan
containing all of these elements, it may be difficult for the Navy to
determine whether its efforts to improve the fleet‘s readiness are
achieving the desired results, adequately measuring overall progress,
or identifying what resources are needed to implement the Fleet
Response Plan.
The Navy has not fully tested and evaluated the Fleet Response Plan or
developed lessons learned to identify the effectiveness of its
implementation and success over time. Systematic testing and evaluation
of new concepts is an established practice to gain insight into how
systems and capabilities will perform in actual operations. However,
instead of methodically conducting realistic tests to evaluate the
Fleet Response Plan, the Navy has tried to demonstrate the viability of
the plan by relying on loosely linked events that were not part of an
overall test and evaluation strategy. This approach could impair the
Navy‘s ability to validate the plan and evaluate its success over time.
In addition, the Navy has not used its lessons learned system to share
the results of its Fleet Response Plan events or as an analytical tool
to evaluate the progress of the plan and improve implementation, which
limits the Navy‘s ability to identify and correct weaknesses across the
fleet.
The U.S.S. George Washington Carrier Strike Group Sailing to
Participate in an Exercise:
[See PDF for image]
[End of figure]
What GAO Recommends:
To facilitate implementation of the Fleet Response Plan, GAO recommends
that the Navy develop a comprehensive management plan with goals and
performance measures. GAO also recommends that the Navy develop a
comprehensive testing and evaluation plan to help determine whether the
plan has been successful. The Department of Defense generally agreed
with GAO‘s recommendations and described efforts to address them.
www.gao.gov/cgi-bin/getrpt?GAO-06-84.
To view the full product, including the scope and methodology, click on
the link above. For more information, contact Janet St. Laurent at
(202) 512-4402 or stlaurentj@gao.gov.
[End of section]
Contents:
Letter:
Results in Brief:
Background:
Fleet Response Plan Does Not Fully Incorporate Sound Management
Practices:
Navy Has Not Fully Tested and Evaluated the Fleet Response Plan or
Developed Lessons Learned:
Conclusions:
Recommendations for Executive Action:
Agency Comments and Our Evaluation:
Appendix I: Scope and Methodology:
Appendix II: Comments from the Department of Defense:
Appendix III: GAO Contact and Staff Acknowledgments:
Figures:
Figure 1: The U.S.S. George Washington Carrier Strike Group Sailing to
Participate in an Exercise:
Figure 2: Fleet Response Plan Readiness Cycle Is Intended to Provide
Eight Carriers for Surge Deployments:
Figure 3: Best Practice Steps for a Methodical Study Plan Process:
United States Government Accountability Office:
Washington, DC 20548:
November 22, 2005:
Congressional Committees:
A key goal of Navy senior leadership is to transform the Navy to better
meet 21st century security challenges. Since 2000, Congress has
appropriated about $50 billion annually for the Navy to operate and
maintain its forces and support around 376,000 active military
personnel. Nonetheless, the Navy recognizes it is facing affordability
issues related to sustaining readiness while developing and procuring
several types of new ships within its historical share of the defense
budget. One area where the Navy has made significant changes is in its
operational posture. In March 2003, the Chief of Naval Operations
initiated the development of a concept, which became known as the Fleet
Response Plan,[Footnote 1] to enhance the Navy's deployment readiness
status. The Fleet Response Plan, as implemented by Fleet Forces Command
in May 2003, is designed to more rapidly prepare and then sustain
readiness in ships and squadrons. To achieve this capability, the plan
alters prior manning, maintenance, and training practices to allow for
a more responsive and ready naval force. The Navy expects this new
readiness approach will enable its forces to provide not only presence
and engagement in forward areas, but also surge a greater number of
ships on short-notice in response to significant crises without
increasing the readiness budget. The Fleet Response Plan modifies the
Navy's pre-2001 rotational deployment policy, replacing 6-month routine
deployments with more flexible deployment options for as many as eight
carrier strike groups when and where needed.
The Fleet Response Plan represents a major change in the way the Navy
manages its forces. Implementing large-scale change management
initiatives, such as organizational transformations, can be a complex
endeavor. Our prior work shows that failure to adequately address--and
often even consider--a wide variety of management issues is at the
heart of unsuccessful transformations. We have identified a number of
key best practices and lessons learned from major public and private
sector organizational mergers, acquisitions, and
transformations.[Footnote 2] These sound management practices include,
for example, establishing a coherent mission and integrated strategic
goals to guide the transformation, including resource commitments;
setting implementation goals and a timeline to build momentum and show
progress from day one; and establishing a communication strategy to
create shared expectations and report related progress.
We prepared this report under the Comptroller General's authority and
are providing it to you because of your oversight of defense issues. We
have previously reported on the maintenance aspects of the Navy's Fleet
Response Plan.[Footnote 3] This report focuses on the following two
questions: (1) To what extent has the Navy employed a sound management
approach in implementing the Fleet Response Plan? (2) To what extent
has the Navy tested and evaluated the effectiveness of its Fleet
Response Plan and shared results to improve its implementation?
To assess the Navy's management approach in implementing the Fleet
Response Plan, we obtained and analyzed key messages, briefings, and
instructions on the Fleet Response Plan and interviewed Department of
Defense (DOD) and Navy headquarters and fleet officials, and compared
the Navy's approach with best practices for transformations of large
organizations. To assess the extent to which the Navy has tested the
effectiveness of the Fleet Response Plan and shared results to improve
its implementation, we obtained briefings from and interviewed Navy
officials, reviewed and queried the Navy Lessons Learned System to
determine relevant lessons recorded, and examined Navy guidance on test
and evaluation efforts. We reviewed and validated the Navy Lessons
Learned System data and determined the data were sufficiently reliable
for our analysis. We conducted our review from January 2005 through
August 2005 in accordance with generally accepted government auditing
standards. The scope and methodology used in our review are described
in further detail in appendix I.
Results in Brief:
While the Navy has taken a number of positive actions to establish the
Fleet Response Plan, the Navy has not fully developed a comprehensive
management approach to effectively guide, monitor, and assess
implementation. Sound management practices that provide a framework for
implementing and managing programs include (1) establishing a coherent
mission and integrated strategic goals to guide the transformation,
including resource commitments; (2) establishing a communication
strategy to create shared expectations and report related progress; and
(3) setting implementation goals and a timeline to build momentum and
show progress. The Navy's implementation of the Fleet Response Plan has
included some aspects of these practices. For example, the Navy has
established strategic goals for progressive readiness levels for
carrier strike groups. However, the Navy has not established specific
readiness goals for the rest of the fleet or determined the resources
needed to achieve its goals, although it has stated the plan will be
budget neutral. The Navy also does not have an official written
definition of the Fleet Response Plan or communications strategy that
clearly establishes a coherent mission and integrated strategic goals
to guide the transformation, including resource commitments. These
sound management practices were not fully implemented because senior
leaders wanted to implement the Fleet Response Plan as quickly as
possible in response to the Chief of Naval Operations' direction.
Although Navy officials recently tasked the Center for Naval Analyses
to conduct the study to identify potential goals and performance
measures, it is not clear what the study will recommend or how long it
will take for the Navy to take action. Until an overall management plan
is developed, neither the Navy nor Congress may be able to determine
whether the Fleet Response Plan has effectively achieved its goals,
measure the plan's overall progress, or determine what resources are
needed to implement the plan.
In addition, the Navy has not fully developed a comprehensive set of
plans to test and evaluate the Fleet Response Plan and has not
developed formal lessons learned from past exercises to evaluate the
plan's effectiveness. DOD has long recognized the importance of testing
new concepts by using war games and experimentation, and recent Navy
guidance stresses the importance of establishing a long-range plan for
testing complex and novel problems. The Navy has identified three
loosely linked events that Navy officials say demonstrate the viability
of the plan. However, none of the three events cited by the Navy were
part of an overall test and evaluation strategy to assess the value of
the plan in increasing readiness. The Navy has not developed an
overarching test and evaluation plan because Navy officials believe
existing readiness reports provide adequate information to assess the
Fleet Response Plan. However, readiness reports do not produce
information on important factors such as costs, long-term maintenance
implications, or quality of life issues. Additionally, the Navy did not
analyze and evaluate the results of these three events and submit
formal lessons learned to the Navy Lessons Learned System. Without
systematic testing and evaluation and use of the lessons learned
system, the Navy's ability to validate a complex change like the Fleet
Response Plan, identify and correct problem areas, and disseminate
lessons learned throughout the fleet is limited. This not only prevents
ship and command staffs from learning from the experiences of others,
but it also prevents the Navy Lessons Learned System from possibly
identifying problems and patterns across the fleet that may require a
high-level, Navy-wide response.
To facilitate implementation of the Fleet Response Plan, we recommend
that the Navy develop a comprehensive management plan with goals and
performance measures. We also recommend that the Navy develop a
comprehensive testing and evaluation plan to help determine whether the
Fleet Response Plan has been successful. In its comments on a draft of
this report, DOD generally concurred with the report's recommendations.
DOD concurred with our recommendation to develop a comprehensive
management plan with goals and performance measures, citing several
actions it has underway or planned. DOD partially concurred with our
recommendation to test and evaluate the Fleet Response Plan. However,
DOD does not plan to conduct no-notice surges as we recommended because
it views such exercises as unnecessary and costly. We continue to
believe that no-notice surges are important because they can serve as
an effective means of gauging whether the Navy is ready to respond to
real world events, which can occur with little notice. DOD comments and
our evaluation are discussed on pages 21 and 22.
Background:
Composition of a Carrier Strike Group:
Carrier strike groups are typically centered around an aircraft carrier
and its air wing, and also include a guided missile cruiser; two guided
missile destroyers; a frigate; an attack submarine; and one or more
supply ships with ammunition, fuel, and supplies (such as food and
spare parts). These groups are formed and disestablished on an as
needed basis, and their compositions may differ though they contain
similar types of ships. Figure 1 shows a carrier strike group sailing
in a group formation as it prepares to participate in an exercise.
Figure 1: The U.S.S. George Washington Carrier Strike Group Sailing to
Participate in an Exercise:
[See PDF for image]
[End of figure]
Origin of the Fleet Response Plan:
Prior to the September 11, 2001, terrorist attacks, only those Navy
ships and air squadrons at peak readiness were deployed overseas,
usually for 6 months at a time. Most of the Navy's remaining units were
not available because they were in early stages of their maintenance or
training cycles, or because the Navy did not have good visibility of
the readiness of these units. This prompted the Chief of Naval
Operations in March 2003 to task the Commander, Fleet Forces Command,
to develop the Fleet Response Plan concept to enhance the Navy's surge
capability. The Chief of Naval Operations approved the concept in May
2003 and further directed the Commander, Fleet Forces Command, to be
responsible and accountable for effectively implementing the plan.
Fleet Response Plan Believed to Provide Increased and Flexible
Readiness:
The Fleet Response Plan emphasizes an increased level of readiness and
the ability to quickly deploy naval forces to respond to crises,
conflicts, or homeland defense needs. The plan applies broadly to the
entire fleet; however, it only sets specific requirements for carrier
strike groups. For example, the plan calls for eight carrier strike
groups to be ready to deploy within 90 days of notification. Six of
them would be available to deploy within 30 days and the other two
within 90 days. This is commonly referred to as the 6 + 2 goal. Under
the Fleet Response Plan, the Navy has developed a surge capability
schedule that it uses to manage and identify the level of training a
ship has completed and its readiness to deploy. The schedule contains
three progressive readiness goals: emergency surge, surge-ready, and
routine deployable status.[Footnote 4] Each readiness goal specifies
phases of training that must be completed to achieve the goal. To be
placed in emergency surge status, a ship or an air squadron needs to
have completed its unit-level phase training. Achieving surge-ready
status requires the completion of integrated phase training. Attaining
routine deployable status requires achievement of all necessary
capabilities, completion of underway sustainment phase training, and
certification of the unit for forward deployed operations.[Footnote 5]
The surge capabilities schedule provides a readiness snapshot for each
ship, allowing decision makers to quickly determine which ships are
available to meet the needs of the mission. Figure 2 illustrates how
the Navy notionally identifies the eight aircraft carriers available
for surge deployments. The carriers numbered 1 through 6 are expected
to be ready to deploy within 30 days notice. The carriers labeled "+1"
and "+2" are expected to able to surge within 90 days notice. The six
surge-ready carriers include two carriers on deployment (numbered 3 and
4), one carrier that is part of the forward deployed naval force based
in Japan (number 6), and three carriers in the sustainment phase
(numbered 1, 2, and 5). These six carriers are expected to have
completed postdeployment depot-level maintenance and their unit-level
phase training. The two additional surge carriers are expected to have
completed depot-level maintenance but not to have completed unit-level
phase training. The remaining four carriers are in the maintenance
phase or deep maintenance.[Footnote 6]
Figure 2: Fleet Response Plan Readiness Cycle Is Intended to Provide
Eight Carriers for Surge Deployments:
[See PDF for image]
[End of figure]
Revised Fleet Response Plan Is Being Developed:
Based on the Navy's experiences during the past 2 years, Fleet Forces
Command has convened a cross-functional working group to develop a
refined version of the Fleet Response Plan. This update, known as Fleet
Response Plan-Enhanced, is intended to further define the Fleet
Response Plan, modify terminology for progressive readiness states to
better reflect their meaning, tie in elements such as a human capital
strategy, and expand the focus of the plan beyond carrier strike groups
to the entire Navy. It may also extend the Fleet Response Plan's
current employment cycle length of 27 months. The Fleet Response Plan-
Enhanced is still under development at this time.
Fleet Response Plan Does Not Fully Incorporate Sound Management
Practices:
The Navy's management approach in establishing the Fleet Response Plan
as its new readiness construct has not fully incorporated sound
management practices needed to effectively guide, monitor, and assess
implementation.[Footnote 7] Studies by several organizations have shown
that successful organizations in both the public and private sectors
use sound management practices to assist agencies in measuring
performance, reporting results, and achieving desired outcomes. These
practices provide management with a framework for effectively
implementing and managing programs and shift program management focus
from measuring program activities and processes to measuring program
outcomes. Sound management practices include (1) establishing a
coherent mission and integrated strategic goals to guide the
transformation, including resource commitments; (2) setting
implementation goals and a timeline to build momentum and show progress
from day one; and (3) establishing a communication strategy to create
shared expectations and report related progress.
The Navy's implementation of the Fleet Response Plan has included some
aspects of these practices. For example, the Navy has established some
strategic goals needed to meet the intent of the plan, such as the
progressive readiness levels of emergency surge, surge-ready, and
routine deployable status. The Navy also has established specific
training actions to support these goals, such as that carrier strike
groups must complete unit-level training to be certified as emergency
surge-ready. However, other actions taken by the Navy do not fully
incorporate these practices. For example, the Navy has identified the 6
+ 2 surge capability as a readiness goal and performance measure for
carrier strike groups, but no such goal was established for the rest of
the fleet. The Navy also has some unofficial goals and performance
measures regarding manning and maintenance, but these unofficial goals
and performance measures have not been formally established. For
example, briefings on the Fleet Response Plan state that the Navy
desires and needs fully manned ships (i.e., manning at 100 percent of a
ship's requirement) for the program to be successful. Moreover,
according to Navy officials, the Navy has not established milestones
for achieving its results.
In addition, 2 years after initiating implementation of the Fleet
Response Plan, the Navy still does not have an official written
definition of the Fleet Response Plan that clearly establishes a
coherent mission and integrated strategic goals to guide the
transformation, including resource commitments. This definition would
describe the Fleet Response Plan's total scope and contain guidance
with formal goals and performance measures. The Navy recently has taken
some action to address this area. In February 2005, the Navy directed
the Center for Naval Analyses to conduct a study to develop formal
definitions and guidance as well as identify goals and performance
measures for the plan. However, it remains to be seen whether this
study will be completed as planned by November 2005; if it will
recommend developing and implementing sound management practices, such
as goals, measures, milestones, and timelines; and whether any
management improvement recommendations made in the study will be
implemented by the Fleet Forces Command, the Navy command responsible
for implementing the Fleet Response Plan. Without goals, performance
measures, timelines, milestones, benchmarks, and guidance to help
effectively manage implementation of the Fleet Response Plan and
determine if the plan is achieving its goals, the Navy may find it more
difficult to implement the Fleet Response Plan across the entire naval
force.
Moreover, despite the Navy's unofficial goal that the Fleet Response
Plan be budget neutral, as articulated in briefings and by senior
leaders, the Navy has not yet clearly identified the resources needed
to achieve its goals or provided a rationale for how these resources
will contribute to achieving the expected level of performance. Navy
officials have said that current operations and maintenance funding
levels, as well as manning at 100 percent of required positions, have
contributed to successful implementation of the Fleet Response Plan.
However, officials do not know what level of manning or funding is
actually required for program success over the long term to avoid any
unintended consequences, such as greater amounts of deferred
maintenance. According to Navy officials, it is difficult to attribute
costs to the plan because there is no single budget line item that
tracks the costs associated with the Fleet Response Plan. Without
knowing the funding needed, the Navy may not be able to assess the
impact of possible future changes in funding on implementing the plan.
Furthermore, without a comprehensive plan that links costs with
performance measures and outcomes, neither the Navy nor Congress may be
able to determine if the Fleet Response Plan is actually achieving its
unofficial goal of being budget neutral.
Finally, the Navy also has not developed a comprehensive communications
strategy that reaches out to employees, customers, and stakeholders and
seeks to genuinely engage them in a two-way exchange, which is a
critical step in successfully implementing cultural change or
transformation. We looked for formal mechanisms that communicated the
details of the Fleet Response Plan and spoke with personnel from
carrier strike groups, aircraft carriers, air wings and an air
squadron, one surface combatant ship, and other command staff.[Footnote
8] We found that while the Fleet Response Plan was communicated
extensively to senior-level officers, and the Navy provided numerous
briefings and messages related to the plan, communication and
understanding of the plan did not flow through to the lower ranks.
While the concept of the Fleet Response Plan is generally understood by
some senior-level officials, many of the lower grade personnel on these
ships were unaware of the scope, goals, and other aspects of the plan.
In the absence of clear communication throughout the fleet via an
overall communications strategy that could increase employee awareness
of the Fleet Response Plan, its successful implementation could be
impeded.
Sound management practices, such as those noted above, were not fully
used by the Navy because senior leaders wanted to quickly implement the
Fleet Response Plan in response to the Chief of Naval Operations'
desires. However, without an overall management plan containing all of
these elements to guide the implementation of such a major change, it
may be difficult for the Navy and Congress to determine the extent to
which the Fleet Response Plan is achieving the desired results, measure
its overall progress, or determine the resources needed to implement
the plan.
Navy Has Not Fully Tested and Evaluated the Fleet Response Plan or
Developed Lessons Learned:
The Navy has not fully tested and evaluated the Fleet Response Plan or
developed lessons learned to identify the effectiveness of its
implementation and success over time. The methodical testing,
exercising, and evaluation of new doctrines and concepts is an
established practice throughout the military to gain insight into how
systems and capabilities will perform in actual operations. However,
instead of methodically conducting realistic tests to evaluate the
Fleet Response Plan, the Navy has tried to demonstrate the viability of
the plan by relying on loosely linked events that were not part of an
overall test and evaluation strategy, which impairs the Navy's ability
to validate the plan and evaluate its success over time. In addition,
the Navy has not used its lessons learned system to share the results
of its Fleet Response Plan tests or as an analytical tool to evaluate
the progress of the plan and improve implementation, which limits the
Navy's ability to identify and correct weaknesses across the fleet.
Methodical Tests and Evaluations of New Concepts Are Important:
Methodically testing, exercising, and evaluating new doctrines and
concepts is an important and established practice throughout the
military. DOD has long recognized the importance of using tabletop
exercises, war games, and experimentation[Footnote 9] to explore
military doctrine, operational concepts, and organizational
arrangements. Collectively, these tests and experiments can provide
important insight into how systems and capabilities will perform in
actual operations. U.S. Joint Forces Command, which has lead
responsibility for DOD experimentation on new concepts of operation and
technologies, states that its experimental efforts aim to foster
military innovation and improvement by exploring, developing, and
transferring new concepts and organizational ideas into operational
reality.
Particularly large and complex issues may require long-term testing and
evaluation that is guided by study plans. Joint Forces Command's Joint
Warfighting Center has an electronic handbook that provides guidance
for conducting exercises and lays out the steps in an exercise life
cycle: design; planning; preparation; execution; and analysis,
evaluation, and reports. The Army also has well-established
guidance[Footnote 10] governing service studies, analyses, and
evaluations that the Navy feels is representative of best practices for
military operations research. This provides an important mechanism
through which problems pertaining to critical issues and other
important matters are identified and explored to meet service needs. As
shown in figure 3, the Army's process involves six major steps that
create a methodical process for developing, conducting, documenting,
and evaluating a study. Following a formal study process enables data
evaluation and development of lessons learned that could be used to
build on the existing knowledge base. In a roundtable discussion with
the Fleet Forces Command on the rationale behind Summer Pulse 2004, the
Navy's major exercise for the Fleet Response Plan, a senior Navy
official stated, "From the concept, — you need to exercise, — you need
to practice, — you need to demonstrate it to know you got it right and
what lessons are there to learn from how we did it."[Footnote 11]
Figure 3: Best Practice Steps for a Methodical Study Plan Process:
[See PDF for image]
[End of figure]
Other governmental agencies, like GAO, and the private sector also rely
on detailed study plans, or data collection and analysis plans, to
guide the development of studies and experiments and the collection and
analysis of data, and to provide a feedback loop that links the
outcomes of the study or experiment event and subsequent analysis to
the original goals and objectives of the study or event. GAO guidance
states that data collection and analysis plans "should carry forward
the overall logic of the study so that the connection between the data
that will be collected and the answers to the study questions will
become evident."[Footnote 12]
Recent Navy guidance also recognizes the need for a thorough evaluation
of complex initiatives. In April 2005, the Navy issued a Study Planning
and Conduct Guide assembled by the Navy Warfare Development
Command.[Footnote 13] This guide stresses the importance of
establishing a long-range plan for complex and novel problems and lays
out the rationale for detailed study plans for exercises and
experiments, as they establish a structure in which issues are explored
and data are collected and analyzed in relation to the established
goals or objectives for the event. Furthermore, the Navy's guide notes
that random, inadequately prepared events and a determination just to
study the problem do not lead to successful resolution of problems that
may arise in programs and concepts that the Navy is testing and
evaluating.
Navy Events to Show Viability of the Fleet Response Plan Have Lacked
Methodical Testing and Evaluation:
The Navy has not methodically conducted realistic tests of the Fleet
Response Plan to demonstrate the plan's viability and evaluate its
progress and success over time, instead relying on loosely linked
events and some routine data to demonstrate the viability of the plan.
The events identified by the Navy as successful tests of the Fleet
Response Plan are Summer Pulse 2004, the emergency deployment of the
U.S.S. Abraham Lincoln, and Global War on Terrorism Surge 2005, but of
these events only Summer Pulse 2004 was driven by the Fleet Response
Plan with the intent of demonstrating that large numbers of ships could
be surged. In addition, these events were not part of an overall test
and evaluation strategy that yielded specific information from which to
assess the value of the plan in increasing readiness and meeting the
new 6 + 2 surge capability goal for carrier strike groups.
Summer Pulse 2004 encompassed a number of previously scheduled
deployments, exercises, and training events that took place between
June and August of 2004. The intent of Summer Pulse 2004 was to
demonstrate the Fleet Response Plan's new readiness construct and the
Navy's ability to deploy multiple carrier strike groups of varying
levels of readiness. However, Summer Pulse 2004 was not a methodical
and realistic test of the Fleet Response Plan for three reasons. First,
Summer Pulse 2004 did not follow best practices regarding study plans
and the ability to evaluate the impact and outcomes of the plan. The
Navy did not develop a formal study plan identifying study objectives,
data collection requirements, and analysis, or produce a comprehensive
after-event report describing the study's findings. Navy officials have
stated that the elements of a formal study plan were there for the
individual deployments, exercises, and training events constituting
Summer Pulse 2004, but were not brought together in a single package.
While the Navy may have had the study elements present for the
individual exercises, they were not directly linked to testing the
Fleet Response Plan. Without such a comprehensive study plan and
overall evaluation, there is no ability to discern potential impacts on
fleet readiness, maintenance, personnel, and other issues that are
critical to the Fleet Response Plan's long-term success. Second, Summer
Pulse 2004 was not a realistic test because all participating units had
several months' warning of the event. As a result, five carriers were
already scheduled to be at sea and only two had to surge. Because six
ships are expected to be ready to deploy with as little as 30 days'
notice under the plan and two additional carriers within 90 days, a
more realistic test of the Fleet Response Plan would include no-notice
or short-notice exercises.[Footnote 14] Such exercises conducted
without advance notification to the participants would provide the
highest degree of challenge and realism. Without such exercises, the
Navy might not be able to realistically practice and coordinate a full
surge deployment. Third, Summer Pulse 2004 was not a sufficient test
because the Navy involved only seven carriers instead of the eight
carriers called for in the plan. Therefore, it did not fully test the
Navy's ability to meet deployment requirements for the expected force.
Another event cited by the Navy as evidence of the Fleet Response
Plan's success is the deployment of the U.S.S. Abraham Lincoln carrier
strike group while it was in surge status in October 2004. Originally
scheduled to deploy in the spring of 2005, the Lincoln was deployed
early to support operations in the Pacific Command area of operation
and provide aid to areas devastated by a tsunami in the Indian Ocean in
December 2004. Navy officials said that the Fleet Response Plan enabled
the Navy to identify a carrier to send to the Pacific and to quickly
tailor its training package based on its progressive readiness status.
The Navy touted this rapid response relief work by a strike group
deployed during surge status as a Fleet Response Plan success story. We
agree that the Lincoln carrier strike group was able to respond
quickly. However, the extent to which this event realistically tested
the Fleet Response Plan's expectations for surging one carrier strike
group is not known. As with Summer Pulse 2004, the Lincoln deployment
was not a methodical test of the Fleet Response Plan because there was
no plan to systematically collect or analyze data that would evaluate
the outcomes of the Lincoln deployment against Fleet Response Plan-
related study goals.
The Navy also pointed to a third event, its recent Global War on
Terrorism Surge 2005,[Footnote 15] as an indicator that the Fleet
Response Plan works. The Global War on Terrorism surge was a response
to a request for forces[Footnote 16] from which the Navy is looking to
glean Fleet Response Plan-related information about what did and did
not work when the ships return. However, this is not a good test of the
Fleet Response Plan because there is no plan showing what specific data
are being collected or what analytical approaches are being employed to
assess the ships' experiences. As of September 2005, no other events
had been scheduled to further test and evaluate the Fleet Response
Plan.
The Navy has not developed the kind of comprehensive plans to test and
evaluate the Fleet Response Plan as recommended by DOD and Navy
guidance and best practices because Navy officials have stated that
existing readiness reporting processes effectively evaluate the Fleet
Response Plan's success on a daily basis. They said after-action
reports[Footnote 17] from training exercises and the Joint Quarterly
Readiness Review[Footnote 18] assist with this function. Navy officials
explained that they implemented the Fleet Response Plan the same way
they had implemented the Inter-Deployment Training Cycle, the
predecessor to the Fleet Response Plan's Fleet Readiness Training Plan.
While this may be true, the Inter-Deployment Training Cycle was focused
on the specific training needed to prepare units for their next
deployment, not for implementing a new readiness construct that
emphasized surge versus routine deployments. Furthermore, the Inter-
Deployment Training Cycle did not contain stated goals whose validity
the Navy needed to test. In addition, ongoing readiness reports do not
provide information on important factors such as costs, long-term
maintenance implications, and quality of life issues.
The Summer Pulse 2004, Lincoln surge deployment, and Global War on
Terrorism Surge 2005 testing events were not part of a methodical test
and evaluation approach. Therefore, the Navy is unable to convincingly
use these events to evaluate the Fleet Response Plan and determine
whether the plan has been successful in increasing readiness or
achieving other goals. Moreover, without effective evaluation of the
Fleet Response Plan, the Navy may be unable to identify and correct
potential problem areas across the fleet. Without a comprehensive long-
range plan that establishes methodical and realistic testing of the
Fleet Response Plan, the Navy may be unable to validate the Fleet
Response Plan operational concept, evaluate its progress and success
over time, and ensure that it can effectively meet Navy goals over the
long term without any adverse, unintended consequences for maintenance,
quality of life, and fleet readiness.
Navy Lessons Learned System's Repository and Analytic Resources Have
Not Been Used to Catalog and Share Fleet Response Plan Lessons:
The formal Navy repository for lessons learned, the Navy Lessons
Learned System, has not been used to disseminate Fleet Response Plan-
related lessons learned or to analyze test results to evaluate the
progress of the plan and improve implementation. The Navy Lessons
Learned System has been designated by the Chief of Naval Operations as
the singular Navy program for the collection, validation, and
distribution of unit feedback as well as the correction of problems
identified and derived from fleet operations, exercises, and
miscellaneous events. However, there are no mechanisms or requirements
in place to force ships, commands, and numbered fleet staffs to submit
all lessons learned to the Navy Lessons Learned System, although such
mechanisms exist for the submission of port visit and other reports.
For the events that the Navy cites as tests of the Fleet Response Plan,
it did not analyze and evaluate the results and produce formal lessons
learned to submit to the Navy Lessons Learned System for recordation
and analysis. Any evaluation done of the testing events has not been
incorporated into the Lessons Learned System, preventing comprehensive
analyses of lessons learned and identification of problems and patterns
across the fleet that may require a high-level, Navy-wide response.
Some ship and carrier strike group staff informed us that they prefer
informal means of sharing lessons learned, because they feel the
process through which ships and commands have to submit lessons learned
for validation and inclusion in the database can be complex and
indirect. This may prevent ship and command staffs across the fleet
from learning from the experiences of others, but it also prevents the
Navy Lessons Learned System from performing comprehensive analyses of
the lessons learned and possibly identifying problems and patterns
across the fleet that may require a high-level Navy-wide response. In
addition, the lessons learned are recorded by mission or exercise
(e.g., Operation Majestic Eagle) and not by operational concept (e.g.,
the Fleet Response Plan), making identification of Fleet Response Plan-
specific lessons learned difficult and inconsistent.
Over the last 10 years, we have issued several reports related to
lessons learned developed by the military. We have found that service
guidance does not always require standardized reporting of lessons
learned[Footnote 19] and lessons learned are not being used in training
or analyzed to identify trends and performance weaknesses.[Footnote 20]
We emphasized that effective guidance and sharing of lessons learned
are key tools used to institutionalize change and facilitate efficient
operations. We found that despite the existence of lessons learned
programs in the military services and the Joint Staff, units repeat
many of the same mistakes during major training exercises and
operations. Our current review indicates that the Navy still does not
include all significant information in its lessons learned database.
Therefore, Navy analysts cannot use the database to perform
comprehensive analyses of operational concepts like the Fleet Response
Plan to evaluate progress and improve implementation.
Officials from the Navy Warfare Development Command stated that the
Navy is currently drafting a new Chief of Naval Operations Instruction
governing the Navy Lessons Learned System that will address some of
these issues.[Footnote 21] Navy Warfare Development Command officials
hope that the new instruction will result in several improvements over
the current system. First, they would like to see a dual reporting
system, so that lessons learned are simultaneously sent to the Navy
Lessons Learned System for preliminary evaluation when they are
submitted to the numbered fleets for validation. This would allow Navy
Lessons Learned analysts to look at unvarnished data for patterns or
issues of interest to the Chief of Naval Operations, without taking
away the numbered fleets' validation processes. In addition, officials
would like to establish deadlines for the submission of lessons learned
to ensure timeliness. Not only will these changes add value to the data
stored in the Navy Lessons Learned System, but they will keep the data
flowing while ensuring that data are actually submitted and not lost as
they move up the chain of command. According to Navy Lessons Learned
officials, other branches of the military already allow operators in
the field to submit lessons learned directly to their lessons learned
systems, enabling value-added analysis and the timely posting of
information. By addressing these issues, the Navy can help ensure that
the lessons learned process will become more efficient, be a command
priority, and produce actionable results.
Conclusions:
Two years after implementing a major change in how it expects to
operate in the future, the Navy has not taken all of the steps needed
to enable the Navy or Congress to assess the effectiveness of the Fleet
Response Plan. As the Navy prepares to implement the Fleet Response
Plan across the entire naval force, it becomes increasingly important
that the Navy effectively manages this organizational transformation so
that it can determine if the plan is achieving its goals. The absence
of a more comprehensive overarching management plan to implement the
Fleet Response Plan has left essential questions about definitions,
goals, performance measures, guidance, timelines, milestones,
benchmarks, and resources unanswered, even though sound management
practices recognize the need for such elements to successfully guide
activities and measure outcomes. The absence of these elements could
impede effective implementation of the Fleet Response Plan.
Furthermore, without a comprehensive plan that links costs with
performance measures and outcomes, neither the Navy nor Congress may be
able to determine if the Fleet Response Plan is budget neutral. More
effective communications throughout the fleet using an overall
communications strategy could increase employee awareness of the plan
and help ensure successful implementation.
The Navy also has not developed a comprehensive long-range plan for
testing and evaluating the Fleet Response Plan. Without a well-
developed plan and methodical testing, the Navy may not be aware of all
of the constraints to successfully surging its forces to crises in a
timely manner. Moreover, the absence of an overarching testing and
evaluation plan that provides for data collection and analysis may
impede the Navy's ability to use its testing events to determine
whether the Fleet Response Plan has been successful in increasing
readiness and to identify and correct problem areas across the fleet.
Failure to document and record the results of testing and evaluation
efforts in the Navy Lessons Learned System could limit the Navy's
ability to validate the value of the concept, identify and correct
performance weaknesses and trends across the fleet, perform
comprehensive analyses of lessons learned, and disseminate these
lessons and analyses throughout the fleet.
Recommendations for Executive Action:
To facilitate successful implementation of the Fleet Response Plan and
enhance readiness and ensure the Navy can determine whether the plan
has been successful in increasing readiness and is able to identify and
correct performance weaknesses and trends across the fleet, we
recommend that the Secretary of Defense take the following two actions:
* Direct the Secretary of the Navy to develop a comprehensive
overarching management plan based on sound management practices that
will clearly define goals, measures, guidance, and resources needed for
implementation of the Fleet Response Plan, to include the following
elements:
* establishing or revising Fleet Response Plan goals that identify what
Fleet Response Plan results are to be expected and milestones for
achieving these results,
* developing implementing guidance and performance measures based on
these goals,
* identifying the costs and resources needed to achieve each
performance goal, and:
* communicating this information throughout the Navy.
* Direct the Secretary of the Navy to develop a comprehensive plan for
methodical and realistic testing and evaluation of the Fleet Response
Plan. Such a comprehensive plan should include a description of the
following elements:
* how operational tests, exercises, war games, experiments,
deployments, and other similar events will be used to show the
performance of the new readiness plan under a variety of conditions,
including no-notice surges;
* how data will be collected and analyzed for these events and
synthesized to evaluate program success and improvements; and:
* how the Navy Lessons Learned System will collect and synthesize
lessons from these events to avoid repeating mistakes and improve
future operations.
Agency Comments and Our Evaluation:
In written comments on a draft of this report, DOD generally concurred
with our recommendations and cited actions it will take to implement
the recommendations.
DOD concurred with our recommendation that the Navy should develop a
comprehensive overarching management plan based on sound management
practices that would clearly define the goals, measures, guidance, and
resources needed for successful implementation of the Fleet Response
Plan, including communicating this information throughout the Navy. DOD
noted that the Navy has already taken action or has plans in place to
act on this recommendation, and described several specific
accomplishments and ongoing efforts in this regard. DOD also noted that
the Navy intends to communicate through message traffic, white papers,
instructions, lectures, and meetings with Navy leadership. We agree
that these means of communication are an important part of an effective
communication strategy; however, we do not believe that these methods
of communication constitute a systemic strategy to ensure communication
at all personnel levels. We believe the Navy would benefit from a
comprehensive communication strategy that builds on its ongoing
efforts, but encompasses additional actions to ensure awareness of the
plan throughout the Navy.
DOD partially concurred with our recommendation to test and evaluate
the Fleet Response Plan. DOD noted that it plans to use a variety of
events and war games to evaluate the Fleet Response Plan, but it does
not see a need to conduct no-notice surges to test the Fleet Response
Plan. DOD stated that it believes no-notice surges are expensive and
unnecessary and could lead to penalties on overall readiness and the
ability to respond to emergent requirements. DOD also noted that the
Navy has surged single carrier strike groups, expeditionary strike
groups, and individual ships or units under the Fleet Response Plan,
and it cited several examples of such surges. We commend the Navy's
plans to use a variety of events to evaluate the Fleet Response Plan
and its use of the Navy Lessons Learned System to report and evaluate
the lessons learned in the Global War on Terrorism Surge 2005 exercise
held earlier this year. However, we continue to believe that no-notice
surges are critical components of realistic testing and evaluation
plans and that the benefits of such exercises can outweigh any
additional costs associated with conducting such tests on a no-notice
basis. Both we and Congress have long recognized the importance of no-
notice exercises. For example, in a 1989 report, we noted that DOD was
instituting no-notice exercises to assess the preparedness of combatant
commands' state of training of their staffs and components.[Footnote
22] In addition, in 1990 the Department of Energy conducted no-notice
tests of security personnel in response to our work and out of
recognition that such tests are the best way to assess a security
force's ability at any given time.[Footnote 23] Furthermore, in recent
years, the Department of Homeland Security, Department of Energy, and
others have conducted no-notice exercises because they add realism and
demonstrate how well organizations are actually prepared to respond to
a given situation. Despite the importance of no-notice exercises, the
Navy has not conducted no-notice exercises to test and evaluate the
centerpiece surge goal of 6 + 2 for carrier strike groups. We believe
that the smaller surges cited by DOD can provide insights into the
surging process, but we do not believe that such surges can effectively
test the Navy's readiness for a full 6 + 2 carrier strike group surge.
DOD also provided technical and editorial comments, which we have
incorporated as appropriate. DOD's comments are reprinted in appendix
II of this report.
We are sending copies of this report to other interested congressional
committees; the Secretary of Defense; the Secretary of the Navy; and
the Director, Office of Management and Budget. We will make copies
available to others upon request. In addition, the report will be
available at no charge on the GAO Web site at http://www.gao.gov.
If you or your staff have any questions about this report, please
contact me at (202) 512-4402 or stlaurentj@gao.gov. Contact points for
our Offices of Congressional Relations and Public Affairs may be found
on the last page of this report. GAO staff who made major contributions
to this report are listed in appendix III.
Signed by:
Janet St. Laurent:
Director, Defense Capabilities and Management:
List of Committees:
The Honorable John Warner:
Chairman:
The Honorable Carl Levin:
Ranking Minority Member:
Committee on Armed Services:
United States Senate:
The Honorable Ted Stevens:
Chairman:
The Honorable Daniel K. Inouye:
Ranking Minority Member:
Subcommittee on Defense:
Committee on Appropriations:
United States Senate:
The Honorable Duncan L. Hunter:
Chairman:
The Honorable Ike Skelton:
Ranking Minority Member:
Committee on Armed Services:
House of Representatives:
The Honorable C. W. Bill Young:
Chairman:
The Honorable John P. Murtha
Ranking Minority Member:
Subcommittee on Defense:
Committee on Appropriations:
House of Representatives:
[End of section]
Appendix I: Scope and Methodology:
To assess the extent to which the Navy has employed a sound management
approach in implementing the Fleet Response Plan, we interviewed Navy
headquarters and fleet officials; received briefings from relevant
officials; and reviewed key program documents. In the absence of a
comprehensive planning document, we compared best practices for
managing and implementing major efforts to key Navy messages,
directives, instructions, and briefings, including, but not limited to,
the Culture of Readiness message sent by the Chief of Naval Operations
(March 2003); the Fleet Response Concept message sent by the Chief of
Naval Operations (May 2003); the Fleet Response Plan Implementation
message sent by the Commander, Fleet Forces Command (May 2003); the
Fleet Response Plan Implementation Progress message sent by the
Commander, Third Fleet (September 2003); and the U.S. Fleet Forces
Command's Fleet Training Strategy instruction (May 2002 and an undated
draft). We also conducted meetings with several of the commanding
officers, executive officers, and department heads of selected carrier
strike groups, aircraft carriers, and air wings to obtain information
on how the plan had been communicated, how the plan had changed their
maintenance and training processes, the impact on their quality of
life, the cost implications of the plan, and other factors.
To assess the extent to which the Navy has tested the effectiveness of
the Fleet Response Plan and shared results to improve its
implementation, we obtained briefings; interviewed Navy headquarters
and fleet officials; and reviewed test and evaluation guidance for both
the Navy and other federal agencies. To evaluate the three Fleet
Response Plan demonstrations identified by the Navy, we interviewed
officials from the Fleet Forces Command and the Navy Warfare
Development Command, reviewed existing documentation on the
demonstrations, queried the Navy Lessons Learned System for lessons
learned from the demonstrations, and compared our findings to accepted
best practices for tests and evaluations. Further, we reviewed Navy
Lessons Learned System instructions and queried the system to determine
recorded lessons learned pertaining to the Fleet Response Plan.
We validated the Navy Lessons Learned System data and determined the
data were sufficiently reliable for our analysis. We conducted our
review from January 2005 through August 2005 in accordance with
generally accepted government auditing standards at the following
locations:
* The Joint Staff, Washington, D.C.
* U.S. Pacific Command, Camp H. M. Smith, Hawaii:
* Offices of the Chief of Naval Operations, Washington, D.C.
* Naval Sea Systems Command, Washington, D.C.
* U.S. Fleet Forces Command, Norfolk, Virginia:
* Offices of the Fleet Forces Command:
* Commander, U.S. Second Fleet:
* Commander, Naval Air Forces:
* Commander, Submarine Forces:
* Commander, Naval Surface Force:
* U.S. Marine Corps Forces:
* Afloat Training Group:
* Navy Warfare Development Command, Newport, Rhode Island:
* Commander, U.S. Pacific Fleet, Pearl Harbor, Hawaii:
* Offices of the U.S. Pacific Fleet:
* Commander, Naval Submarine Force:
* Asia-Pacific Center for Security Studies, Honolulu, Hawaii:
We held group discussions with selected personnel such as commanding
officers, executive officers/chief of staffs, department heads, and
crew members from the following units, all located in the Norfolk,
Virginia, area:
* U.S.S. Bulkeley:
* U.S.S. Enterprise:
* U.S.S. Theodore Roosevelt:
* U.S.S. Harry S. Truman:
* Carrier Air Wing 3:
* Carrier Air Wing 8:
* Carrier Strike Group 2:
* Carrier Strike Group 10:
[End of section]
Appendix II: Comments from the Department of Defense:
OFFICE OF THE UNDER SECRETARY OF DEFENSE:
PERSONNEL AND READINESS:
4000 DEFENSE PENTAGON:
WASHINGTON, D.C. 20301-4000:
OCT 31 2005:
Janet A. St. Laurent:
Director, Defense Capabilities and Management:
U.S. Government Accountability Office:
441 G Street, N.W.:
Washington, DC 20548:
Dear Ms. St. Laurent:
This is the Department of Defense (DoD) response to the GAO draft
report, GAO-06-84, "MILITARY READINESS: Navy's Fleet Response Plan
Would Benefit From a Comprehensive Management Approach and Rigorous
Testing" (GAO Code 350625)," dated September 27, 2005.
DoD appreciates the opportunity to comment on the draft report. DoD
concurs with Recommendation 1 and partially concurs with Recommendation
2. Detailed comments on the GAO recommendations and report are
enclosed:
Sincerely,
Signed by:
Joseph J. Angello:
Director:
Readiness Programming & Assessment:
Enclosure: As Stated:
GAO DRAFT REPORT - DATED September 27, 2005 GAO CODE 350625/GAO-06-84:
"MILITARY READINESS: Navy's Fleet Response Plan Would Benefit From a
Comprehensive Management Approach and Rigorous Testing"
DEPARTMENT OF DEFENSE COMMENTS TO THE RECOMMENDATIONS:
RECOMMENDATION 1: The GAO recommended that the Secretary of Defense
direct the Secretary of the Navy to develop a comprehensive overarching
management plan based on sound management practices that will clearly
define goals, measures, guidance, and resources needed for
implementation of the Fleet Response Plan, to include the following
elements:
* establish or revise Fleet Response Plan goals that identify what
Fleet Response Plan results are to be expected and milestones for
achieving these results;
* develop implementing guidance and performance measures based on these
goals;
* identify the costs and resources needed to achieve each performance
goal; and:
* communicate this information throughout the Navy.
DoD RESPONSE: Concur. The Navy has either already developed or is in
process of developing guidance, performance measures, and a methodology
to capture Fleet Response Plan (FRP) costs and efficiencies.
Additionally, the Navy leadership is continuously working to educate
and inform personnel, at all levels, on the dynamics, requirements, and
benefits of the FRP.
Since its very inception, the FRP is designed as a dynamic and flexible
operational process that affords all levels of command an ability to
adapt to changing world environments and requirements. Therefore, the
management and process tracking of FRP will also change in concert with
those emergent requirements.
Specific accomplishments and ongoing efforts are described below:
* FRP goals and expectations were initially established and
communicated to the Fleet via formal message traffic and guidance from
subordinate commanders. CNO's "personal for" Commanders, Commanding
Officers, and Officers in Charge dated 22 May 2003 outlined the
requirement to develop a process that would improve Navy's speed of
response. It announced the approval of the Fleet Response Concept and
directed CFFC to develop FRP to make the concept a reality.
Supplemental guidance from CFFC explained FRP and delineated key
implementation milestones. More recent and expanded guidance can be
seen in the draft CFFC Instruction 3501.3A titled Fleet Training
Continuum. This document discusses and identifies in-depth guiding
principles, roles and responsibilities, business rules, and critical
training enablers.
* Overarching guidance in the form of a CNO FRP instruction is in work
with OPNAV N43 as lead agent. This instruction will address all facets
of FRP and will include references and descriptions of performance
based measures and consolidate implementation guidance. Additionally,
significant progress has been made to standardize performance measures
via a capabilities-based approach, using the Navy Mission Essential
Task Lists (NMETLs). These are the basic building blocks used in
determining Fleet training requirements, plans, execution, and
readiness assessment.
* Identification of costs and resources needed to efficiently operate
the FRP has been undertaken by several organizations under the
cognizance of CFFC. The Center for Naval Analysis (CNA) will deliver
the results of a comprehensive study on the "Resource Implications of
the FRP". This study is currently expected to be published in November
2005. Fleet Forces Command is also developing the Fleet Training
Capability Cost System (FTCCS) which is an Activity Based
Costing/Management system (ABC/M) designed to trace Fleet resource
utilization and costs.
* As previously described, Navy leadership has worked diligently to
communicate evolving changes to Fleet plans to prepare and respond to
our changing security environment. This work will continue through
message traffic, white papers, instructions, and face-to-face lectures
and meetings with Navy leadership.
RECOMMENDATION 2: The GAO recommended that the Secretary of Defense
direct the Secretary of the Navy to develop a comprehensive plan for
methodical and realistic testing and evaluation of the Fleet Response
Plan. Such a comprehensive plan should include a description of the
following elements:
* how operational tests, exercises, war games, experiments,
deployments, and other similar events will be used to show the
performance of the new readiness plan under a variety of conditions,
including no-notice surges;
* how data will be collected and analyzed for these events and
synthesized to evaluate program success and improvements, and:
* how the Navy Lessons Learned System will collect and synthesize
lessons from these events to avoid repeating mistakes and improve
future operations.
DoD RESPONSE: Partially Concur. Continual evaluation is a key tenet of
the revised Fleet Response Plan (FRP) and the Navy plans on using a
variety of events to include deployments and war games to evaluate FRP.
However, conducting no-notice surges to "test' 'the Fleet Response Plan
is an expensive and unnecessary means to evaluate the FRP, and could
lead to penalties on overall readiness and ability to respond to future
emergent requirements. The FRP is an operational construct which is
subject to methodical and realistic "testing and evaluation" during
ongoing deployments and operations. To expand:
The Navy's mission is global response and providing our nation and
allies dominant naval power wherever and whenever required. Our
maritime forces do this through persistence, precision, reach, speed,
and agility. The FRP is the operational framework that capitalizes on
investments that have been made to readiness accounts and leverages
force provider capability to meet global Combatant Commander demand
signals for traditional roles, e.g. forward presence, and new emerging
mission areas such as Detention Facility Security, JTF HOA, and Air
Ambulance. The FRP is an on-going mission-driven response plan that
provides the right readiness at the right time, enables responsive
forward presence, and drives our ability to answer the Combatant
Commanders' demand signals. With FRP the Navy has deployed and
developed a more agile, flexible and scalable naval force capable of
surging quickly to deal with unexpected threats, humanitarian disasters
and contingency operations.
* A key objective in the FRP evolution is the continual evaluation and
re-evaluation of mission essential tasks, in response to Combatant
Commander demands, to create an adaptive capability earlier in the
training cycle. The Navy has and will continue to use both Unit Level
Training (ULT) and the ability to respond to Requests For Forces (RFF)
to evaluate both current training funding levels and FRP sustainment
funding levels. By aligning key training milestones in accordance with
national strategy, and Component and Combatant Commander
feedback/lessons learned our naval forces will deliver a tailored
combat ready force in the highest state of readiness.
* To specifically address the no-notice surge requirement, any FRP
surge is in fact no notice. Should world events require a large no-
notice surge, the Navy will capture those lessons learned and
incorporate improvements into the FRP. For example, Global War on
Terrorism Surge 05 (European and Central Command surges) was no notice.
The demand signal and supply signal were both generated in May 05 and
ships deployed 27 May 05, NWDC is currently analyzing. Other surge
examples:
- HSV surge in support of PACOM Tsunami relief:
- Hurricanes Katrina and Rita Navy response (HARRY S TRUMAN underway 96
hours after notification):
* Salvage/Recovery ship/EOD/Divers:
* Mine Countermeasure ships:
* Hospital ship (surge from reduced operational status):
* SEABEES/Cargo Handling/Helicopters/Tactical Air Control/AEW/Maritime
Preposition:
* Expeditionary medical facilities:
* Causeway transport:
- SAIPAN Expeditionary Strike Group readiness to surge in 96 hours to
support Haiti operations (JCS did not execute):
- SAIPAN Expeditionary Strike Group 96 hour readiness to surge to PACOM
in support of Tsunami relief operations (JCS did not execute):
- THEODORE ROOSEVELT Carrier Strike Group surged 7 days early from
planned deployment to meet CENTCOM increased demand signal for Iraqi
elections.
- LINCOLN Carrier Strike Group surge (from Emergency Surge status) to
cover for KITTY HAWK in maintenance:
- BOXER/BATAAN and KEARSARGE surge in support of Marine Corps Air
Combat Element transport to Iraq.
* The FRP was demonstrated in response to Operation Iraqi Freedom,
where 6 carriers where deployed to Iraq, and again during SUMMER PULSE
2004. During SUMMER PULSE over 40 Broad categories of data were
captured. Prior to conduct of the surge, analysis was conducted to
predict performance, and was then compared to SUMMER PULSE results, all
of which were in the expected (fiscally, flight hour, weapons
expenditure and steaming days constrained) and predicted readiness
levels. Navy analyzed:
- Cost/sorties/flight hours;
- Combat readiness/Mobility ratings;
- Steaming days/PASSEX events;
- Mil to Mil exchanges/Conduct of Argentinean and Brazilian aircraft
aboard USS REAGAN;
- Training events conducted/weapons expended/ranges utilized (TAN TAN
in Morocco noteworthy);
- Anti-submarine warfare time;
- Logistics/maintenance;
- Personnel qualification gains;
- Information Operations/Public Affairs;
- FRP total ordnance posture;
- Strategic/Operational and Tactical gains (Theater Security
Cooperation) with International organizations and foreign nations.
The Navy will continue to capitalize on the investments to the
readiness accounts under the Fleet Response Plan framework/operational
construct. FRP provides a superb base to meet the challenges faced in
the Global War on Terror and transforming, trans-national threats while
maintaining significant presence abroad; to assure our allies, dissuade
our adversaries and respond with scalable Major Combat Operations
capable forces. It provides the framework to rapidly adapt to new
mission requirements and meet the demands of Combatant Commanders. It
is an evolving operational construct that incorporates lessons learned,
adapts to the changing landscape and maintains the capability to manage
risk and fiscal responsibilities.
[End of section]
Appendix III: GAO Contact and Staff Acknowledgments:
GAO Contact:
Janet St. Laurent, (202) 512-4402 or stlaurentj@gao.gov:
Acknowledgments:
In addition to the contact named above, Richard Payne, Assistant
Director; Renee Brown; Jonathan Clark; Nicole Collier; Dawn Godfrey;
David Marroni; Bethann Ritter; Roderick Rodgers; John Van Schaik; and
Rebecca Shea made significant contributions to this report.
FOOTNOTES
[1] The Fleet Response Plan is also known as the Fleet Readiness
Program, or simply "FRP."
[2] See GAO, Highlights of a GAO Forum on Mergers and Transformation:
Lessons Learned for a Department of Homeland Security and Other Federal
Agencies, GAO-03-293SP (Washington, D.C.: Nov. 14, 2002); and Results-
Oriented Cultures: Implementation Steps to Assist Mergers and
Organizational Transformations, GAO-03-669 (Washington, D.C.: July 2,
2003).
[3] See GAO, Defense Logistics: GAO's Observations on Maintenance
Aspects of the Navy's Fleet Response Plan, GAO-04-724R (Washington,
D.C.: June 18, 2004).
[4] Emergency surge status means a unit can be employed in case of
urgent need but does so at levels of operational risk correlating to
the level of capability achieved at the time of emergency surge. Surge-
ready status means that units are ready to be employed at more
acceptable levels of operational risk commensurate with the level of
capability achieved at the time of the requirement to surge. Routine
deployable status means a unit has achieved all required capabilities,
completed underway training requirements, and is certified for forward
deployed operations.
[5] Unit-level phase training focuses on completion of unit-level
training requirements, including team training both on board and
ashore; unit-level exercises in port and at sea; and unit inspections,
assessments, certifications, and qualifications. During this phase, a
unit becomes proficient in all required capabilities, meets the
training commander's certification criteria, and becomes ready for more
complex integrated training events. Integrated phase training brings
individual units together to conduct strike-group-level integrated
training and operations in a challenging operational environment as a
foundation for performing their anticipated deployed mission.
Sustainment phase training exercises units and staffs in multimission
planning and execution, including the ability to interoperate
effectively in a wartime environment.
[6] The maintenance phase consists of depot-level shipyard maintenance
for a period of 6 to 12 months. Deep maintenance is for a period of 2
to 3 years for a nuclear-refueling overhaul.
[7] The Congress enacted the Government Performance and Results Act of
1993 to provide for, among other things, the establishment of strategic
planning and performance measurement in the federal government. Pub. L.
No. 103-62 (1993).
[8] We met with people individually and in groups. Personnel included
commanding officers, executive officers, and senior and junior enlisted
personnel. Additionally, we met with command master chiefs and command
career counselors.
[9] Tabletop exercises are analytical tools that require fewer
resources than full-fledged live exercises. They provide a means to
develop both immediate and long-term solutions among functional areas,
to develop standardization and interoperability of procedures, and to
document best practices for others to utilize.
[10] Army Regulation 5-5, Army Studies and Analyses (June 30, 1996).
[11] Rear Admiral John D. Stufflebeem, "Roundtable: Summer Pulse
Discussions With Rear Admiral John D. Stufflebeem," Summer Pulse '04
News Archive (July 8, 2004),
http://www.cffc.navy.mil/summerpulse04/stufflebeem-transcript.htm.
[12] See GAO, Quantitative Data Analysis: An Introduction, GAO/PEMD-
10.1.11, ch. 7 (Washington, D.C.: May 1992).
[13] The Navy Warfare Development Command's responsibilities include
being a champion for Navy warfare innovation, operating concepts, and
concept of operations development in a naval, joint, and coalition
environment; coordinating the planning and implementation of the Navy's
experimentation process; managing development, approval, and rapid
dissemination of naval, joint, and allied doctrine; and managing the
Navy's Lessons Learned Program.
[14] No-notice exercises demonstrate participants' ability to rapidly
respond to unexpected situations. This type of exercise is valued
because it can lead to improvements in procedures by exercising
participants in a near-real-world context.
[15] In the spring of 2005, the Navy surged five ships in support of
the Global War on Terrorism to work with allies to detect, disrupt, and
deny international terrorist organizations the use of the maritime
environment. These ships will also work to build regional security and
long-term stability. The five ships are the U.S.S. Saipan (LHA 2),
U.S.S. Nashville (LPD 13), U.S.S. Nicholas (FFG 47), the U.S.S. Gunston
Hall (LSD 44), and the U.S.S. Philippine Sea (CG 58).
[16] A request for forces is a special request by a geographic
combatant commander through the Chairman, Joint Chiefs of Staff, for
forces in addition to the normal, preapproved force deployments or
forces assigned.
[17] An after-action report is a professional discussion of an event,
focused on performance standards, that enables participants to discover
for themselves what happened, why it happened, and how to sustain
strengths and address weaknesses. It is a tool that leaders, trainers,
and units can use to get maximum benefit from every mission or task.
[18] The Joint Quarterly Readiness Review is a quarterly readiness
assessment that identifies capability shortfalls and risks in mission
execution and identifies appropriate measures for risk reduction.
[19] See GAO, Chemical and Biological Defense: Army and Marine Corps
Need to Establish Minimum Training Tasks and Improve Reporting for
Combat Training Centers, GAO-05-8 (Washington, D.C.: Jan. 28, 2005),
and Force Structure: Navy Needs to Fully Evaluate Options and Provide
Standard Guidance for Implementing Surface Ship Rotational Crewing,
GAO- 05-10 (Washington, D.C.: Nov. 11, 2004).
[20] See GAO, Military Training: Potential to Use Lessons Learned to
Avoid Past Mistakes Is Largely Untapped, GAO/NSIAD-95-152 (Washington,
D.C.: Aug. 9, 1995).
[21] Chief of Naval Operations Instruction 3500.37D, currently being
drafted, would replace Chief of Naval Operations Instruction 3500.37C,
March 19, 2001, Navy Lessons Learned System.
[22] See GAO, Defense Reorganization: Progress and Concerns at JCS and
Combatant Commands, GAO/NSIAD-89-83 (Washington, D.C.: Mar. 1, 1989).
[23] See GAO, Nuclear Safety: Potential Security Weaknesses at Los
Alamos and Other DOE Facilities, GAO/RCED-91-12 (Washington, D.C.: Oct.
11, 1990).
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: