Best Practices
Better Support of Weapon System Program Managers Needed to Improve Outcomes
Gao ID: GAO-06-110 November 30, 2005
The Department of Defense (DOD) relies on a relatively small cadre of officials to develop and deliver weapon systems. In view of the importance of DOD's investment in weapon systems, we have undertaken an extensive body of work that examines DOD's acquisition issues from a perspective that draws lessons learned from the best commercial product development efforts to see if they apply to weapon system acquisitions. In response to a request from the Chairman and Ranking Minority Member of the Subcommittee on Readiness and Management Support, Senate Committee on Armed Services, this report assesses (1) how successful commercial companies position their program managers, (2) how DOD positions its program managers, and (3) underlying reasons for the differences. In compiling this report, GAO conducted a survey of program managers. See GAO-06-112SP.
U.S. weapons are among the best in the world, but the programs to acquire them often take significantly longer and cost more money than promised and often deliver fewer quantities and capabilities than planned. It is not unusual for estimates of time and money to be off by 20 to 50 percent. When costs and schedules increase, quantities are cut, and the value for the warfighter--as well as the value of the investment dollar--is reduced. When we examined private sector companies that developed complex and technical products similar to DOD, we found that their success hinged on the tone set by leadership and disciplined, knowledge-based processes for product development and execution. More specifically, long before the initiation of a new program, senior company leaders made critical investment decisions about the firm's mix of products so that they could commit to programs they determined best fit within their overall goals. These decisions considered long-term needs versus wants as well as affordability and sustainability. Once high level investment decisions were made, senior leaders ensured that programs did not begin unless they had a business case that made sure resources were in-hand to execute the program--that is, time, technology, money, and people. Once a business case was established, senior leaders tasked program managers with executing that business case for each new product from initiation to delivery, but required their program managers to use a knowledge-based product development process that demanded appropriate demonstrations of technology, designs, and processes at critical junctures. The program manager was empowered to execute the business case, but also held accountable for delivering the right product at the right time for the right cost. Requiring the program manager to stay throughout the length of a project was a principal means of enforcing accountability. Overall, by providing the right foundation and support for program managers, the companies we visited were able to consistently deliver quality products within targets, and in turn, transform themselves into highly competitive organizations. DOD program managers are put in a very different situation. DOD leadership rarely separates long-term wants from needs based on credible, future threats. As a result, DOD starts many more programs than it can afford--creating a competition for funds that pressures program managers to produce optimistic cost estimates and to overpromise capabilities. Moreover, our work has shown that DOD allows programs to begin without establishing a formal business case. And once they begin, requirements and funding change over time. In fact, program managers personally consider requirements and funding instability--which occur throughout the program--to be their biggest obstacles to success. Program managers also believe that they are not sufficiently empowered to execute their programs, and that because much remains outside of their span of control, they cannot be held accountable.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-06-110, Best Practices: Better Support of Weapon System Program Managers Needed to Improve Outcomes
This is the accessible text file for GAO report number GAO-06-110
entitled 'Best Practices: Better Support of Weapon System Program
Managers Needed to Improve Outcomes' which was released on December 1,
2005.
This text file was formatted by the U.S. Government Accountability
Office (GAO) to be accessible to users with visual impairments, as part
of a longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to the Subcommittee on Readiness and Management Support,
Committee on Armed Services, U.S. Senate:
United States Government Accountability Office:
GAO:
November 2005:
best practices:
Better Support of Weapon System Program Managers Needed to Improve
Outcomes:
Best Practices:
GAO-06-110:
GAO Highlights:
Highlights of GAO-06-110, a report to the Subcommittee on Readiness and
Management Support, Committee on Armed Services, U.S. Senate:
Why GAO Did This Study:
The Department of Defense (DOD) relies on a relatively small cadre of
officials to develop and deliver weapon systems. In view of the
importance of DOD‘s investment in weapon systems, we have undertaken an
extensive body of work that examines DOD‘s acquisition issues from a
perspective that draws lessons learned from the best commercial product
development efforts to see if they apply to weapon system acquisitions.
In response to a request from the Chairman and Ranking Minority Member
of the Subcommittee on Readiness and Management Support, Senate
Committee on Armed Services, this report assesses (1) how successful
commercial companies position their program managers, (2) how DOD
positions its program managers, and (3) underlying reasons for the
differences. In compiling this report, GAO conducted a survey of
program managers. See GAO-06-112SP.
What GAO Found:
U.S. weapons are among the best in the world, but the programs to
acquire them often take significantly longer and cost more money than
promised and often deliver fewer quantities and capabilities than
planned. It is not unusual for estimates of time and money to be off by
20 to 50 percent. When costs and schedules increase, quantities are
cut, and the value for the warfighter”as well as the value of the
investment dollar”is reduced.
When we examined private sector companies that developed complex and
technical products similar to DOD, we found that their success hinged
on the tone set by leadership and disciplined, knowledge-based
processes for product development and execution. More specifically,
long before the initiation of a new program, senior company leaders
made critical investment decisions about the firm‘s mix of products so
that they could commit to programs they determined best fit within
their overall goals. These decisions considered long-term needs versus
wants as well as affordability and sustainability. Once high level
investment decisions were made, senior leaders ensured that programs
did not begin unless they had a business case that made sure resources
were in-hand to execute the program”that is, time, technology, money,
and people. Once a business case was established, senior leaders tasked
program managers with executing that business case for each new product
from initiation to delivery, but required their program managers to use
a knowledge-based product development process that demanded appropriate
demonstrations of technology, designs, and processes at critical
junctures. The program manager was empowered to execute the business
case, but also held accountable for delivering the right product at the
right time for the right cost. Requiring the program manager to stay
throughout the length of a project was a principal means of enforcing
accountability. Overall, by providing the right foundation and support
for program managers, the companies we visited were able to
consistently deliver quality products within targets, and in turn,
transform themselves into highly competitive organizations.
DOD program managers are put in a very different situation. DOD
leadership rarely separates long-term wants from needs based on
credible, future threats. As a result, DOD starts many more programs
than it can afford--creating a competition for funds that pressures
program managers to produce optimistic cost estimates and to
overpromise capabilities. Moreover, our work has shown that DOD allows
programs to begin without establishing a formal business case. And once
they begin, requirements and funding change over time. In fact, program
managers personally consider requirements and funding instability”which
occur throughout the program”to be their biggest obstacles to success.
Program managers also believe that they are not sufficiently empowered
to execute their programs, and that because much remains outside of
their span of control, they cannot be held accountable.
What GAO Recommends:
GAO recommends the Secretary of Defense develop an investment strategy
to prioritize needed capabilities; require senior stakeholders to
formally commit to business cases for new weapon system developments;
and develop a process to instill and sustain accountability for
successful program outcomes. DOD agreed with our recommendations.
www.gao.gov/cgi-bin/getrpt?GAO-06-110
www.gao.gov/cgi-bin/getrpt?GAO-06-112SP
To view the full product, including the scope
and methodology, click on the links above.
For more information, contact Michael J. Sullivan at (202) 512-4841 or
sullivanm@gao.gov.
[End of section]
Contents:
Letter:
Executive Summary:
Purpose:
Background:
Results in Brief:
Best Practice: Corporate Leadership and Disciplined, Knowledge-Based
Processes Are Critical to Program Manager Success:
DOD: Critical Support Factors Are Missing:
Differences in Incentives Contribute to Differences in Support for
Program Managers:
Chapter 1: Introduction:
Long-Standing Problems Hamper Weapons Systems Acquisitions:
DOD Program Managers Are Central Executors of the Acquisition Process:
Legislation to Improve Program Manager Proficiency:
Objectives, Scope, and Methodology:
Chapter 2: Senior Leader Support and Disciplined Knowledge-Based
Processes Are Critical Enablers for Program Managers:
Senior Leadership Provides Program Managers with a Strong Foundation
for Success:
Knowledge-Based Process Followed to Execute Programs:
Continued Senior Leadership during Product Development Further Enabled
Success:
Chapter 3: DOD Is Not Supporting Its Program Managers Effectively:
Senior Leadership Does Not Provide a Strong Foundation for Success:
Execution in DOD Does Not Provide Adequate Support and Accountability:
Senior Leader Support during Execution:
Chapter 4: Basic Incentives Drive Differences in How Program Managers
Are Supported and Held Accountable:
Definition of Success:
Means for Success:
Other Differences Put Additional Pressures on DOD Program Managers:
Chapter 5: Conclusions and Recommendations:
Recommendations for Executive Action:
Agency Comments and Our Evaluation:
Appendix I: Comments from the Department of Defense:
Appendix II: GAO Staff Acknowledgments:
Related GAO Products:
Tables:
Table 1: Acquisition Categories:
Table 2: Are Best Practices Present in DOD?
Table 3: Technology Maturity and Program Outcomes:
Table 4: Are Best Practices Present in DOD?
Table 5: Program Manager Views on Formal vs. Informal Authority:
Figures:
Figure 1: Critical Support and Accountability Factors:
Figure 2: 2005 Toyota Avalon:
Figure 3: Siemens Bi-Plane AXIOM Artis:
Figure 4: Best Practice Roles, Responsibilities, and Behaviors of
Senior Managers:
Figure 5: Breakdowns in Support and Accountability Factors:
Figure 6: Highlights of Program Manager Comments Regarding Competition
for Funding:
Figure 7: To What Extent Were the Parameters of Your Program Reasonable
at Program Start?
Figure 8: How Program Managers Responded to an Open-ended Question on
What Were the Biggest Obstacles They Faced:
Figure 9: Highlights of Program Manager Comments on What Types of
Authority They Need:
Figure 10: Key Differences in Definition of Success and Resulting
Behaviors:
Figure 11: Commercial vs. DOD Oversight Environments:
Abbreviations:
DAWIA: Defense Acquisition Workforce Improvement Act:
DOD: Department of Defense:
OSD: Office of the Secretary of Defense:
PEO: program executive officer:
United States Government Accountability Office:
Washington, DC 20548:
November 30, 2005:
The Honorable John Ensign:
Chairman:
The Honorable Daniel K. Akaka:
Ranking Minority Member:
Subcommittee on Readiness and Management Support:
Committee on Armed Services:
United States Senate:
As you requested, this report examines how program managers in the
Department of Defense are supported and how they are held accountable
for program outcomes. It compares department polices and practices to
those of leading commercial companies we visited and discusses actions
DOD could take to improve the accountability of program managers and
provide them with timely support as they manage the development of
complex systems. We make recommendations to the Secretary of Defense to
(1) develop an investment strategy to prioritize needed capabilities,
(2) require, for each new program, that senior level stakeholders
formally commit to a business case for program approval at the start of
a new program, and (3) implement a process to instill and sustain
accountability for successful program outcomes.
We are sending copies of this report to the Secretary of Defense; the
Secretary of the Army; the Secretary of the Navy; the Secretary of the
Air Force; the Director, Missile Defense Agency; the Director of the
Office of Management and Budget; and interested congressional
committees. We will also make copies available to others upon request.
In addition, the report will be available at no charge on the GAO Web
site at http://www.gao.gov.
If you have any questions regarding this report, please call me at
(202) 512-4841. Staff acknowledgments are listed in appendix II.
Signed by:
Michael J. Sullivan:
Director, Acquisition and Sourcing Management:
[End of section]
Executive Summary:
Purpose:
The Department of Defense (DOD) plans to increase its investment in the
research, development, and procurement of new weapon systems from $144
billion in fiscal year 2005 to $185 billion in fiscal year 2009. U.S.
weapons are among the best in the world, but the programs to acquire
them often take significantly longer and cost more money than promised
and often deliver fewer quantities and other capabilities than planned.
It is not unusual for estimates of time and money to be off by 20 to 50
percent. When costs and schedules increase, quantities are cut, and the
value for the warfighter--as well as the value of the investment
dollar--is reduced.
In view of the importance of DOD's investment in weapon systems, we
have undertaken an extensive body of work that examines DOD's
acquisition issues from a different, more cross-cutting perspective--
one that draws lessons learned from the best commercial product
development efforts to see if they apply to weapon system acquisitions.
In response to a request from the Chairman and Ranking Minority Member
of the Subcommittee on Readiness and Management Support, Senate
Committee on Armed Services, this report assesses (1) how successful
commercial companies position their program managers, (2) how DOD
positions its program managers, and (3) underlying reasons for the
differences.
Background:
DOD relies on a relatively small cadre of military and civilian
officials--known as program managers--to lead the development and
delivery of its weapon systems. The responsibility placed on this group
is enormous. The systems that program managers are responsible for
range from highly complex and sophisticated aircraft, missile
interceptors, submarines, and space-based sensors, to new communication
and ground control systems that support and interconnect this
equipment, to smaller, less complex systems that support the
warfighter. In these times of asymmetric threats and netcentricity,
individual weapon system investments are getting larger and more
complex. The development process itself is very challenging as many
systems require successful management and coordination of a broad array
of military service and DOD officials, outside suppliers, internal and
external oversight entities, as well as technical, business,
contracting, and management expertise. Moreover, in many cases, weapon
systems are also expected to incorporate technologies that push the
state-of-the-art while operating in harsh and even untested
environments--adding daunting technical challenges to the already
existing business, management, and logistical challenges. Lastly, GAO
has reported many of the business processes that support weapons
development--strategic planning and budgeting, human capital
management, infrastructure, financial management, information
technology, and contracting--are beset with pervasive, decades-old
management problems, which include outdated organizational structures,
systems, and processes.[Footnote 1]
Weapon system program managers are the central executors of the
acquisition process. They are responsible for all aspects of
development and delivery of a new system and for assuring that systems
are high quality, affordable, supportable, and effective. In carrying
out this responsibility, they are also responsible for balancing
factors that influence cost, schedule, and performance. DOD employs
about 729 program managers to run its weapons programs. Both military
officers and civilians serve as program managers, but the majority is
from the military. DOD's program managers typically report to program
executive officers (PEO) who are charged with overseeing the execution
of a portfolio of related systems. PEOs, in turn, typically report to a
military service acquisition executive, who reports to a service
secretary, or for some programs, the PEO reports to the Defense
Acquisition Executive.
Results in Brief:
Program managers from the leading companies we spoke with believed that
two critical enablers--(1) support from top leadership and (2)
disciplined, knowledge-based processes for product development
execution--empowered them to succeed in delivering new products when
needed within cost, quality, and performance targets originally set by
the company. Long before the initiation of a new product development,
senior company leaders make critical strategic investment decisions
about the firm's mix of products and the return on investment they may
yield. Once high-level investment decisions were made, senior leaders
ensured that programs did not begin unless they had a business case
that demonstrated the program was aligned with the company's goals and
that resources were in-hand to execute the program--that is, time,
technology, money, and people. Once a business case was established,
senior leaders tasked program managers with executing that business
case for each new product from initiation to delivery, but required
their program managers to use a knowledge-based product development
process that demanded appropriate demonstrations of technology,
designs, and processes at critical junctures. The program manager was
empowered to execute the business case, but also held accountable for
delivering the right product at the right time for the right cost.
Throughout execution, company senior leaders supported their program
managers by encouraging open and honest communication and continually
assured that the right levels of resources and management attention
were available for the project.
While DOD has taken action in recent years to better position programs
for success, it puts its program managers in a very different
situation. Program managers themselves believe that rather than making
strategic investment decisions, DOD starts more programs than it can
afford and rarely prioritizes them for funding purposes. The result is
a competition for funds that creates pressures to produce optimistic
cost and schedule estimates and to overpromise capability. Our own work
has shown that many programs begin without a business case, that is,
without adequate knowledge about technology, time, and cost and without
demonstrating that the program itself is the optimal approach for
achieving a needed capability. Moreover, once programs begin, the
program manager is not empowered to execute the program. In particular,
program managers cannot veto new requirements, control funding, or
control staff. In fact, program managers personally consider
requirements and funding instability to be their biggest obstacles to
success. Program managers also believe that they are not sufficiently
supported once programs begin. In fact, they must continually advocate
for their programs in order to sustain support. Our past reports also
show that programs are incentivized to suppress bad news and to
continually produce optimistic estimates--largely due to continual
funding competition.
Many of these differences can be attributed to how success is defined
within the commercial and DOD environment. Success for the commercial
world is straightforward and simple: maximize profit. In turn, this
means selling products to customers at the right price, the right time,
and the right cost. With this imperative in hand, companies have no
choice but to adopt processes and cultures that emphasize basing
decisions on knowledge, reducing risks prior to undertaking new
efforts, producing realistic cost and schedule estimates, and assuring
consistency and quality pervade all efforts. At first glance, DOD's
definition of success is very similar: deliver capability to the
warfighter at the right price, the right time, and the right cost. But,
for various reasons, it is clear that the implied definition for
success is to attract funds for new programs and to keep funds for
ongoing programs. While the annual appropriations process and the wide
variety of mission demands placed on DOD contribute to this condition,
DOD has made matters worse by not making hard tradeoff decisions to
ensure it does not pursue more programs than it can afford. Once
attracting funds becomes "success," harmful practices emerge. For
example, it is not in a program manager's interest to develop accurate
estimates of cost, schedule, and technology readiness, because honest
assessments could result in lost funding. Delayed testing becomes
preferred over early testing because that will keep "bad news" at bay.
In turn, knowing data being reported to them may not be reliable,
senior leaders believe they cannot trust it and must instill multiple
oversight mechanisms. Any attempts to improve policy and processes
eventually succumb to funding competition because no one wants to risk
loss of support.
We are making recommendations to DOD to better position program
managers for success. These recommendations focus on what is needed to
be done to provide the strategic leadership needed to provide the right
foundation for starting programs, ensure an executable business case is
delivered to program managers, and to hold program managers accountable
for successful outcomes. It is important to note that the success of
all of our recommendations hinge on DOD's ability to instill more
discipline and leadership over the investment process. After a review
of a draft of this report, DOD concurred with our recommendations and
provided some additional comments. The full text of DOD's comments may
be found in appendix I.
Best Practice: Corporate Leadership and Disciplined, Knowledge-Based
Processes Are Critical to Program Manager Success:
At all of the companies we visited, support for program managers began
well before they were assigned to a new product development effort--
with high-level strategic planning and investment decisions and
concerted efforts to make sure that any new initiative the company
undertook was achievable within the time and money and other resources
the company had available. Technology development and program advocacy
were also generally kept out of a program manager's domain. Once new
efforts got off the ground, program managers were empowered to manage
resources, encouraged to bring up problems and propose solutions, and
consult with senior leaders without fear of losing their support. At
the same time, however, they were expected to base their decisions on
hard data and to assure the right knowledge was in-hand before
proceeding into the next phases of development. They were also held
accountable for their choices, though companies generally found that
with good pre-program decisions, a good launch, a sound, disciplined
process for execution, and continued support, there was little need to
punish or remove their program managers. Ultimately, as long as a
program manager could deliver the right product at the right time for
the right cost, he was incentivized to do so without interference from
above.
According to commercial program managers we spoke with, the most
critical support factors included the following:
* Investment strategies. Each of the companies we visited followed a
rigorous process to forecast market needs against company resources,
economic trends, available technologies, and its own strategic vision.
These exercises culminated in short-and long-term investment strategies
that provided program managers with confidence that the company was
committed to their particular program and showed them where the project
fit within overall corporate goals.
* Evolutionary development. All of the companies followed an
incremental path toward meeting market needs rather than attempting to
satisfy all needs in a single step. This provided program managers with
more achievable requirements, which, in turn, facilitated shorter cycle
times. With shorter cycle times, the companies could ensure both
program managers and senior leaders stayed with programs throughout the
duration.
* Matching requirements to resources. Once specific product concepts
were identified, the companies worked vigorously to close gaps between
requirements/customer needs, and resources--time, money, and
technology. In effect, this took the investment strategy down to a
project level, assuring that the program manager would be well
positioned to execute within cost and schedule.
* Matching the right people to the program. All of the companies we
visited took steps to ensure that they assigned the right people to the
right programs. These included long term efforts to train and groom
technical staff into program managers, mentoring on the part of senior
leaders with program management experience, handpicking program
managers based on their expertise and experience, and supporting
program managers with teams of highly qualified functional and
technical experts.
* Knowledge-driven development decisions. Once a new product
development began, program managers and senior leaders used
quantifiable data and demonstrable knowledge to make go/no-go
decisions. These covered critical program facets such as cost,
schedule, technology readiness, design readiness, production readiness,
relationships with suppliers, etc. Development was not allowed to
proceed until certain thresholds were met, for example, a high
percentage of engineering drawings completed or production processes
under statistical control. Development processes were also continually
tailored based on lessons learned. Program managers themselves placed
high value on these requirements, as they ensured programs were well
positioned to move into subsequent phases and were less likely to
encounter disruptive problems.
* Empowerment. At all the companies we visited, program managers were
empowered to make decisions as to whether programs were ready to move
forward and to resolve problems and implement solutions. They could
redirect available funding, if needed. They could change team members.
Prior to development, they often had a say in what requirements they
would be handed.
* Accountability. With authority, came accountability. Program managers
at all of the companies we visited were held accountable for their
choices. To assure accountability, senior leaders set goals that were
clear to the entire project team and provided incentives for program
managers and others to meet those goals.
* Tenure. All of the companies we visited required that program
managers stay on until the end of the program. This was a primary means
of assuring accountability.
* Continued senior leadership. In addition to empowering them, program
managers credited senior leaders with other vital levels of support.
Namely, senior leaders' commitment to their programs were unwavering,
they trusted their program managers, they encouraged them to share bad
news, and they encouraged collaboration and communication. At the end
of the day, it was the senior leaders' job to anticipate and remove
obstacles and provide the right levels of support so that the path was
cleared for the program manager to execute the program.
DOD: Critical Support Factors Are Missing:
At DOD, program managers are not put in a position to deliver a product
within estimates, nor are they held accountable when there are failures
to deliver products within estimates. While senior leaders work hard to
develop a short-and long-term vision for the defense of the United
States, these visions are rarely translated into realistic investment
strategies that assure the right mix of programs is being pursued.
Moreover, while recognized in policy as a best practice, DOD does not
always make sure that there is a business case for new initiatives.
Lastly, program managers are not empowered to execute programs once
they begin or held accountable when programs get off track.
The primary problem, according to many program managers and verified by
GAO's work, is that DOD starts more programs than it can afford and
does not prioritize programs for funding. This creates an environment
where programs must continually compete for funding. Before programs
are even started, advocates are incentivized to underestimate both cost
and schedule and overpromise capability.
A second problem is that gaps between resources and requirements are
not closed before or even during program development. For example, we
have reported that DOD allows many programs to go forward without
knowing whether critical technologies--such as satellite's main sensor,
a fighter aircraft's stealth technology, a new tank's networking
capability--can work as intended. Invariably, when programs start with
such unknowns, they spend a great deal of time and money later on
fixing technical glitches while simultaneously trying to get other
program aspects on track. One reason programs begin with immature
technologies is that program advocates are rushed to start the
acquisition program because it assures at least an initial commitment
of funding. Compounding this problem is the fact that acquisition
programs tend to attract funds over other activities, including science
and technology efforts that ultimately support acquisition. As a
result, program managers are incentivized to take on tasks that really
should be accomplished within a laboratory environment, where it is
easier and cheaper to discover and address technical problems.
A third problem is that program managers themselves are not empowered
to execute their programs. First, they have little control over funding
and they cannot count on funding to be stable. When funding is taken
away, program managers often find themselves in a negative spiral of
funding-related problems--particularly because they've already made
commitments to contractors based on certain anticipated levels of
funding. Second, they cannot veto new requirements. Faced with long
development life cycles and promising technology advances, users often
ask for new or better capabilities as a program proceeds forward.
Program managers themselves are not always empowered to say "no" to
demands that may overly stretch their programs, and few senior leaders
above them have been willing to. In addition, program managers have
little authority over staffing and the ability to shift funds within
the program. With so much outside their span of control, program
managers say that DOD is unable to hold them accountable when programs
get off track. Another reason that it is difficult to hold program
managers accountable is that their tenure is relatively short. The
problems being encountered today may well be the result of a poor
decision made years ago by another program manager.
DOD has tried to improve its processes and policies to better position
programs for success. For example, policies embrace the concept of
closing gaps between requirements and resources before launching new
programs, and DOD is making changes to requirements setting and funding
processes in an attempt to strengthen investment decisions. At this
point, however, program managers do not see trade-offs being made in
the front-end of product development that would ensure DOD could fully
commit to their programs and allow program managers themselves to focus
solely on executing their programs. The level of trust, collaboration
and communication is low, while the level of oversight and second
guessing is high.
Differences in Incentives Contribute to Differences in Support for
Program Managers:
Differences between how program managers are supported and held
accountable are rooted in differences in incentives and resulting
behaviors. This begins with the definition of success. The commercial
firms we studied concluded their survival hinged on their ability to
increase their market share, which, in turn, meant developing higher
quality products, at the lowest possible price, and delivering them in
a timely fashion--preferably before their competitors could do the
same. This imperative meant that they had no choice but to narrow the
gap between requirements and resources in a manner that not only
ensured they met their market targets, but did so in a manner that
consumed resources fairly efficiently. It also meant that they had no
choice but to fully support the development effort, instill strategic
planning and prioritization, work collaboratively, follow a knowledge-
based process that makes product development manageable, and,
ultimately, make everyone accountable for success. Ultimately, the
companies developed processes that embodied these tenets for success.
At the strategic level, these include accurate, strategic planning and
prioritization to ensure the right mix of products are pursued and
strong systems engineering to help them establish a realistic business
case. At the tactical level, companies developed development processes
that required certain thresholds of knowledge to be gained before a
decision to proceed forward is made.
In theory, DOD's success likewise hinges on its ability to deliver high-
quality weapons to the warfighter in a timely fashion. But in practice,
success is defined as the ability of a program to win support and
attract funds. Of course, there are reasons for this disconnect.
Corporate revenue is generated by customer sales while DOD's funding is
dependent on annual appropriations. Corporations go out of business
when their product development efforts do not succeed; DOD does not.
Selling products to customers is the single focus of a private-sector
company while DOD is charged with a myriad of important missions--each
of which also competes for budget share. Nevertheless, these conditions
create a vastly different set of processes and behaviors affecting
program managers. Program managers are incentivized, for example, to be
optimistic and suppress bad news because doing otherwise could result
in a loss of support and funding and further damage their program. In
short, unknowns become acceptable and desirable rather than
unacceptable as they are in the corporate environment. And
accountability becomes much more difficult to define.
[End of section]
Chapter 1:
Introduction:
DOD plans to spend about $1.3 trillion for its major programs between
2005 and 2009 and increase its investment in research and development
during that period by about 28 percent--from $144 billion to $185
billion. Although DOD's weapons are widely regarded as unrivaled in
superiority, DOD has not received a predictable return on investment in
major weapon systems acquisitions. For decades, many of DOD's weapon
systems acquisitions have experienced large cost increases and extended
schedules, which, in turn, have jeopardized performance and, more
broadly, undermined DOD's buying power.
To help better position DOD to successfully field weapons, we have
undertaken a body of work over the past decade that has examined
lessons learned from the best commercial product development efforts to
see if they can be applied to DOD weapon system development. Leading
commercial firms have developed increasingly sophisticated products in
significantly less time and at lower costs. Our previous best practices
reports[Footnote 2] have examined such topics as matching resources
with requirements, controlling total ownership costs, effective use of
testing, and product development. This report examines the program
manager's role and the mechanisms that DOD and leading commercial
companies use to position program managers for success and hold them
accountable. As the central executor of the acquisition process, DOD
depends on its program managers to efficiently and effectively run its
large range of complex weapon systems acquisitions.
The challenge that program managers now face is massive. Weapon systems
themselves are becoming increasingly sophisticated and interdependent
and, therefore, more complicated and difficult to develop. At the same
time, however, DOD is faced with threats that are constantly evolving,
requiring quicker development cycles and more flexibility within
weapons programs. Moreover, many of the business processes that support
weapons development--strategic planning and budgeting, human capital
management, infrastructure, financial management, information
technology, and contracting--are beset with pervasive, decades-old
management problems, including outdated organizational structures,
systems, and processes. In fact, these areas--along with weapons system
acquisitions--are on GAO's high-risk list of major government programs
and operations. Lastly, while DOD plans to considerably ramp up weapons
system spending in the next 5 years in an effort to dramatically
transform how it carries out its military operations, it is likely to
face considerable pressure to reduce its investment in new weapons as
the nation addresses long-term fiscal imbalances.
Long-Standing Problems Hamper Weapons Systems Acquisitions:
While DOD's acquisition process has produced weapons that are among the
best in the world, it also consistently yields undesirable
consequences--such as cost increases, late deliveries to the
warfighter, and performance shortfalls. Such problems have been
highlighted, for example, in our past reviews of DOD's F/A-22 Raptor,
Space-Based Infrared System, Airborne Laser, the Joint Strike Fighter,
and other programs. Our past work has found that problems occur because
DOD's weapon programs do not capture early on the requisite knowledge
that is needed to efficiently and effectively manage program risks. For
example, programs move forward with unrealistic cost and schedule
estimates, lack clearly defined and stable requirements, use immature
technologies to launch the product development, and fail to solidify
design and manufacturing processes at appropriate junctures in
development.
When costs and schedules increase, quantities are cut and the value for
the warfighter, as well as the value of the investment dollar, is
reduced. Moreover, in these times of asymmetric threats and
netcentricity, individual weapon system investments are getting larger
and more complex. Just 4 years ago, the top five weapon systems cost
about $281 billion; today, in the same base year dollars, the top five
weapon systems cost about $521 billion.[Footnote 3] If these
megasystems are managed with traditional margins of error, the
financial consequences--particularly the ripple effects on other
programs--can be dire.
DOD has long recognized such problems and initiated numerous
improvement efforts. In fact, between 1949 and 1986 five commissions
studied issues such as cycle time and cost increases as well as the
acquisition workforce. DOD has also undertaken a number of acquisition
reforms. Specifically, DOD has restructured its acquisition policy to
incorporate attributes of a knowledge-based acquisition model and has
reemphasized the discipline of systems engineering. In addition, DOD
recently introduced new policies to strengthen its budgeting and
requirements determination processes in order to plan and manage
systems based on joint warfighting capabilities. While these policy
changes are positive steps, we recently testified that implementation
in individual programs has not occurred because of inherent funding,
management, and cultural factors that lead managers to develop business
cases for new programs that overpromise on cost, delivery, and
performance of weapon systems.
DOD Program Managers Are Central Executors of the Acquisition Process:
DOD relies on a cadre of military and civilian officials--known as
program managers--to lead the development and delivery of hundreds of
weapon systems and subsystems. The services report a combined total of
729 program managers currently executing programs at all acquisition
category levels. The systems that program managers are responsible for
range from highly sophisticated air, land, sea, and space-based systems
to smaller, less complex communications or support equipment that
interconnects or supports larger systems. Program managers are
responsible for assuring that these systems are reliable, affordable,
supportable, and effective. They carry out multiple roles and
responsibilities and are expected to have a working knowledge in such
diverse areas as contracting, budgeting, systems engineering, and
testing.
DOD classifies its acquisition programs into categories based upon a
number of factors such as their size, cost, complexity and importance.
The largest, most complex and expensive programs generally fall under
the responsibility of the Under Secretary of Defense (Acquisition,
Technology and Logistics) while less complex and risky programs are
overseen by the service or component acquisition executive. Table 1
provides more details.
Table 1: Acquisition Categories:
Acquisition category: Category I;
Definition: Research, development, test, and evaluation > $365M
Procurement > $2.19B;
Milestone decision authority: 1D: Under Secretary of Defense
(Acquisition, Technology and Logistics);
1C: Service Acquisition Executive;
Program examples: Future Combat System; DD(X) Destroyer; B-1 Aircraft.
Acquisition category: Category II;
Definition: Research, development, test, and evaluation > $140M
Procurement > $660M;
Milestone decision authority: Service or Component Acquisition
Executive;
Program examples: All Source Analysis System; KC-130J Aircraft; Joint
Surveillance and Target Attack Radar System.
Acquisition category: Category III;
Definition: No fiscal criteria;
Program examples: 10k W Auxiliary Power Unit; Assault Breaching
Vehicle; C-5 Avionics.
Acquisition category: Category IV;
Definition: No fiscal criteria (Navy and Marine Corps only);
Program examples: C-130 Night Vision Lighting; Advanced Recovery
Control System.
Source: GAO.
Note: Category I systems are referred to as "programs" and smaller
related subsystems are called "projects" or "products." For example,
the Air Force's B-1 aircraft system-a category IC program-includes
category II and III projects that may have a designated manager.
Category 1D and 1C programs are distinguished by their milestone
decision authority.
[End of table]
Program managers typically supervise a large staff of engineers,
contracting personnel, logisticians, business, financial, and
administrative personnel. The number of people assigned to program
offices varies widely and depends on factors such as the complexity of
the system, the category level, and the availability of staff. For
example, the Joint Strike Fighter, a category ID program, is managing
the development of three configurations of a new aircraft for the Navy,
Marines and Air Force, and currently has about 200 government and
international personnel assigned. By contrast the Light Utility
Helicopter, a category II project relying largely on commercial off-
the-shelf components, has a staff of 34.
To successfully deliver a weapon system to the user, program managers
must also work with a range of individuals outside their sphere of
influence such as those charged with independent cost estimating,
testing, funding, writing requirements, security, and ensuring
interoperability. Simultaneously, the program manager is responsible
for overseeing, integrating, and evaluating the defense contractor's
work as the development progresses. Moreover, some program managers
lead international teams. For example, the Joint Strike Fighter Program
Office, in addition to the military, civilian, and contract team
members, has eight international partners and approximately 40
international team members.
The majority of DOD program managers for category I programs are
military officers at the rank of colonel or (Navy) captain. Subsystem
program managers are usually lower in rank and report directly to the
system program manager. DOD also employs civilian program managers,
usually GS-15s for its category I programs. As a rule, program managers
report to a Program Executive Officer--a civilian at the senior
executive level or military officer at the general officer rank--who
typically manages a portfolio of related weapon systems. However, some
program executive officers are responsible for a single large program,
such as the Joint Strike Fighter or the F-22 aircraft. One level up
from the program executive officer is the Service Acquisition
Executive, a civilian (often a political appointee) who reports to the
service Secretary. Programs classified as a category ID report through
the defense acquisition executive, Undersecretary of Defense
(Acquisition, Technology and Logistics), as their milestone decision
authority.
Legislation to Improve Program Manager Proficiency:
Program manager training and tenure is now governed by legislation
known as the Defense Acquisition Workforce Improvement Act
(DAWIA),[Footnote 4] enacted in 1990 after studies showed that a key
problem affecting acquisitions was that program managers did not stay
in their positions long enough to be accountable for outcomes and that
many simply lacked the training and experience needed to assume their
leadership roles. Congress amended the law in the fiscal year 2004 and
2005 defense authorization acts to allow the Secretary of Defense more
flexibility to tailor tenure, experience, and education qualifications
for program managers.
The act specifically created a formal acquisition corps and defined
educational, experience, and tenure criteria needed for key positions,
including program managers as well as contracting officers and others
involved the acquisition process. The act also provided for the
establishment of a defense acquisition university to provide
educational development and training for acquisition personnel. Under
DOD regulations program managers are required to attend training and
meet course requirements through the university in order to meet
certification requirements for the program management track.
There are three progressive certification levels: basic, intermediate,
and advanced. Program managers of major defense acquisition programs
are required to have Level 3 certification, which requires four years
of acquisition experience and an advanced level Defense Acquisition
University course in program management. DOD prefers that individuals
with Level 3 certification have a Master's degree in engineering,
systems acquisition management, or business administration, and
complete additional external coursework in relevant fields.
Objectives, Scope, and Methodology:
The Chairman and the Ranking Member, Subcommittee on Readiness and
Management Support, Senate Committee on Armed Services, requested that
we examine best practices and DOD procedures for factors that affect
program manager effectiveness. Our overall objectives for this report
were to (1) identify best practices that have enabled organizations to
successfully position their program managers for success, (2) identify
DOD practices for supporting program managers and holding them
accountable, and (3) compare and contrast DOD and commercial practices
in order to identify possible improvements to DOD practices.
To identify the best practices and processes that commercial companies
employ to position their program managers for success, we used a case
study methodology. We selected companies that, like DOD, research,
develop, and field products, using program managers as the central
executors of the programs. Selection of the companies was also based
upon recognition by the American Productivity and Quality Center and
the Project Management Institute and the recommendations of experts.
Below are descriptions of the three companies that are specifically
featured in this report.
* Toyota Motor Manufacturing of North America, Inc.
Toyota Motor Manufacturing of North America, Inc., the third largest
automobile producer in the world and the fifth largest industrial
company in the world, designs, manufactures, and markets cars, trucks,
and buses worldwide. In 2005, the company reported total net sales of
$172.7 billion. We met with individuals involved with the development
of the 2005 Toyota Avalon, a full-size sedan, at Toyota Motor
Manufacturing in Erlanger, Kentucky.
* Siemens Medical Solutions USA, Inc.
Siemens Medical Solutions is one of the world's largest suppliers in
the healthcare industry. Siemens Medical manufactures and markets a
wide range of medical equipment, including magnetic resonance imaging
systems, radiation therapy equipment, ultrasound equipment, and patient
monitoring systems. We met with individuals from the Angiography,
Cardiology, and Neurology business unit, located in Hoffman Estates,
Illinois.
* Motorola, Inc.
Motorola is a Fortune 100 global communications leader that provides
seamless mobility products and solutions across broadband, embedded
systems and wireless networks. Seamless mobility harnesses the power of
technology convergence and enables smarter, faster, cost-effective,
flexible communication in homes, autos, workplaces and all spaces in
between. Motorola had sales of $31.3 billion in 2004. We visited its
offices in Arlington Heights, Illinois, and discussed program
management practices and processes with representatives from the
Networks sector.
In addition to the three companies featured in this report, we visited
two additional successful firms to assess whether they employed similar
processes and practices for program management. These include Molson
Coors Brewing Company and Wells Fargo. Both companies have undertaken
projects that reflect some of the complexity and challenges that a DOD
weapon systems program would face. For example, we met with managers of
a Molson project intended to automate day-to-day marketing operations
for digital assets. We also met with Wells Fargo officials who
developed an electronic imaging process for paperless check clearance.
At both companies, we also discussed broader corporate investment
processes that supported these particular internal projects as well as
the companies' main service lines.
For each of the five companies, we interviewed senior management
officials and program managers to gather consistent information about
processes, practices, and metrics the companies use to support program
managers and hold them accountable. In addition to the case studies, we
synthesized information from GAO's past best practices work about
product development.
We also examined key best practices studies related to program
management, including studies from organizations such as the Project
Management Institute and the American Productivity and Quality Center.
Moreover, we relied on our previous best practice studies, which have
examined incentives and pressures affecting weapon system programs, the
optimal levels of knowledge needed to successfully execute programs,
and complementary management practices and processes that have helped
commercial and DOD programs to reduce costs and cycle time.
In order to determine DOD practices for supporting program managers and
holding them accountable, we conducted five separate focus groups
between July and October 2004. Each group was composed of project
managers from one of the services or the Missile Defense Agency. A
total of 28 acquisition category I program managers representing a
range of DOD programs were identified by their respective services for
the meetings held in separate locations in Huntsville, Ala; El Segundo,
Calif; Dayton, Ohio; Arlington, Va; and Ft. Belvoir, Va. For each focus
group, the facilitators introduced discussion topics to discover how
program managers define success, as well as what they are accountable
for and how they are held accountable. In addition, participants were
asked to discuss how program managers are supported and what obstacles
they encounter in performing their duties.
We analyzed the content of focus group transcripts and used the themes
we identified to design a survey to gather information about
acquisition category I and II program managers' perceptions about
factors that assist or block their success and to more clearly define
other issues in the DOD acquisition process that affect program manager
effectiveness. We elicited input from several experts-retired program
managers, active-duty members with program management experience, and
senior acquisition officials who reviewed the questions and provided
feedback on the draft survey.
We pretested the survey with five program managers. During the pretest
we asked the program managers questions to determine whether (1) the
survey questions were clear, (2) the terms used were precise, (3) the
questionnaire placed an undue burden on the respondents, and (4) the
questions were unbiased. We then incorporated their comments into the
survey, finalized the questions, and sent the web-based survey to
acquisition category I and II program managers. We selected the
category I and II program managers because they manage the more complex
and expensive programs. We identified the program managers through
consultation with each of the services. The survey consisted of open-
ended and close-ended questions concerning support for program managers
and how they are held accountable for program outcomes. Originally we e-
mailed 237 program managers but later determined that 52 should not be
included because they managed programs other than acquisition category
I and II. Of the 185 remaining program managers, we received completed
surveys from 69 percent.
The surveys were conducted using self-administered electronic
questionnaires posted on the World Wide Web. We sent e-mail
notifications to all acquisition category I and II program managers on
April 12, 2005. We then sent each potential respondent a unique
password and username by e-mail to ensure that only members of the
target population could participate in the survey. To encourage
respondents to complete the questionnaire, we began sending e-mail
messages to prompt each nonrespondent between April 26, 2005 and May
19, 2005. Additionally, the team contacted nonrespondents through
telephone calls between May 31, 2005 and July 12, 2005. We closed the
survey on July 19, 2005.
In this report we discuss some of the results obtained from the survey.
A more complete tabulation of survey questions together with tables
indicating the levels of response can be found on our Web site at GAO-
06-112SP. The survey contained close-ended questions and open-ended
questions. We conducted a content analysis of the open-ended questions
and constructed tables showing the results of the analysis arranged
into broad categories. Some of the respondents to our survey provided
more than one answer to the open-ended questions. All responses that
indicated equally important factors were tabulated in the appropriate
categories. However, because some respondents provided more than one
answer, the percentages may add up to more than 100 percent of
respondents. The web-based report does not contain all the results from
the survey. For example, we do not report responses for questions about
demographics, some open-ended questions, or questions with high item
nonresponse rates.
In addition to the focus groups and survey, we conducted in-depth
interviews with individual program managers, program executive officers
from across the services, as well as program managers from Boeing and
Lockheed Martin for two major weapon systems. To further assess the
conditions and environment program managers were operating in, we
relied on previous GAO reports. For example, we relied on a recent
study of space acquisition problems that incorporated interviews of
more than 40 individuals, including experienced program managers,
program executive officials, officials responsible for science and
technology activities, and former and current officials within the
Office of Secretary of Defense who have specific responsibility for
space system oversight or more general weapon system oversight.
To further determine relevant DOD policies and practices, we analyzed
documents describing the roles and responsibilities of program
managers, acquisition force career management, promotion rates,
performance reporting, and training requirements. Moreover, we analyzed
relevant legislation and the DOD 5000 series of directives and
instructions. We also interviewed career acquisition service officials,
Defense Acquisition University course managers, and the Director of
Training. We reviewed studies from the Rand Corporation, the Center for
Strategic and International Studies, and the Defense Science Board,
among others, on weapons system program management and acquisition
issues as well as studies performed by past commissions focused on
acquisition reform.
We conducted our review between April 2004 and November 2005 in
accordance with generally accepted government auditing standards.
[End of section]
Chapter 2: Senior Leader Support and Disciplined Knowledge-Based
Processes Are Critical Enablers for Program Managers:
Program managers from the leading companies we spoke with believed that
two critical enablers--(1) support from top leadership and (2)
disciplined, knowledge-based processes for strategic investment,
program selection, and product development execution--empowered them to
succeed in delivering new products when needed within cost, quality,
and performance targets originally set by the company. At all of the
companies we visited, corporate leadership began at a strategic level,
long before the initiation of a new product development, with senior
company leaders making critical strategic investment decisions about
the firm's mix of products and the return on investment they may yield.
Once high-level investment decisions were made, senior leaders assured
that new programs did not begin until there was a business case for
them--meaning there was assurance that the program fit in with the
corporation's goals and investment strategy and that there were
resources available to execute the program. Once a business case had
been made, senior leaders selected and tasked program managers with
executing the program. They also required the program managers to use a
knowledge-based product development process that demanded appropriate
demonstrations of technology, designs, and processes at critical
junctures. They also empowered program managers as appropriate to
execute the program and held them accountable for delivering the
program within estimates. While they were empowered to execute the
program, program managers were still supported by senior leaders, who
encouraged open and honest communication and continually assured that
the right levels of resources and management attention were available
for the project. Figure 1 maps critical support and accountability
factors.
Figure 1: Critical Support and Accountability Factors:
[See PDF for image]
[End of figure]
Senior Leadership Provides Program Managers with a Strong Foundation
for Success:
At each of the companies we visited, senior leaders invested a great
deal of time and effort positioning new development efforts for
success. Before even considering initiating a new project, senior
leaders made high level trade-off decisions between their long-term
corporate goals, projected resources, market needs, and alternative
ways of meeting those needs. These efforts culminated in investment
strategies that assured that the company could fully commit to any
product development effort it pursued. With a broad strategy in place,
senior leaders would then begin concept development for potential new
products, analyzing proposed products in terms of what requirements
could be achieved today versus future versions of the product and what
resources would be needed--not just in terms of cost, but in terms of
technologies, time, and people. Once a specific concept was selected,
senior leaders would follow rigorous systems engineering processes to
narrow the gap between requirements and resources to a point where they
were assured that they were pursuing a product that would meet market
needs and could be developed within cost and schedule goals. The end
point of this process was a sound business case that senior leaders
could then hand off to a program manager--who was then empowered to
deliver the product on time and within cost. Program managers
themselves highly valued this support because it ensured the companies
were committed to their particular efforts, reduced the level of
unknowns that they were facing, and kept them focused solely on
executing their programs. Put more simply, they believed senior leaders
consistently provided a sound foundation on which they could launch
their programs.
The most critical characteristics of the strategic leadership provided
include the following:
* Investment strategies. Because there are more product ideas than
there is funding to pursue them, the commercial companies we visited
used a knowledge-based process to make decisions about which product
development efforts to invest in. They began by developing an
investment strategy that supports a corporate vision. For the most
profitable mix of new products, companies analyzed factors such as
customer needs, available technology, and available resources.
Companies ensured that decisions to start new product developments fit
within the investment strategy. The investment strategy determined
project priority as well as providing a basis for trade-off decisions
against competing projects. Program managers found their company's use
of investment strategies helpful because it gave them confidence that
their project had commitment from their organization and from their top
leaders and managers and clearly identified where their project stood
within the company's overall investment portfolio and funding
priorities.
* Evolutionary development. All of the companies generally followed an
evolutionary path toward meeting market needs rather than attempting to
satisfy all needs in a single step. In effect, the companies evolved
products, continuously improving their performance as new technologies
and methods allow. These evolutionary improvements to products
eventually result in full desired capability, but in multiple steps,
delivering enhanced capability to the customer more quickly through a
series of interim products. For example, the 2005 Avalon involved
redesign of about 60 per cent of the vehicle, but component sections
such as the electronics and such features as the keyless ignition
system and the reclining rear seat were either developed by suppliers
or had been used on the Lexus. By using this method, the company
changed the Avalon's overall design and functionality by increments. In
more strategic investment planning, Toyota maintains an ongoing
research into such technology areas as alternative fueled automobiles
and environmental implications of automotive developments that will
feed into its long-term planning. Our previous work has found that this
approach reduces the amount of risk in the development of each
increment, facilitating greater success in meeting cost, schedule, and
performance requirements. The approach permits program managers to
focus more on design and manufacturing with a limited array of new
content and technologies in a program. It also ensures that the company
has the requisite knowledge for a product's design before investing in
the development of manufacturing processes and facilities. Conversely,
our past work has found that organizations that set exceedingly high
technology advancement goals invariably spend more time and money than
anticipated trying to address technology-related challenges amid other
product development activities, including design and production
stabilization.
* Matching of Requirements and Resources. The companies we visited were
able to achieve their overall investment goals by matching requirements
to resources--that is time, money, technology, and people--before
undertaking a new development effort. Any gaps that existed were
relatively small, and it was the program manager's job to quickly close
them as development began. More specifically:
* The companies had already extensively researched and defined
requirements to ensure that they are achievable given available
resources before initiating new efforts.
* Technologies were mature at the start of a program, that is, they had
been proven to work as intended. More ambitious technology development
efforts were assigned to corporate research departments until they were
ready to be added to future generations (increments) of the product. In
rare instances when less mature technologies were being pursued, the
company accepted the additional risk and planned for it.
* Companies committed to fully fund projects before they began. Not one
of the program managers we spoke with mentioned funding as a problem at
the beginning of a development effort and throughout. Funding was a
given once senior leaders had committed to their project.
* Systems engineering was typically used to close gaps between
resources and requirements before launching the development process. As
our previous work has shown, requirements analysis, the first phase of
any robust systems engineering regimen, is a process that enables the
product developer to translate customer wants into specific product
features for which requisite technological, software, engineering, and
production capabilities can be identified. Once these are identified, a
developer can assess its own capabilities to determine if gaps exist,
and then analyze and resolve them through investments, alternate
designs, and, ultimately, trade-offs. The companies we visited allowed
their engineers to analyze and weigh-in on the customers needs as
determined by its marketers.
Our previous best practice work has consistently found the practice of
matching requirements and resources prior to initiating a new program
to be a hallmark for successful companies. Simply put, we have found
that when wants and resources are matched before a product development
is started, the development is more likely to meet performance, cost,
and schedule objectives. When this match does not take place at the
start of a program, programs typically encounter problems such as
increased costs, schedule delays, and performance shortfalls as they
try to meet requirements during product development. Program managers
we spoke with for this review specifically cited this process as an
enabler for their own success because it ensured they were in a good
position to commit to cost and schedule estimates that were attainable,
and it did not require them to perform "heroic" efforts to overcome
problems resulting from large gaps between wants and resources, such as
technology challenges or funding shortages.
In addition to these critical strategic enablers, program managers at
the companies also stated that senior leaders made concerted efforts to
match program manager skills and experience to appropriate projects and
to train and mentor program managers. In selecting program managers
themselves, the companies placed high value on strong leadership
qualities, including decision making skills, diplomacy, communication
skills, ability to motivate others, and integrity, as well as how
individual personalities fit with the job or team. Most of the program
managers we spoke with had been groomed for their positions through
formal training on budgeting, scheduling, and regulatory compliance and
other aspects of program management; informal mentoring by senior
executives or experienced program managers; and by being placed in
positions that gradually increased their management responsibilities.
In addition, many of the program managers we spoke with also possessed
considerable technical experience. In fact, they often started at the
company as engineers. The companies we visited were similarly
deliberate in developing and deploying teams of functional experts to
support a program manager. In some cases, the teams reported directly
to the program manager. In others, they reported to their respective
home units and worked collaboratively with the program managers. In
either case, the program managers themselves valued the support they
were getting from these teams--particularly because they enabled the
program manager to employ a broad array of expertise from day-one of
the development effort and to facilitate an exchange of ideas. The
program managers we spoke with believed that their functional teams
were also highly skilled--to the point where they could easily delegate
major tasks.
Strategic Leadership at Toyota and Siemens Medical:
Strategic leadership for the development of Toyota's Avalon luxury
sedan ties back to conscious decisions made by senior leaders in Japan
when they built a Toyota facility in the United States 25 years ago. To
assure that the vehicles could be made to the same levels of quality as
those in Japan, Toyota replicated its manufacturing facilities, used
Japanese suppliers, and sent its managers to the United States to
supervise development. As U.S. employees gained experience and
demonstrated their capability, the reliance on Japanese suppliers and
personnel gradually decreased. A second step Toyota took was to
replicate its training and mentoring of program managers--pairing them
with more experienced chief engineers, who oversee long-term planning
across projects, and even bringing them to Japan to study how Toyota
approached development.
To support all of its new development efforts, senior leaders have
developed an overall strategic plan--which takes a long-and short-term
investment perspective. Over the long run, the plan envisions the
company achieving significant advancements in capabilities, such as
alternatively fueled engines, through incremental improvements to
technologies. Over the short run, a specific vehicle development
program uses a marketing analysis about features customers desire in
new models; and the staff determines whether a market exists for a
certain type of product at a certain price. In establishing a business
case for the Avalon, Toyota embarked on a formal concept development
effort, which was led by a chief engineer. The chief engineer, a high-
level executive, was largely responsible for setting the vision for the
new Avalon, securing resources needed for development effort prior to
initiating the development program, and working with representatives
from its sales division to make sure that the design and technologies
being pursued still fit within market needs--not just in terms of cost,
but in terms of vehicle features. A variety of functional experts were
consulted during this phase, though the chief engineer had the most
formal authority over concept development. At the conclusion of this
effort, Toyota decided to take on a very extensive redesign of the
Avalon but also set a goal bringing the vehicle to market in only 18
months. Redesign features included a reinforced body, improvements to
the engine and to the braking system, as well as features customers
desired such as a keyless ignition system and reclining rear seats.
Toyota leadership also decided to include mature technologies, often
borrowed from other vehicle lines, or purchased from outsider
suppliers. Once the design was approved, day to day project management
shifted to the Chief Production Engineer, whose responsibility it was
to see the vehicle through production to distribution.
Figure 2: 2005 Toyota Avalon:
[See PDF for image]
[End of figure]
Corporate leadership at Siemens Medical took a similar shape in the
development of new medical equipment. For example, senior leaders
developed an overall investment strategy, based largely on researching
their customers' technology needs as well as their own technology
readiness, the direction their competitors were going in, economic
trends, and projected manpower resources. From these assessments, a
team within Siemens developed a portfolio of potential new projects to
pursue, which upper management then prioritized based on their
potential profit, how they fit in with corporate goals and projected
resources. Ultimately, senior leaders produce a short-term (1 year)
investment plan as well as a longer-term (3 to 5 year) plan. Once a
specific project is selected, Siemens employs systems engineering
practices to narrow down the gap between customer requirements and
resources--working with both business and technical managers. A
"product manager" is charged with making trade-offs between
requirements, schedule, and cost prior to initiating product
development and is held accountable for systems engineering decisions
made to level requirements with resources for the business case. This
person sits at a relatively high level within the company and possesses
marketing and business expertise. A "project manager" who reports to
the product manager is ultimately assigned to execute the business
case, but he or she plays a role in the concept development by
participating in trade-off decisions and raising concerns about how
decisions can be executed.
At Siemens Medical, many project managers begin by serving as the
technical leader working with three to five people in systems
engineering or another technical area of a project. As the technical
team lead, they gain experience with scheduling, communicating, and
managing people. Over time the individual is given more
responsibilities such as becoming a subsystem project leader; as the
manager gains experience, he or she transitions to handling cross-
functional areas including business, budgeting, staffing, technology,
and testing.
Siemens Medical project managers are also given formal training,
including courses on regulatory and quality requirements as well as
courses that help program managers learn about their management styles.
In addition, Siemens ensures that project managers are well-trained on
risk management so that they can identify and mitigate potential risks
at the beginning of the project. Also, since project managers function
within a centralized project management department, they are mentored
both by the head of the department and by their peers.
Figure 3: Siemens Bi-Plane AXIOM Artis:
[See PDF for image]
[End of figure]
Knowledge-Based Process Followed to Execute Programs:
Once a new development effort began, program managers were empowered to
execute the business case and were held accountable for doing so. At
all of the companies we visited, program managers believed that
following a disciplined, knowledge-based development process and
continued support from senior leaders were essential to their success.
The process itself was typically characterized by a series of gates or
milestone decisions, which demanded programs assess readiness and
remaining risk within key sectors of the program as well as overall
cost and schedule issues, and it required go/no-go decisions to be made
fairly quickly. The most important aspect of the process, in the view
of the program managers, was that it empowered them to make decisions
about design and manufacturing trade-offs, supplier base, staffing on
the program team, etc.--as long as they were within the parameters of
the original business case. At the same time, the process held program
managers accountable and set clear goals and incentives.
Common critical characteristics of the knowledge-based process followed
to execute programs include the following:
* Knowledge-driven development decisions. Once a new product
development began, program managers and senior leaders used
quantifiable data and demonstrable knowledge to make go/no-go
decisions. These covered critical facets of the program such as cost,
schedule, technology readiness, design readiness, production readiness,
and relationships with suppliers. Development was not allowed to
proceed until certain thresholds were met, for example, a high
proportion of engineering drawings completed or production processes
under statistical control. Program managers themselves placed high
value on these requirements, as it ensured they were well positioned to
move into subsequent phases and were less likely to encounter
disruptive problems.
* Empowerment for program managers to make decisions. At all the
companies we visited, program managers were empowered to make decisions
on the direction of the program and to resolve problems and implement
solutions. They could make trade-offs among schedule, cost, and
performance features, as long as they stayed within the confines of the
original business case. When the business case changed, senior leaders
were brought in for consultation--at this point, they could become
responsible for trade-off decisions.
* Accountability. Program managers at all the companies we visited were
held accountable for their choices. Sometimes this accountability was
shared with the program team and/or senior leaders. Sometimes, it
resided solely with the program manager on the belief that company had
provided the necessary levels of support. In all cases, the process
itself clearly spelled out what the program manager was accountable
for--the specific cost, performance, schedule, and other goals that
needed to be achieved.
* Tenure. To further ensure accountability, program managers were also
required to stay with a project to its end. Sometimes senior leaders
were also required to stay. At the same time, program managers were
incentivized to succeed. If they met or exceeded their goals, they
received substantial bonuses and/or salary increases. Awards could also
be obtained if the company as a whole met larger objectives. In all
cases, companies refrained from removing a program manager in the midst
of a program. Instead, they chose first to assess whether more support
was needed in terms of resources for the program or support and
training for the program manager.
Other important aspects within the development process included the
following:
* Common templates and tools to support data gathering and analysis.
These tools included databases of demonstrated, historical cost,
schedule, quality, test, and performance data that helped program
managers produce metrics as well as standard forms and guidance for
conducting the meetings. Program managers valued these tools because
they greatly reduced the time needed to prepare for milestone meetings.
In all cases, program managers did not believe they were spending time
collecting data that was valuable to senior management but not to them.
* Common processes that supported product development. The companies
generally found that requiring program managers to employ similar risk
management, project management, requirements approval, testing, quality
management, problem resolution, and other processes enabled them to add
additional discipline and consistency to product development. Some
companies were certified by professional organizations as achieving the
highest level of proficiency within supporting development processes.
For example, Motorola was certified as a level 5 software development
organization by Carnegie Mellon's Software Engineering Institute.
* Lessons learned. All of the companies we visited continually refined
and enhanced their development process via some sort of lessons-learned
process. The program managers themselves placed a great deal of value
on these processes--as they were seen as the primary means for learning
how to tailor the process to better fit a project and to prevent the
same mistakes from recurring.
Program managers also cited flexibility as an enabling quality of their
processes. All of the companies allowed their processes to be tailored
as needed. Milestones that were deemed unnecessary could be dropped.
More often, however, additional meetings were added to gain consensus
on how to address particular problems. Another enabling factor was that
their processes ensured decisionmakers were not flooded with data.
Often, program teams boiled down data into one or two pages, using
simple metrics depicting status and risk on critical facets of the
program such as cost, schedule, technology readiness, design readiness,
and production readiness. Program managers valued the process of
translating detailed data into higher level metrics because it required
them to think about their programs in more strategic terms and focus on
the highest problem areas.
Knowledge-Based Development at Motorola and Toyota:
Motorola's development process is comprised of 16 milestones or
"gates"--the first five of which pertain to processes employed to
develop a product concept and the business case. Eleven gates comprise
the execution of the business case, from project initiation, to systems
requirements definition, design readiness, testing, controlled
introduction, and then full deployment. Each gate demands an array of
indicators on status and progress, including resources, cost, scope,
risk, and schedule. A centralized database helps program managers
produce this data and allows users to obtain data at any time and at
any level of detail that they need. For meetings themselves, program
managers are required to produce a set of "vital few" performance
measures relating to cost, quality, program status, and customer
satisfaction. At the gates themselves, program managers discuss the
status of the program with senior leaders, but they are ultimately
responsible for making decisions on whether to proceed to the next
phase. In the past, program managers did not have this responsibility
and acted more as an administrator than a leader, according to senior
executives. With less responsibility and accountability, programs were
not managed as well--often employing disjointed management processes
with less attention to efficiency and effectiveness. By increasing
program manager's ownership and accountability over the project, senior
leaders found that they were more incentivized to meet and exceed cost,
schedule, and performance goals. To support this change, the company
also adopted common supporting processes, including configuration
management, design, training, testing, defect prevention, quality
management, supplier management, and system upgrades. The common
processes assured program managers employed the same set of quality
controls and that deployed tools and guidance enabled program managers
to reduce cycle times as well as to produce better and more consistent
management data.
Toyota's process is comprised of eight key milestones--starting with a
lessons-learned gate. At this point, senior leaders and project teams
formally review what worked well and not so-well in the prior
development effort and assess whether the process needs to be tailored
as a result. The Avalon program manager told us that these
"reflections" are not taken lightly; they are developed through a very
detailed and soul-searching process during which people have to openly
admit errors and inadequacies so that better processes and procedures
can be devised. The next gate, "image" represents the process by which
the chief engineer derives the business case. Once he is done, direct
supervision of the project is transferred to a "chief production
engineer," who is charged with its execution of the business case
although the chief engineer continues to be involved in the
development. The next few milestones come as the car is designed,
prototyped, tested, and put through quality assurance. The last
milestone, the production stage also contains a customer feedback
phase, which is used to refine the next development effort.
Within the business case itself, Toyota places highest importance on
schedule because a number other vehicle development efforts are
dependent on the same resources and staff being used by the current
effort. As a result, the chief production engineer is more inclined to
make trade-offs that favor schedule over other factors. At each
milestone meeting, the chief production engineer reviews the status of
the program with senior leaders, focusing first on what problems are
occurring and what his solutions are for overcoming them. The meeting
itself employs streamlined reporting with simple indicators of
remaining risk on critical facets of the program--specifically, a
circle, meaning low remaining risk and okay to proceed; a triangle,
meaning there are problems but they can be fixed, or an "x," meaning
there is a problem without a solution. The chief production engineer is
responsible for making decisions as to how to proceed at these
milestones, unless there is a problem that significantly affects the
business case. If so, senior leaders become more involved in the
decision-making rather than simply advising the chief production
engineer.
While the Toyota process only employs eight formal milestones, the
chief production engineer actually involves functional experts, senior
executives, and other stakeholders in frequent meetings to make
tactical decisions about the program. For example, the Avalon chief
production engineer told us that he held "obeya" (literally "big room,"
signifying that all inputs are desired) meetings twice a week, which
involved all functional areas as well as "namewashi" (literally binding
the roots together, signifying gathering facts and moving toward a
decision) meetings before a formal milestone meeting--at which
functional officials consulted with each other to identify problems and
develop potential solutions that would be presented to senior leaders
at the milestone. Overall, the accountability for meeting the Avalon
program's goals was shared between the chief production engineer, the
functional team, and senior executives. At Toyota, senior leaders
assume that the processes they have in place will work, and if the
process is not delivering a suitable quality outcome, then it was the
shared responsibility of managers and staff to resolve the issue. If
performance issues arose, senior leaders attempted to address them
first through training, mentoring, and additional support, rather than
removing the program manager.
Continued Senior Leadership during Product Development Further Enabled
Success:
Empowering program managers to make decisions in executing the business
case was seen as the most significant type of support provided by
senior leaders. But program managers themselves pointed to other types
of support that made it easier for them to succeed. Primarily, senior
leaders did the following:
* Provided unwavering commitment to the development effort. At all the
firms we visited, senior leaders were champions of the project
throughout its life and fully committed to supporting it. When
significant problems arose that jeopardized the business case, they
found ways to address those problems, rather than rejecting the program
in favor of another one.
* Trusted their program managers. Senior leaders trusted the
information being provided by the program manager as well as his or her
expertise. This reduced the need for instilling additional layers of
oversight that could slow down the program. At the same time, however,
senior leaders took personal responsibility for assuring their program
managers had the knowledge and capability needed to succeed--in some
cases, by personally mentoring them for a long period of time.
* Encouraged program managers to share bad news. Senior leaders went
out of their way to encourage program managers to share problems. In
fact, program managers were often expected to discuss problems before
anything else at key milestones. And, in some cases, program managers
were evaluated based on their ability to identify and share problems.
At the same time, senior leaders expected their program managers to
come up with solutions--to take ownership over their efforts.
* Encouraged collaboration and communication. Senior leaders spent a
great deal of time breaking down stovepipes and other barriers to
sharing information. The Avalon chief production engineer, in fact,
told us that Toyota's development processes alone were much like other
automobile manufacturers he had worked for. What separated Toyota from
the others was its emphasis on open information exchange, cooperation,
and collaboration. He believed that this was the key enabler for
Toyota's superior systems integration.
Figure 4: Best Practice Roles, Responsibilities, and Behaviors of
Senior Managers:
[See PDF for image]
[End of figure]
[End of section]
Chapter 3: DOD Is Not Supporting Its Program Managers Effectively:
While DOD's leadership has taken action in recent years it hopes will
better position programs and improve planning and budgeting, it is
still not effectively positioning or supporting program managers for
success. For example, rather than making strategic investment
decisions, DOD starts more programs than it can afford and rarely
prioritizes them for funding purposes. The result is a competition for
funds that creates pressures to produce optimistic cost and schedule
estimates and to overpromise capability. Moreover, our work has shown
that DOD often starts programs without establishing a business case.
Specifically, technologies are not always mature at start, requirements
are not fully defined, and cost and schedule estimates are not always
realistic. In addition, program managers are not empowered to execute
programs. They cannot veto requirements and they do not control funding
or other resources. In fact, program managers who responded to our
survey personally consider requirements and funding instability to be
their biggest obstacles to success. Program managers also believe that
they are not sufficiently supported once programs begin. In particular,
they believe that program decisions are based on funding needs of other
programs rather than demonstrable knowledge; they lack tools needed to
enable them to provide leadership consistent with cost, schedule and
performance information; they are not trusted; they are not encouraged
to share bad news; and they must continually advocate for their
programs in order to sustain commitment.
Figure 5: Breakdowns in Support and Accountability Factors:
[See PDF for image]
[End of figure]
Senior Leadership Does Not Provide a Strong Foundation for Success:
According to program managers we interviewed as well as comments to our
survey and our past reviews, senior leadership within DOD does not
provide a strong foundation for success. While DOD is adept at
developing long-term visions and strategic plans, it does not develop
realistic, integrated investment strategies for weapons acquisitions to
carry out these plans. Instead, more programs are started than can be
funded and too many programs must compete for funding, which, in turn,
creates incentives to produce overly optimistic estimates and to
overpromise capability. Moreover, when faced with a lower budget,
program managers believe that senior executives within the Office of
the Secretary of Defense (OSD) and the services would rather make
across-the-board cuts to a span of programs rather than hard decisions
as to which ones to keep and which ones to cancel or cut back. Our work
continues to show that, while DOD has adopted evolutionary development
in its policies, programs are being encouraged to pursue significant
leaps in capability. In addition, DOD's policy now encourages programs
to match resources to requirements before program initiation, but
program managers reported in our survey that requirements and funding
are not stabilized and were the biggest obstacles to their success.
Further, while program managers believe their training has been
adequate, they also believe that mentoring has been uneven and that
they could benefit with tours of duty inside the Pentagon, for example,
in offices that oversee budget or financial management. Table 2
highlights differences between strategic senior leadership support
within the commercial companies we visited and DOD.
Table 2: Are Best Practices Present in DOD?
Best practices: Develop long-term vision and investment strategy;
DOD: DOD has long-term vision, but not an investment strategy. Lack of
investment strategy has created competition for funding and spurred low
cost-estimating, optimistic schedules, and suppression of bad news.
Best practices: Adopt evolutionary path toward meeting customer needs;
DOD: DOD has adopted evolutionary development in policy but not in
practice.
Best practices: Match requirements and resources before starting new
product development;
DOD: DOD has encouraged achieving match in policy but not in practice.
Requirements are not stable; funding commitments are not enforced;
key technologies are not matured before development. Requirements and
funding are biggest obstacles in view of program managers.
Source: GAO.
[End of table]
Investment Strategy and Evolutionary Development:
DOD is attempting to address some of the problems identified, but it is
too early to determine how effective its solutions are. For example, it
is implementing a new requirements setting processes--known as the
Joint Capabilities Integration and Development System--in an attempt to
bring more discipline to investment decisions. The system is organized
around key functional concepts and areas, such as command and control,
force application, battlespace awareness, and focused logistics. For
each area, boards of high-ranking military and civilian officials
identify long-term joint needs and make high-level trade-offs on how
those needs should be met. Once specific programs are proposed, the
process is designed to encourage a more evolutionary approach by
allowing requirements setters the flexibility to define requirements in
terms of capabilities as well as to defer final requirements
formulation to later in the development process. DOD has also been
attempting to implement complementary planning and budgeting processes-
-for example, by asking the military services to plan budgets around
guidance that takes a joint perspective and by taking a portfolio
planning approach. However, there is no evidence to date that shows
these enhancements are providing DOD with a sound investment strategy
as well as the right controls for enforcing that strategy.
While some program managers we spoke with believed the process' focus
on capabilities versus requirements promised more flexibility, program
managers comments to our survey show that they also still widely
believed that they were operating in an environment where there was
unfair competition for funding. Figure 6 highlights specific views.
Figure 6: Highlights of Program Manager Comments Regarding Competition
for Funding:
[See PDF for image]
[End of figure]
DOD has also adopted policies that encourage evolutionary
development.[Footnote 5] However, our reviews continue to find that
programs are still pursuing significant leaps in capabilities. For
example, we reported this year[Footnote 6] that the Joint Strike
Fighter acquisition strategy was striving to achieve the ultimate
fighter capability within a single product development increment, and
that it had bypassed early opportunities to trade or defer to later
increments those features and capabilities that could not be readily
met. We also testified[Footnote 7] that while DOD's space acquisition
policy states its preference for evolutionary development, programs
still attempt to achieve significant leaps in one step.
Matching Resources to Requirements:
In recent years, DOD has changed its acquisition policy to encourage
decisionmakers to match requirements to resources before starting a new
program. For example, the policy specifically encourages that
technologies be demonstrated in a relevant environment before being
included in a program; that a full funding commitment be made to a
program before it is started and that requirements be informed by the
systems engineering process. Concurrently, DOD's new requirements
process is designed to instill more discipline during initial
requirements development and postpone final determination of
requirements to assure that requirements being set are achievable.
In practice, however, our work has shown that there are still
significant gaps between requirements and technology resources when
programs begin. Our most recent annual assessment of major weapon
systems programs,[Footnote 8] for example, showed that only 15 percent
of the programs we reviewed began development having demonstrated that
all of their technologies were mature. More often than not, programs
had to worry about maturing technologies well into system development,
when they should have focused on maturing system design and preparing
for production. These assessments also show that programs that started
development with mature technologies experienced lower development and
unit cost increases than those programs that started with immature
technologies. Table 3 provides some examples.
Table 3: Technology Maturity and Program Outcomes:
Program: Advanced Threat Infrared Countermeasures/Common Missile
Warning System;
Percent increase in R&D (first full estimate to latest estimate): 5.6%;
Percent of critical technologies and associated maturity level at
development start: 50 % (3 of 6) at 6 or higher.
Program: C-5 Reliability Enhancement and Reengining Program;
Percent increase in R&D (first full estimate to latest estimate): 2.1%;
Percent of critical technologies and associated maturity level at
development start: 100 % (11 of 11) at 6 or higher.
Program: DD(X) Destroyer;
Percent increase in R&D (first full estimate to latest estimate):
417.3%;
Percent of critical technologies and associated maturity level at
development start: 25 % (3 of 12) at 6 or higher.
Program: Future Combat System;
Percent increase in R&D (first full estimate to latest estimate): 50.8%;
Percent of critical technologies and associated maturity level at
development start: 32 % (17of 52) at 6 or higher.
Program: Joint Strike Fighter;
Percent increase in R&D (first full estimate to latest estimate): 30.1%;
Percent of critical technologies and associated maturity level at
development start: 25 % (2 of 8) are 6 or higher.
Source: GAO.
Note: Technology readiness level of 7 or higher at program launch is
considered best practice; a technology readiness level of 6 or higher
is DOD standard.
[End of table]
Although the majority of respondents to our survey believed that the
initial baselines of their programs were reasonable, a significant
group, about 24 percent of program managers, responded that their
program parameters were not reasonable at the start and 45 program
managers responded that their program had been re-baselined one or more
times for cost and schedule increases; 18 percent said one or more key
technologies fell below a readiness level of 7, which is proven to work
in an operational environment. They also noted that the most frequently
missing critical skill was systems engineering--a key function for
matching requirements to the technologies needed and for providing
reasonable baselines at the beginning of development. In addition, in
written comments and individual interviews, program managers noted
pressure to agree to cost commitments that could be attained only if
programs enjoyed higher-level support. They also noted that
requirements were often not fully defined at the onset of a program,
and many also pointed out that users and stakeholders often did not
stick to the agreements they made when programs were launched,
especially if technologies did not mature as planned.
Figure 7: To What Extent Were the Parameters of Your Program Reasonable
at Program Start?
[See PDF for image]
[End of figure]
Figure 8: How Program Managers Responded to an Open-ended Question on
What Were the Biggest Obstacles They Faced:
[See PDF for image]
[End of figure]
Program managers' views were mixed when it came to whether human
capital resources were well matched to new programs. They cited major
improvements in DOD's training programs and credited cross-functional
teams as a valuable resource. They also generally believed they
personally had the right mix of experience and training to do their
jobs well. Ninety-four percent of the program managers responding to
our survey reported that they had been certified at the highest level
for program management by DOD's Defense Acquisition University. More
than 80 percent also believed they had adequate training in the areas
of systems engineering, business processes, contracting, management,
program representation, cost control, and planning and budgeting.
Slightly less, about 76 percent, believed they had enough leadership
training. In addition, about 92 percent said that they believed that
their service consistently assigned people with the skills and
experience to be effective program managers.
At the same time, however, program managers comments and interviews
with program executive officers pointed to critical skill shortages for
staff that support them--including program management, systems
engineering, cost estimating, and software development. Some of these
officials attributed the shortages to shifts in emphasis from oversight
to insight of contractor operations. Lastly, in their written comments,
about 18 percent of program managers who provided written comments
cited shortcomings in their career path, such as lack of opportunities
at the general officer level and requirements to move often as a
disincentive; 13 percent cited the lack of financial incentives. Some
program managers also noted that DOD loses opportunities to retain
valuable experience, merely because there are no formal incentives for
military officers to stay on as program managers after they are
eligible for retirement. Civilians in program management also cited a
lack of career opportunities; one problem cited was having to find
their next job in contrast to military program managers, whose
subsequent job is presented to them.
Execution in DOD Does Not Provide Adequate Support and Accountability:
According to program managers and our past reviews, the execution
process does not provide adequate support and accountability. In
particular, knowledge-based development processes are not employed,
program managers are not empowered to execute, and they are not held
accountable for delivering programs within targets.
More specifically, DOD has encouraged following knowledge-based
development processes in its acquisition policy but not always in
practice. The acquisition process itself mirrors many aspects of the
commercial companies. For example, it requires a variety of senior,
functional, and program-level personnel to come together, assess
progress, identify problems, and make go/no-go decisions at key points
in development. It encourages oversight personnel to base these
decisions on quantifiable data and demonstrated knowledge. To enhance
product development, DOD has also been attempting to adopt and improve
policies in areas such as software development and systems engineering.
However, program managers who responded to our survey believe that the
acquisition process does not enable them to succeed because it does not
empower them to make decisions on whether the program is ready to
proceed forward or even to make relatively small trade-offs between
resources and requirements as unexpected problems are encountered.
Program managers assert that they are also not able to make shifts
within personnel to respond to changes affecting the program. At the
same time, program managers commented that requirements continue to be
added as the program progresses and funding instability continues
throughout. These two factors alone cause the greatest disruption to
programs, according to program managers. Compounding this problem is
the fact that because acquisition programs tend to attract funds over
other activities, including science and technology efforts that
ultimately support acquisition, program managers are incentivized to
take on tasks that really should be accomplished within a laboratory
environment, where it is easier and cheaper to discover and address
technical problems.
With many factors out of their span of control, program managers in our
focus groups also commented that it was difficult to hold them
accountable for mistakes. In addition, in their written comments to the
survey, many program managers expressed frustration with the time
required of them to answer continual demands for information from
oversight officials--many of which did not seem to add value. Some
program managers in fact estimated that they spent more then 50 percent
of their time producing and tailoring and explaining status information
to others.
More broadly, in interviews and written comments, many program managers
and program executive officials said that did not believe that DOD's
acquisition process really supported or enabled them. Instead, they
viewed the process as cumbersome and the information produced as non-
strategic. When strategic plans or useful analyses were produced, they
were done so apart from the acquisition process.
Our own reviews have pointed to a number of structural problems with
the acquisition process.[Footnote 9] In particular, while DOD's
acquisition policy has embraced best practice criteria for making
decisions, it does not yet have the necessary controls to ensure
knowledge is used for decision-making purposes. As a result, programs
can move forward into design, integration, and production phases
without demonstrating that they are ready to. Without a means to ensure
programs and senior managers are adhering to the process, the process
itself can become an empty exercise--and, in the view of program
managers, a time-consuming one.
Table 4 highlights differences between DOD and commercial knowledge-
based development support and accountability factors--collectively from
the perspective of program managers, our past reports, and observations
we made during the course of the review.
Table 4: Are Best Practices Present in DOD?
Best practices: Base decisions on quantifiable data and demonstrated
knowledge;
DOD: DOD policy encourages decisions to be based on quantifiable data
and demonstrated knowledge, but not happening in practice.
Best practices: Empower program managers to make decisions;
DOD: Program managers say they are not empowered in the same way as
commercial companies. They do not control resources. They do not have
authority to move programs to next phases.
Best practices: Hold program managers accountable;
DOD: Difficult to enforce accountability.
Best practices: Program managers stay through execution;
DOD: Tenure has been lengthened, but program managers generally do not
stay after 3 to 4 years.
Source: GAO.
[End of table]
Data Supporting Oversight and Management Decisions:
We reported that while DOD's acquisition policy has embraced best
practice criteria for making decisions, it does not yet have the
necessary controls to ensure demonstrable data is used for decision-
making purposes. We recommended that DOD assure that program launch
decisions capture knowledge about cost and schedule estimates based on
analysis from a preliminary design using systems engineering tools. In
transitioning from system integration to system demonstration, we
recommended that DOD ensure the capture of knowledge about the
completion of engineering drawings; completion of subsystem and system
design reviews; agreement from all stakeholders that the drawings are
complete; and identification of critical manufacturing processes, among
other indicators. And in transition to production, we recommended that
DOD capture knowledge about production and availability of
representative prototypes along with statistical process control data.
We recommended adopting these controls because, in our view, they would
help set program managers up for success by (1) empowering them with
demonstrated knowledge as they move toward production and (2) bringing
accountability to their positions and making the business case more
understandable. Without these types of controls, the process can become
an empty and time-consuming exercise in the view of program managers.
At present, our reports continue to show that programs are allowed to
proceed without really showing that they are ready to. In our most
recent annual assessment of major weapon systems, for example, only 42
percent of programs had achieved design stability at design review and
almost none of the programs in production or nearing production planned
to assure production reliability through statistical control of key
processes.
Our survey also indicated that a relatively small percentage of
programs used knowledge indicators that successful commercial companies
use. For example, in responding to our survey, only 32 percent of
program managers said they used design drawing completion extensively
to measure design maturity; only 26 percent said they used production
process controls to a great extent. Even fewer program managers
reported that their immediate supervisor used these measures
extensively to evaluate progress.
In our survey and interviews, program managers and program executive
officers also frequently commented that they spend too much time
preparing data for oversight purposes that is not strategic or very
useful to them. In fact, more than 90 percent of survey respondents
said that they spent either a moderate, great, or very great extent of
their time representing their program to outsiders and developing and
generating information about program progress. In addition, program
managers told us that they do not have standard tools for preparing
program-status data. Instead, they must hand-tailor data to the
requester's particular demands for format and level of detail. The Air
Force was cited by some program managers as taking initiative in
developing a database (known as the System Management Analysis
Reporting Tool) that could save time in answering internal oversight
demands for data, but they also wanted to be able to use such a tool to
answer outside demands. While program managers said they were spending
a great deal of time reporting on program status to outsiders, some
program executive officers and program managers also commented that
they had to separately produce data, analyses, and strategic plans for
their own purposes in order to keep their programs on track--the types
of plans and analyses that they used were simply not called for by the
process itself. One program executive officer said that he used three
documents, the approved program baseline, the acquisition strategy, and
the test plan to evaluate the program manager's plans--all of these
documents and many more are required under current acquisition
planning--but these were of most significance. In addition the
executive officer held a one-day review per quarter with each program
manager and reviewed metrics such as earned value, use of award fee,
contract growth, and schedule variation.
Program Manager Authority:
In several key areas, program managers said that they do not have the
necessary authority to overcome obstacles and make trade-offs to
achieve program goals. About 60 percent of the program managers that
responded to our survey said that program managers should have more
authority to manage their programs--particularly when it comes to
funding, deciding when programs are ready to proceed to the next phase,
and shifting staff. In interviews and written comments, some program
managers commented that they were seeking the ability to make
relatively small trade-offs--for example, moving a staff member from
one section of a program to another and shifting a small amount of
funds from procurement accounts to research and development accounts,
while others advocated for greater authority, as long as their program
stayed on track. In addition, program managers often commented that
they should have a larger role in requirements decisions that are made
after a program is started--specifically, the ability to veto new
requirements that would put too much strain on the program. A few
program managers we interviewed, however, believed that they did have
sufficient authority and that many program managers have not learned
how to exercise it or are risk averse. Others commented that program
managers were simply not allowed by senior managers to exercise their
authority. At the same time, program executive officers, who manage a
set of programs, commented in interviews that they also lacked
authority over simple matters such as moving staff or shifting small
amounts of funds. Lastly, in our focus groups and in written comments,
program managers who specifically worked for the Missile Defense Agency
indicated that they did have authority to make trade-offs among cost,
schedule, and performance and to set requirements for the business
case. They found that this authority alone greatly separated their
current positions from past program manager positions and consistently
cited it as a major enabler.
Table 5 shows how program managers answered survey questions regarding
the types of formal and information authority they have. Figure 9
highlights comments that were provided by program managers.
Table 5: Program Manager Views on Formal vs. Informal Authority:
(Continued From Previous Page)
In percent: Type of authority: Developing program requirements;
I have formal authority[A]: 10%;
I have informal authority: 82%;
No authority: 7%.
In percent: Type of authority: Changes in program requirements;
I have formal authority[A]: 13%;
I have informal authority: 85%;
No authority: 2%.
In percent: Type of authority: Flexibility within program to reallocate
funding;
I have formal authority[A]: 81%;
I have informal authority: 15%;
No authority: 5%.
In percent: Type of authority: Developing technology;
I have formal authority[A]: 42%;
I have informal authority: 45%;
No authority: 9%.
In percent: Type of authority: Setting testing requirements;
I have formal authority[A]: 48%;
I have informal authority: 49%;
No authority: 2%.
In percent: Type of authority: Selecting contractor sources;
I have formal authority[A]: 48%;
I have informal authority: 33%;
No authority: 11%.
In percent: Type of authority: Administering contracts;
I have formal authority[A]: 60%;
I have informal authority: 37%;
No authority: 3%.
In percent: Type of authority: Addressing difficulties to meet
requirements;
I have formal authority[A]: 66%;
I have informal authority: 31%;
No authority: 2%
Source: GAO.
[A] Note: Numbers may not total 100 percent due to rounding.
[End of table]
Figure 9: Highlights of Program Manager Comments on What Types of
Authority They Need:
[See PDF for image]
[End of figure]
Accountability:
Program manager views with regard to accountability are mixed. In our
interviews and our focus groups, many program managers stated they
personally held themselves accountable; however, many also commented
that it is difficult to be accountable when so much is outside their
span of control. During our focus groups, program managers cited
sporadic instances when program managers were removed from their
positions or forced to retire. They also cited instances when a program
manager was promoted, even though the program was experiencing
difficulties. In their written comments for our survey, program
managers often commented that it was a disincentive that senior leaders
who were impacting their program negatively were not being held
accountable.
We observed some key differences between the commercial companies we
visited and DOD when it comes to accountability.
* Commercial companies make it very clear who is accountable on a
program and for what. Goals that must be achieved are clearly spelled
out and understood by the entire program team. In DOD, it is not always
clear who is responsible. Moreover, the expectations set for program
managers by their supervisors may not necessarily match up with the
goals of their program--particularly when the program manager is a
military officer who reports to both a PEO and another commanding
official.
* Program managers and senior managers in the commercial sector are
required to stay with programs until they are done; at DOD they are
not.
* Program managers in the commercial sector are incentivized to stay
with programs and be accountable for them--principally through
empowerment and financial incentives, but also through their desire to
help the company achieve its goals. At DOD, program managers strongly
asserted that they are incentivized to help the warfighter, but few
said they were incentivized by financial or promotion incentives or by
empowerment.
Senior Leader Support during Execution:
In commenting on senior leader support during program execution,
program managers had mixed views on whether they received sustained
commitment from their program executive officers, but widely believed
that they did not receive sustained commitment from other senior
leaders and stakeholders--unless their programs enjoyed priority and
support from very high level officials, Congress, or the President.
More often than not, programs struggled to compete for funding and were
continually beset by changing demands from users. Others noted that
while DOD is emphasizing jointness in programs more and more,
collaboration among senior leaders needed to achieve jointness is not
always happening. Some program managers lamented that they felt they
were not respected in DOD, while others believed their service was
taking some positive actions to put program managers on a par with
military officers in operational positions.
Program managers were also troubled by constant demands for information
for oversight purposes as well as interruptions from stakeholders (for
example, in department-wide budget or testing offices) that seemed to
be non value-added. As we noted earlier, over 90 percent of the survey
respondents said that they spent either a moderate, great, or very
great extent of their time representing their program to outsiders and
developing and generating information about program progress.
Several program managers also cited reluctance on the part of senior
managers to hear bad news. Our past reviews have similarly noted that
the overall competition for funding in DOD spurs program managers to
suppress bad news because it can result in funding cuts.
[End of section]
Chapter 4: Basic Incentives Drive Differences in How Program Managers
Are Supported and Held Accountable:
Differences between DOD and leading companies in how program managers
are supported and held accountable are rooted in differences in
incentives and resulting behaviors. This begins with the definition of
success. The commercial firms we studied concluded their survival
hinged on their ability to increase their market share, which, in turn,
meant developing higher-quality products, at the lowest possible price,
and delivering them in a timely fashion--preferably before their
competitors could do the same. This imperative meant that they had no
choice but to narrow the gap between requirements and resources in a
manner that not only ensured they met their market targets, but did so
in a manner that consumed resources efficiently. It also meant that
they had no choice but to fully support the development effort, to
instill strategic planning and prioritization, to work collaboratively,
to follow a knowledge-based process that makes product development
manageable, and ultimately, assign accountability to all involved for
success or failure. In theory, DOD's success likewise hinges on its
ability to deliver high quality weapons to the warfighter in a timely
fashion. But in practice, the implied definition of success is the
ability of a program to win support and attract funds. Of course, there
are reasons for this disconnect. Corporate revenue is generated by
customer sales while DOD's funding is dependent on annual
appropriations. Corporations go out of business when their product
development efforts do not succeed; DOD does not. Selling products to
customers is the single focus of a private-sector company while DOD is
charged with a myriad of important missions--each of which also
competes for budget share. As a result, program managers are
incentivized to overpromise on performance because it makes their
program stand out from others. They are incentivized to underestimate
cost and schedule and to suppress bad news because doing otherwise
could result in a loss of support and funding and further damage their
program. In short, unknowns become acceptable and desirable rather than
unacceptable as they are in the corporate environment. And
accountability becomes much more difficult to define.
Figure 10: Key Differences in Definition of Success and Resulting
Behaviors:
[See PDF for image]
[End of figure]
Definition of Success:
Success for the commercial world is straightforward and simple:
maximize profit. In turn, this means selling products to customers at
the right price, right time, and right cost. Each of the commercial
companies we visited enjoyed success to this end, but at some point in
time, as competitors made gains, markets tightened, and the pace of
technology changes grew faster, they realized they needed to do more to
be successful. Toyota decided it needed to expand its role in the world
market place and this need persisted as competition grew stronger over
the years. For Siemens this realization came in the 1990s--when Siemens
Medical Division took a hard look at its profitability for its medical
devices and for Motorola in the 1980's when it began losing market
share for its communication devices. To turn themselves around, all
three companies chose not to depend on technology breakthroughs or
exotic marketing, but rather to improve their position by looking
inward at how they approached development. Each found that there was
room for improvement, starting with corporate cultures and ending with
processes and controls. In Toyota's case, emphasis was largely placed
on collaboration and consistency. In Siemens case, emphasis was placed
on quality, particularly because its medical products come under
extensive Food and Drug Administration regulations. For Motorola,
emphasis was placed on empowerment and commonality, particularly in the
processes that support product development like software development.
At DOD, success is often formally defined in similar terms as the
commercial world: deliver high quality products to customers (the
warfighter) at the right time and the right cost. Virtually all program
managers we spoke with first defined success in terms of enabling
warfighters and doing so in a timely and cost-efficient manner. But
when the point was pursued further, it became clear that the implied
definition for success in DOD is attracting funds for new programs, and
keeping funds for ongoing programs. Program managers themselves say
they spend enormous amounts of time retaining support for their efforts
and that their focus is largely on keeping funds stable. They also
observe that DOD starts more programs than it can afford to begin with,
which merely sets the stage for competition and resulting behaviors. As
noted earlier, there are factors that contribute to how success is
defined in practice, including the fact that DOD depends on annual
appropriations and it must fund a wide variety of missions beyond
weapon systems development. However, according to program managers, the
willingness to make trade-off decisions alone, would help DOD mitigate
these circumstances.
Means for Success:
Regardless of where they placed greatest emphasis, each company we
studied adopted processes and support mechanisms that emphasized the
following:
* risk reduction,
* knowledge-based decisionmaking,
* discipline,
* collaboration,
* trust,
* commitment,
* consistency,
* realism, and:
* accountability.
Such characteristics were seen as absolutely essential to gaining
strength in the market place. With limited opportunities to invest in
new product development efforts, companies understand it is essential,
for example, that they know they are pursuing efforts that will
optimize profits. Therefore, estimates of costs and technology maturity
must be accurate and they must be used for making decisions.
Consistency and discipline are integral to assuring that successful
efforts can be repeated. Ultimately, these characteristics translate
into processes that help companies develop products quicker, cheaper,
and better. At the strategic level, processes include accurate,
strategic planning and prioritization to ensure the right mix of
products are pursued; investment strategies that prioritize projects
for funding; and strong systems engineering to help them establish a
realistic business case that levels market needs with available
resources prior to beginning a product development. At the tactical
level, this includes knowledge-based developments that center on
designing and manufacturing products that will sell well enough to make
an acceptable profit. This combination of focused leadership and
disciplined processes promotes positive behaviors, such as an
insistence that technology development take place separately from
product development programs and trade-offs between requirements and
resources be made before beginning a program; it promotes an atmosphere
of early candor and openness from everyone as to potential program
risks; and underscores the need for realistic, knowledge-based cost and
schedule estimates to support full funding decisions; and the ability
to test early, allowing "red lights" for problems that must be proven
solved before they can be changed to "green lights."
Once attracting and sustaining funds becomes a part of the definition
of success, as it is at DOD, different values and behaviors emerge. For
example, it is not necessarily in a program manager's interest to
develop accurate estimates of cost, schedule, and technology readiness,
because honest assessments could result in lost funding. Delayed
testing becomes preferred over early testing since that will keep "bad
news" at bay.
Ultimately, no matter how well-intentioned or what improvements are
made, DOD's processes and support mechanisms eventually play into
funding competition. On paper, the requirements process may emphasize
realism and the importance of incremental development, but in practice,
it consistently encourages programs to promise performance features
that significantly distinguish them from other systems. Likewise,
changes may be made to make the funding process more strategic, but
because there are still many programs competing for funds, it
encourages cost and schedule estimates to be comparatively soft with
little benefit from systems engineering tradeoffs. By favoring
acquisition programs over science and technology efforts, the funding
process also encourages programs to take on technology development that
should be carried out in research labs. Lastly, the acquisition process
may adopt world-class criteria for making decisions, but because it is
much easier to attract funds for a formal weapons program than funds
for the exercise of gaining knowledge about technologies, the process
encourages programs to move forward despite risks with the assumption
that programs can resolve technical, design, or production "glitches"
later on. Significant unknowns are accepted in this environment.
Delivering a product late and over cost does not necessarily threaten
program success. The cumulative effect of these pressures is
unpredictable cost and schedule estimates at the outset of a program
that are breached, sometimes very significantly, by the time the weapon
system is fielded.
Other Differences Put Additional Pressures on DOD Program Managers:
There are other environmental differences that put additional pressures
on program managers within DOD. They include layers of internal and
external oversight that come with DOD's stewardship responsibilities,
personnel rules that make it more difficult to manage human capital and
hold people accountable, laws and regulations that place additional
constraints on an acquisition, and the mere size and scope of DOD,
which adds significant challenges to communicating and collaborating
effectively.
For example, as shown below, commercial companies we visited tended to
have fairly streamlined oversight. No matter what level the program
manager resided, they had access to top executives who were empowered
to help them make go/no-go decisions. In addition to this structure,
the companies were governed by a degree of oversight from shareholders,
but this was not actualized in day-to-day management of program
development activities. At DOD, by contrast, program managers operate
under many layers of oversight--both internally and externally. These
include Congress, the President, the Secretary of Defense, a myriad of
functional agencies as well as the military services--all of whom have
a say in DOD's overall budget as well as funding for specific programs.
Moreover, within these confines, leaders at all levels shift
frequently. Much of this oversight is necessary for carrying out
stewardship responsibilities for public money, but studies conducted by
a variety of commissions assessing acquisition problems through the
years have consistently found that there are opportunities to reduce
oversight layers and streamline oversight processes and protect
programs from frequent leadership changes. Program managers themselves
understood the need for oversight, but found that responding to
oversight demands was taking too much of their time. They also
identified opportunities to make it easier to work within this
environment, including development of databases to support internal and
external oversight requests, empowering program managers for more day-
to-day decisions, and making stakeholders more accountable.
Figure 11: Commercial vs. DOD Oversight Environments:
[See PDF for image]
[End of figure]
Program managers also cited several trends that have increased
pressures that they face. These include DOD's movement toward
developing more technical complex families of weapon systems as one
package (system of systems), which they believe vastly increases
management challenges and makes it more difficult to oversee
contractors and DOD's reduction in acquisition workforces over the past
decade, which has made it more difficult to carry out day-to-day
responsibilities and retain technical and business expertise. Overall,
however, program managers themselves consistently attribute their
problems to competition for funding over these other factors.
[End of section]
Chapter 5: Conclusions and Recommendations:
Like the commercial world, DOD has a mandate to deliver high-quality
products to its customers, at the right time, and the right price.
Quality and time are especially critical to maintain DOD's superiority
over others, to counter quickly changing threats, and to better protect
and enable the warfighter. Cost is critical given DOD's stewardship
over taxpayer money, long-term budget forecasts which indicate that the
nation will not be able to sustain its current level of investment in
weapon systems, and DOD's desire to dramatically transform the way it
conducts military operations. At this time, however, DOD is not
positioned to deliver high quality products in a timely and cost-
efficient fashion. It is not unusual to see cost increases that add up
to tens of millions of dollars, schedule delays that add up to years,
and large and expensive programs to be continually rebaselined.
Recognizing this dilemma, DOD has tried to embrace best practices in
its policies, instill more discipline in requirements setting,
strengthen training for program managers and has lengthened program
manager tenures. It has also reorganized offices that support and
oversee programs, required programs to use independent cost estimates
and systems engineering, and it has alternately relaxed and
strengthened oversight over contractors in an effort to extract better
performance from them. Yet despite these and many other actions,
programs are still running over cost and over schedule and, in some
cases, changes have merely added tasks for program managers while
adding no value.
Our work shows that this condition will likely persist until DOD
provides a better foundation on which program managers can launch
programs and more consistent and steadfast support once it commits to
programs. At the core of this solution is developing and enforcing an
investment strategy that prioritizes programs based on customer needs
and DOD's long term vision and reduces the burden of advocacy on the
part of the program manager. DOD will always be facing funding
uncertainties due to the environment it operates in. But it has an
opportunity to greatly mitigate the risks that come with this
environment by separating long-term wants from needs, matching them up
against what technologies are available today, tomorrow, and decades
from now, as well as being realistic in determining what resources can
be counted on. Without an investment strategy, all other improvements
will likely succumb to the negative incentives and behaviors that come
with continual competition for funding. With an investment strategy,
senior leaders will be better positioned to formally commit to a
business case that assures new programs fit in with priorities, that
they begin with adequate knowledge about technology, time, and cost,
and that they will follow a knowledge-based approach as they move into
design and production. Another core enabler for improving program
managers' chances for success lies in leadership's ability to implement
evolutionary, incremental acquisition programs that have limited cycle
times from beginning to delivery of the weapon system. This allows DOD
to align program managers' tenures to delivery dates, thereby
substantially increasing accountability for successful outcomes.
Once senior leaders do their part--by providing program managers with
an executable business case and committing their full support to a
program--they can begin to empower program managers more and hold them
accountable. By embracing a model for supporting program managers that
incorporates all these elements, DOD can achieve the same outcomes for
its weapons programs as other high-performing organizations.
Recommendations for Executive Action:
We recommend that the Secretary of Defense take the following actions
to ensure program managers are well positioned to successfully execute
and be held accountable for weapon acquisitions:
* DOD should develop an investment strategy that, at a minimum,
* determines the priority order of needed capabilities with a corollary
assessment of the resources, that is dollars, technologies, time and
people needed to achieve these capabilities. The remaining capabilities
should be set out separately as desirable, resources permitting.
* lays out incremental product development programs for achieving
desired capabilities, and:
* establishes controls to ensure that requirements, funding, and
acquisition processes will work together so that DOD will sustain its
commitment to its priority programs.
As DOD works to develop the strategy, it should take an interim step by
identifying priorities for programs that are already past milestone B
(the formal start of development). Once the strategy is complete, it
should be used by the Office of the Secretary of Defense to prepare and
assess annual budget proposals as well as to balance funding between
science and technology efforts and acquisition efforts to ensure that
robust technology development efforts are conducted, but outside the
acquisition program until reaching maturity.
* For each new major weapons program, require that senior-level
officials from the requirements, science and technology, program
management, testing communities as well as the Office of the
Comptroller formally commit to a business case prior to approving a
program at milestone B. At a minimum, the business case should
demonstrate that:
* a requirement exists that warrants a materiel solution consistent
with national military strategy,
* an independent analysis of alternatives has been conducted:
* the developer has the requisite technical knowledge to meet the
requirement,
* the developer has a knowledge-based product development and
production plan that will attain high levels of design and production
maturity,
* reasonable estimates have been developed to execute the product
development and production plan, and:
* funding is available to execute the plan.
* Develop and implement a process to instill and sustain accountability
for successful program outcomes. At a minimum, this should consider:
* matching program manager tenure with delivery of a product or for
system design and demonstration,
* tailoring career paths and performance management systems to
incentivize longer tenures,
* empowering program managers to execute their programs, including an
examination of whether and how much additional authority can be
provided over funding, staffing, and approving requirements proposed
after milestone B,
* developing and providing automated tools to enhance management and
oversight as well as to reduce time required to prepare status
information.
Agency Comments and Our Evaluation:
In commenting on a draft of our report, DOD's Acting Director for
Procurement and Acquisition Policy concurred with our recommendations.
In doing so, DOD asserted it was already taking actions to address our
recommendations, notably by reviewing its overall approach to
acquisition governance and investment decisionmaking as part of its
Quadrennial Defense Review due in February 2006 and identifying ways to
more effectively implement existing policies. DOD also stated it
intended to develop a plan to address challenges in acquisition
manpower including program manager tenure and career path and it
intends to enhance its information management systems to facilitate
oversight and management decisions. As underscored in our report, DOD
has attempted similar efforts in the past--that is, reviewed its
approach to governance and investment decisions and policies--without
achieving significant improvements. This is because DOD has not assured
such actions were executed in tandem with (1) instilling more
leadership and discipline in investment decisionmaking, in both the
short and long term and (2) instilling accountability--by requiring key
senior officials to sign a business case, based on systems engineering
knowledge, prior to every new acquisition as well as by matching
program managers' tenure to cycle time. Therefore, in pursuing the
actions it identifies in its response to our report, we believe that
DOD should address the important questions of who should be held
accountable for acquisition decisions; how much deviation from the
original business case is allowed before it is no longer considered
valid and the investment reconsidered; and what is the penalty when
investments do not result in meeting promised warfighter needs.
The full text of the department's response is in appendix I.
[End of section]
Appendix I: Comments from the Department of Defense:
Acquisition Technology And Logistics:
Office Of The Under Secretary Of Defense:
3000 Defense Pentagon Washington, DC 20301-3000:
Nov. 22 2005:
DPAP/PAIC:
Mr. Michael Sullivan:
Director, Acquisition and Sourcing Management:
U.S. Government Accountability Office:
441 G Street, N.W.:
Washington, D.C. 20548:
Dear Mr. Sullivan:
This is the Department of Defense response to the GAO draft report,
`BEST PRACTICES: Better Support of Weapon System Program Managers
Needed to Improve Outcomes,' dated October 21, 2005, (GAO Code 120320/
GAO-06-110). I concur with the recommendations in the report and have
provided some amplifying discussion.
Domenic Cipicchio Acting Director, Defense Procurement and Acquisition
Policy:
Attachment: As stated:
GAO Draft Report - Dated October 21, 2005 GAO Code 120320/GAO-06-110:
"Best Practices: Better Support Of Weapon System Program Managers
Needed To Improve Outcomes"
Department Of Defense Comments To The Recommendations:
Recommendation 1: The GAO recommended that the Secretary of Defense
direct the DoD to develop an investment strategy that at a minimum,
* determines the priority order of needed capabilities with a corollary
assessment of the resources, that is dollars, technologies, time and
people needed to achieve these capabilities. The remaining capabilities
should be set out separately as desirable, resources permitting.
* lays out incremental product development programs for achieving
desired capabilities, and:
* establishes controls to ensure that requirements, funding, and
acquisition processes will work together so that DoD will sustain its
commitment to its priority programs.
Also, as DoD works to develop the strategy, it should take an interim
step by identifying priorities for programs that are already past
milestone B. Once the strategy is complete, it should be used by the
Office of the Secretary of Defense to prepare and assess annual budget
proposals as well as to balance funding between science and technology
efforts and acquisition efforts to ensure that robust technology
development efforts are conducted, but outside the acquisition program
until reaching maturity. (p. 61/GAO Draft Report):
DOD Response: Concur:
DoD is currently reviewing its overall approach to department
acquisition governance with the objective of (1) refining the mechanism
for prioritizing materiel acquisition proposals and improving the
alignment between corporate commitment and associated resource
allocation; (2) effectively implementing existing. evolutionary
acquisition policy; (3) ensuring that current statutory and regulatory
controls operate effectively; and 4) creating a more complementary
relationship between technology development and acquisition. The
ongoing Quadrennial Defense Review (QDR) and related efforts will
address these issues. The QDR will be completed in February 2006.
Subsequently, DoD will develop an implementation plan and schedule for
actions that will address these issues.
Recommendation 2: The GAO recommended that the Secretary of Defense,
for each new weapons program, require that senior-level officials from
the requirements, science and technology, program management, testing
communities as well as the Office of the Comptroller formally commit to
a business case prior to approving a program at milestone B. At a
minimum, the business case should demonstrate that:
* a requirement exists that warrants a materiel solution consistent
with national military strategy,
* an independent analysis of alternatives has been conducted:
* the developer has the requisite technical knowledge to meet the
requirement,
* the developer has a knowledge-based product development and
production plan that will attain high levels of design and protection
maturity,
* reasonable estimates have been developed to execute the product
development and production plan, and:
* funding is available to execute the plan. (pages 61 & 62/GAO Draft
Report):
DOD Response: Concur:
Current DoD policy requires a number of criteria to be met before a
program may be formally initiated. These include: (1) a Joint
Requirements Oversight Council validated requirement that is consistent
with national military strategy; (2) an Analysis of Alternatives that
has been completed and assessed; (3) evidence of technology maturity
that has been independently assessed; (4) an approved acquisition
strategy; (5) an approved acquisition program baseline; (6) a
completed, independent cost estimate; and (7) full-funding. Mechanisms
to improve the effectiveness of these policies are being considered as
part of the QDR.
Recommendation 3: The GAO recommended that the Secretary of Defense
direct the DoD to develop and implement a process to instill and
sustain accountability for successful program outcomes. At a minimum,
this should consider:
* matching program manager tenure with delivery of a product or for
system design and demonstration,
* tailoring career paths and performance management systems to
incentivize longer tenures,
* empowering program managers to execute their programs, including an
examination of whether and how much additional authority can be
provided over funding, staffing, and approving requirements proposed
after milestone B, * developing and providing automated tools to
enhance management and oversight as well as to reduce time required to
prepare status information, (p. 62/GAO Draft Report):
DOD Response: Concur:
DoD is currently engaged in the development of a manpower strategy
designed to satisfy our current and future acquisition manpower
challenges. That strategy will be comprehensive, and consider such
issues as program manager tenure and career paths. This strategy will
be developed by the end of FY 2006. In addition the department is
considering policy designed to reduce requirements growth after program
initiation and, is in the process of developing a transparent and
efficient information management system intended to provide management
with accurate and current program information.
[End of section]
Appendix II: GAO Staff Acknowledgments:
GAO Contacts:
Michael J. Sullivan (202) 512-4841 or sullivanm@gao.gov:
Staff Acknowledgments:
Greg Campbell, Cristina Chaplain, Ron La Due Lake, Sigrid McGinty, Jean
McEwen, Carol Mebane, Guisseli Reyes, Lesley Rinner, Lisa Simon,
Bradley Trainor, and Michele Williamson.
[End of section]
Related GAO Products:
DOD Acquisition Outcomes: A Case for Change. GAO-06-257T. Washington,
D.C.: November 15, 2005.
Defense Acquisitions: Stronger Management Practices Are Needed to
Improve DOD's Software-Intensive Weapon Acquisitions. GAO-04-393.
Washington, D.C.: March 1, 2004.
Best Practices: Setting Requirements Differently Could Reduce Weapon
Systems' Total Ownership Costs. GAO-03-57. Washington, D.C.: February
11, 2003:
Best Practices: Capturing Design and Manufacturing Knowledge Early
Improves Acquisition Outcomes. GAO-02-701. Washington, D.C.: July 15,
2002.
Defense Acquisitions: DOD Faces Challenges in Implementing Best
Practices. GAO-02-469T. Washington, D.C.: February 27, 2002.
Best Practices: Better Matching of Needs and Resources Will Lead to
Better Weapon System Outcomes. GAO-01-288. Washington, D.C.: March 8,
2001.
Best Practices: A More Constructive Test Approach Is Key to Better
Weapon System Outcomes. GAO/NSIAD-00-199. Washington, D.C.: July 31,
2000.
Defense Acquisition: Employing Best Practices Can Shape Better Weapon
System Decisions. GAO/T-NSIAD-00-137. Washington, D.C.: April 26, 2000.
Best Practices: DOD Training Can Do More to Help Weapon System Programs
Implement Best Practices. GAO/NSIAD-99-206. Washington, D.C.: August16,
1999.
Best Practices: Better Management of Technology Development Can Improve
Weapon System Outcomes. GAO/NSIAD-99-162. Washington, D.C.: July 30,
1999.
Defense Acquisitions: Best Commercial Practices Can Improve Program
Outcomes. GAO/T-NSIAD-99-116. Washington, D.C.: March 17, 1999.
Defense Acquisition: Improved Program Outcomes Are Possible. GAO/T-
NSIAD-98-123. Washington, D.C.: March 17, 1998.
Best Practices: DOD Can Help Suppliers Contribute More to Weapon System
Programs. GAO/NSIAD-98-87. Washington, D.C.: March 17, 1998.
Best Practices: Successful Application to Weapon Acquisition Requires
Changes in DOD's Environment. GAO/NSIAD-98-56. Washington, D.C.:
February 24, 1998.
Best Practices: Commercial Quality Assurance Practices Offer
Improvements for DOD. GAO/NSIAD-96-162. Washington, D.C.: August 26,
1996.
FOOTNOTES
[1] Defense Management: Key Elements Needed to Successfully Transform
DOD Business Operations, GAO-05-629T (Washington, D.C.: April 28, 2005)
and High-Risk Series: An Update, GAO-05-207 (Washington, D.C.: January
2005).
[2] A complete list of best practices reports is at the end of this
report.
[3] These figures represent the costs for the top five weapon systems
in 2001 and the top five in 2005. For 2001, these systems were F/A-22
Raptor, DDG-51 Guided Missile Destroyer, Virginia Class Submarine, C-17
Globemaster III, and the F/A 18 E/F, Naval Strike Fighter. The 2005
systems include the Joint Strike Fighter, Future Combat System, F/A-22
Raptor, DDG-51 Guided Missile Destroyer, and the Virginia Class
Submarine.
[4] 10 U.S.C.§ 1701 et seq. (P.L. 101-510. Div A. Title XII (November
5, 1990)).
[5] DOD Directive 5000.1, the Defense Acquisition System (May 2003) and
DOD Instruction 5000.2 Operation of the Defense Acquisition System (May
2003). The directive establishes evolutionary acquisition strategies as
the preferred approach to satisfying DOD's operational needs. The
directive also requires program managers to provide knowledge about key
aspects of a system at key points in the acquisition process. The
instruction implements the policy and establishes detailed policy for
evolutionary acquisition.
[6] Tactical Aircraft: Opportunity to Reduce Risks in the Joint Strike
Fighter Program with Different Acquisition Strategy, GAO-05-
271(Washington, D.C.: March 15, 2005).
[7] Space Acquisitions: Stronger Development Practices and Investment
Planning Needed to Address Continuing Problems, GAO-05-891T,
(Washington, D.C.: July 12, 2005).
[8] Defense Acquisitions: Assessments of Selected Major Weapon
Programs, GAO-05-301 (Washington, D.C.: March 31, 2005).
[9] Defense Acquisitions: DOD's Revised Policy Emphasizes Best
Practices but More Controls Are Needed, GAO-04-53 (Washington, D.C.:
Nov. 17, 2003).
GAO's Mission:
The Government Accountability Office, the investigative arm of
Congress, exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. Government Accountability Office
441 G Street NW, Room LM
Washington, D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm
E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director,
NelliganJ@gao.gov
(202) 512-4800
U.S. Government Accountability Office,
441 G Street NW, Room 7149
Washington, D.C. 20548: