DOD Business Systems Modernization

Navy Implementing a Number of Key Management Controls on Enterprise Resource Planning System, but Improvements Still Needed Gao ID: GAO-09-841 September 15, 2009

The Department of Defense (DOD) has long been challenged in effectively implementing key acquisition management controls on its thousands of business system investments. For this and other reasons, GAO has designated DOD's business systems modernization efforts as high-risk since 1995. One major business system investment is the Navy's Enterprise Resource Planning (ERP) system. Initiated in 2003, it is to standardize the Navy's business processes, such as acquisition and financial management. It is being delivered in increments, the first of which is to cost about $2.4 billion over its 20-year useful life and be fully deployed by fiscal year 2013. To date, the program has experienced about $570 million in cost overruns and a 2-year schedule delay. GAO was asked to determine whether (1) system testing is being effectively managed, (2) system changes are being effectively controlled, and (3) independent verification and validation (IV&V) activities are being effectively managed. To do this, GAO analyzed relevant program documentation, traced random samples of test defects and change requests, and interviewed cognizant officials.

The Navy has largely implemented effective controls on Navy ERP associated with system testing and change control. For example, it has established a well-defined structure for managing tests, including providing for a logical sequence of test events, adequately planning key test events, and documenting and reporting test results. In addition, it has documented, and is largely following, its change request review and approval process, which reflects key aspects of relevant guidance, such as having defined roles and responsibilities and a hierarchy of control boards. However, important aspects of test management and change control have not been fully implemented. Specifically, the program's tool for auditing defect management did not always record key data about changes made to the status of identified defects. To its credit, the program office recently took steps to address this, thereby reducing the risk of defect status errors or unauthorized changes. Also, while the program office's change review and approval procedures include important steps, such as considering the impact of a change, and program officials told GAO that cost and schedule impacts of a change are discussed at control board meetings, GAO's analysis of 60 randomly selected change requests showed no evidence that cost and schedule impacts were in fact considered. Without such key information, decision-making authorities lack an adequate basis for making informed investment decisions, which could result in cost overruns and schedule delays. The Navy has not effectively managed its IV&V activities, which are designed to obtain an unbiased position on whether product and process standards are being met. In particular, the Navy has not ensured that the IV&V contractor is independent of the products and processes that it is reviewing. Specifically, the same contractor responsible for performing IV&V of Navy ERP products (e.g., system releases) is also responsible for ensuring that system releases are delivered within cost and schedule constraints. Because performance of this system development and management role makes the contractor potentially unable to render impartial assistance to the government in performing the IV&V function, there is an inherent conflict of interest. In addition, the IV&V agent reports directly and solely to the program manager and not to program oversight officials. As GAO has previously reported, the IV&V agent should report the findings and associated risks to program oversight officials, as well as program management, in order to better ensure that the IV&V results are objective and that the officials responsible for making program investment decisions are fully informed. Furthermore, the contractor has largely not produced the range of IV&V deliverables that were contractually required between 2006 and 2008. To its credit, the program office recently began requiring the contractor to provide assessment reports, as required under the contract, as well as formal quarterly reports; the contractor delivered the results of the first planned assessment in March 2009. Notwithstanding the recent steps that the program office has taken, it nevertheless lacks an independent perspective on the program's products and management processes

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-09-841, DOD Business Systems Modernization: Navy Implementing a Number of Key Management Controls on Enterprise Resource Planning System, but Improvements Still Needed This is the accessible text file for GAO report number GAO-09-841 entitled 'DOD Business Systems Modernization: Navy Implementing a Number of Key Management Controls on Enterprise Resource Planning System, but Improvements Still Needed' which was released on September 15, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: United States Government Accountability Office: GAO: September 2009: DOD Business Systems Modernization: Navy Implementing a Number of Key Management Controls on Enterprise Resource Planning System, but Improvements Still Needed: GAO-09-841: GAO Highlights: Highlights of GAO-09-841, a report to congressional requesters. Why GAO Did This Study: The Department of Defense (DOD) has long been challenged in effectively implementing key acquisition management controls on its thousands of business system investments. For this and other reasons, GAO has designated DOD‘s business systems modernization efforts as high-risk since 1995. One major business system investment is the Navy‘s Enterprise Resource Planning (ERP) system. Initiated in 2003, it is to standardize the Navy‘s business processes, such as acquisition and financial management. It is being delivered in increments, the first of which is to cost about $2.4 billion over its 20-year useful life and be fully deployed by fiscal year 2013. To date, the program has experienced about $570 million in cost overruns and a 2-year schedule delay. GAO was asked to determine whether (1) system testing is being effectively managed, (2) system changes are being effectively controlled, and (3) independent verification and validation (IV&V) activities are being effectively managed. To do this, GAO analyzed relevant program documentation, traced random samples of test defects and change requests, and interviewed cognizant officials. What GAO Found: The Navy has largely implemented effective controls on Navy ERP associated with system testing and change control. For example, it has established a well-defined structure for managing tests, including providing for a logical sequence of test events, adequately planning key test events, and documenting and reporting test results. In addition, it has documented, and is largely following, its change request review and approval process, which reflects key aspects of relevant guidance, such as having defined roles and responsibilities and a hierarchy of control boards. However, important aspects of test management and change control have not been fully implemented. Specifically, the program‘s tool for auditing defect management did not always record key data about changes made to the status of identified defects. To its credit, the program office recently took steps to address this, thereby reducing the risk of defect status errors or unauthorized changes. Also, while the program office‘s change review and approval procedures include important steps, such as considering the impact of a change, and program officials told GAO that cost and schedule impacts of a change are discussed at control board meetings, GAO‘s analysis of 60 randomly selected change requests showed no evidence that cost and schedule impacts were in fact considered. Without such key information, decision-making authorities lack an adequate basis for making informed investment decisions, which could result in cost overruns and schedule delays. The Navy has not effectively managed its IV&V activities, which are designed to obtain an unbiased position on whether product and process standards are being met. In particular, the Navy has not ensured that the IV&V contractor is independent of the products and processes that it is reviewing. Specifically, the same contractor responsible for performing IV&V of Navy ERP products (e.g., system releases) is also responsible for ensuring that system releases are delivered within cost and schedule constraints. Because performance of this system development and management role makes the contractor potentially unable to render impartial assistance to the government in performing the IV&V function, there is an inherent conflict of interest. In addition, the IV&V agent reports directly and solely to the program manager and not to program oversight officials. As GAO has previously reported, the IV&V agent should report the findings and associated risks to program oversight officials, as well as program management, in order to better ensure that the IV&V results are objective and that the officials responsible for making program investment decisions are fully informed. Furthermore, the contractor has largely not produced the range of IV&V deliverables that were contractually required between 2006 and 2008. To its credit, the program office recently began requiring the contractor to provide assessment reports, as required under the contract, as well as formal quarterly reports; the contractor delivered the results of the first planned assessment in March 2009. Notwithstanding the recent steps that the program office has taken, it nevertheless lacks an independent perspective on the program‘s products and management processes. What GAO Recommends: GAO is making recommendations to the Secretary of Defense aimed at improving the program‘s system change request review and approval process and its IV&V activities. DOD concurred with the recommendations and identified actions that it plans to take. View [hyperlink, http://www.gao.gov/products/GAO-09-841] or key components. For more information, contact Randolph C. Hite at (202) 512- 3439 or hiter@gao.gov. [End of section] Contents: Letter: Background: Key Aspects of Navy ERP Testing Have Been Effectively Managed: System Changes Have Been Controlled, but Their Cost and Schedule Impacts Were Not Sufficiently Considered: Navy ERP IV&V Function Is Not Independent and Has Not Been Fully Performed: Conclusions: Recommendations for Executive Action: Agency Comments: Appendix I: Objectives, Scope, and Methodology: Appendix II: Comments from the Department of Defense: Appendix III: GAO Contact and Staff Acknowledgments: Tables: Table 1: Navy Systems Commands and Their Responsibilities: Table 2: Navy ERP Template 1 Releases: Table 3: Organizations Responsible for Navy ERP Oversight and Management: Table 4: Navy ERP Program Contracts: Table 5: Description of the Purpose of Navy ERP Tests: Table 6: Navy ERP Testing-Related Organizations and Respective Roles and Responsibilities: Table 7: Roles and Responsibilities for Change Review and Approval: Figures: Figure 1: Navy ERP Timeline: Figure 2: Navy ERP Life-Cycle Cost Estimates in Fiscal Years 2003, 2004, and 2007: Figure 3: Navy ERP Deployment Schedule: Figure 4: Release 1.0 and 1.1 Test Activity Schedule: Abbreviations: DOD: Department of Defense: DON: Department of the Navy: ERP: Enterprise Resource Planning: FISC: Fleet Industrial Supply Center: FOC: full operational capability: FOT&E: follow-on operational test and evaluation: GCSS-MC: Global Combat Support System--Marine Corps: GDIT: General Dynamics Information Technology: IOC: initial operational capability: IOT&E: initial operational test and evaluation: IST: integrated system testing: IT: information technology: IV&V: independent verification and validation: MDA: milestone decision authority: NAVAIR: Naval Air Systems Command: NAVSEA: Naval Sea Systems Command: NAVSUP: Naval Supply Systems Command: NTCSS: Naval Tactical Command Support System: OT&E: operational test and evaluation: SAP: Systems Applications and Products: SPAWAR: Space and Naval Warfare Systems Command: TEMP: Test and Evaluation Master Plan: UAT: user acceptance testing: [End of section] United States Government Accountability Office: Washington, DC 20548: September 15, 2009: The Honorable Evan Bayh: Chairman: The Honorable Richard Burr: Ranking Member: Subcommittee on Readiness and Management Support: Committee on Armed Services: United States Senate: The Honorable John Ensign: United States Senate: For decades, the Department of the Defense (DOD) has been challenged in modernizing its timeworn business systems.[Footnote 1] In 1995, we designated DOD's business systems modernization program as high-risk, and continue to do so today.[Footnote 2] Our reasons include the modernization's large size, complexity, and its critical role in addressing other high-risk areas, such as overall business transformation and financial management. Moreover, we continue to report on business system investments that fail to effectively employ acquisition management controls and deliver promised benefits and capabilities on time and within budget.[Footnote 3] Nevertheless, DOD continues to invest billions of dollars in thousands of these business systems, 11 of which account for about two-thirds of the department's annual spending on business programs. The Navy Enterprise Resource Planning (ERP) program is one such program. Initiated in 2003, Navy ERP is to standardize the Navy's acquisition, financial, program management, plant and wholesale supply, and workforce management business processes across its dispersed organizational environment. As envisioned, the program consists of a series of major increments, the first of which includes three releases and is expected to cost approximately $2.4 billion over its 20-year life cycle and to be fully operational in fiscal year 2013. We recently reported that Navy ERP program management weaknesses had contributed to a 2-year schedule delay and about $570 million in cost overruns. [Footnote 4] As agreed, our objectives were to determine whether (1) system testing is being effectively managed, (2) system changes are being effectively controlled, and (3) independent verification and validation (IV&V) activities are being effectively managed. To accomplish this, we analyzed relevant program documentation, such as test management documents, individual test plans and procedures and related test results and defect reports; system change procedures and specific change requests and decisions; change review board minutes; and verification and validation plans and contract documents. We also observed the use of tools for recording and tracking test defects and change requests, including tracing a statistically valid sample of transactions through these tools. We conducted this performance audit from August 2008 to September 2009, in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Additional details on our objectives, scope, and methodology are in appendix I. Background: The Department of the Navy's (DON) primary mission is to organize, train, maintain, and equip combat-ready naval forces capable of winning wars, deterring aggression by would-be foes, preserving freedom of the seas, and promoting peace and security. Its operating forces, known as the fleet, are supported by four systems commands. Table 1 provides a brief description of each command's responsibilities. Table 1: Navy Systems Commands and Their Responsibilities: Systems command: Naval Air Systems Command (NAVAIR); Responsibilities: Developing, delivering, and supporting aircraft and weapons used by sailors and marines. Systems command: Naval Supply Systems Command (NAVSUP); Responsibilities: Providing supply, fuel, transportation, and other logistics programs. Systems command: Space and Naval Warfare Systems Command (SPAWAR); Responsibilities: Developing, delivering, and supporting specialized command and control technologies, business information technology, and space capabilities. Systems command: Naval Sea Systems Command (NAVSEA); Responsibilities: Acquiring and maintaining the department's ships and submarines. Source: GAO analysis of DON data. [End of table] To support the department's mission, these commands perform a variety of interrelated and interdependent business functions (e.g., acquisition and financial management), relying heavily on business systems to do so. In fiscal year 2009, DON's budget for business systems and associated infrastructure was about $2.7 billion, of which about $2.2 billion was allocated to operations and maintenance of existing systems and about $500 million to systems in development and modernization. Of the approximately 2,480 business systems that DOD reports having, DON accounts for 569, or about 23 percent, of the total. Navy ERP is one such system investment. Navy ERP: A Brief Description: In July 2003, the Assistant Secretary of the Navy for Research, Development, and Acquisition established Navy ERP to converge the functionality of four pilot systems that were under way at the four commands into one system.[Footnote 5] According to DOD, Navy ERP is to address the Navy's long-standing problems related to financial transparency and asset visibility. Specifically, the program is intended to standardize the Navy's acquisition, financial, program management, plant and wholesale supply, and workforce management business processes across its dispersed organizational components, and support about 86,000 users when fully implemented. Navy ERP is being developed in a series of increments using the Systems Applications and Products (SAP) commercial software package, augmented as needed by customized software. SAP consists of multiple, integrated functional modules that perform a variety of business-related tasks, such as finance and acquisition. The first increment, called Template 1, is currently the only funded portion of the program and consists of three releases (1.0, 1.1, and 1.2).[Footnote 6] Release 1.0, Financial and Acquisition, is the largest of the three releases in terms of Template 1 functional requirements.[Footnote 7] See table 2 for a description of these releases. Table 2: Navy ERP Template 1 Releases: Release: 1.0 Financial and Acquisition; Functionality: * General Fund and Navy Working Capital Fund finance applications, such as billing, budgeting, and cost planning; * Acquisition applications, such as activity-based costing, contract awards, and budget exhibits; * Workforce management applications, such as personnel administration and training, as well as events management. Release: 1.1 Wholesale and Retail Supply; Functionality: * Wholesale applications, such as supply and demand planning, order fulfillment, and supply forecasting; * Retail supply applications, such as inventory management, supply and demand processing, and warehouse management. Release: 1.2 Intermediate-Level Maintenance; Functionality: * Maintenance applications, such as maintenance management, quality management, and calibration management. Source: GAO analysis of DON data. [End of table] DON estimates the life-cycle cost for Template 1 to be about $2.4 billion, including about $1 billion for acquisition and $1.4 billion for operations and maintenance. The program office reported that approximately $600 million was spent from fiscal year 2004 through fiscal year 2008. For fiscal year 2009, about $190 million is planned to be spent. Program Oversight, Management, and Contractor Roles and Responsibilities: To acquire and deploy Navy ERP, DON established a program management office within the Program Executive Office for Executive Information Systems. The program office manages the program's scope and funding and is responsible for ensuring that the program meets its key objectives. To accomplish this, the program office performs program management functions, including testing, change control, and IV&V. In addition, various DOD and DON organizations share program oversight and review activities. A listing of key entities and their roles and responsibilities is provided in table 3. Table 3: Organizations Responsible for Navy ERP Oversight and Management: Entity: Under Secretary of Defense for Acquisition, Technology, and Logistics; Roles and responsibilities: Serves as the milestone decision authority (MDA), which according to DOD, has overall responsibility for the program, to include approving the program to proceed through its acquisition cycle on the basis of, for example, independent operational test evaluation and certification. Entity: Assistant Secretary of the Navy, Research, Development, and Acquisition; Roles and responsibilities: Serves as DON's oversight organization for the program, to include enforcement of Under Secretary of Defense for Acquisition, Technology, and Logistics policies and procedures. Entity: DON, Program Executive Office for Executive Information Systems; Roles and responsibilities: Oversees a portfolio of large-scale projects and programs designed to enable common business processes and provide standard capabilities, to include reviewing and approving overarching test plans and user acceptance test readiness. Entity: Navy ERP Senior Integration Board; Roles and responsibilities: Reviews progress in attaining acceptable system performance at systems commands, including approving new system capabilities. Chaired by the Principal Deputy Assistant Secretary of the Navy. Entity: Navy ERP Program Management Office; Roles and responsibilities: Performs day-to-day program management and serves as the single point of accountability for managing the program's objectives through development, testing, deployment, and sustainment. Source: GAO analysis of DOD data. [End of table] To deliver system and other program capabilities and to provide program management support services, Navy ERP relies on multiple contractors, as described in table 4. Table 4: Navy ERP Program Contracts: Contract: Release 1.0 System Integration; Award date: September 2004; Completion date: February 2008; Contract value: $176 million; Awarded to: BearingPoint; Purpose: Design and development of release 1.0; training and deployment at NAVAIR. Contract: Release 1.1 & 1.2 System Integration; Award date: June 2007; Completion date: September 2011; Contract value: $152.9 million; Awarded to: IBM; Purpose: Design and development of release 1.1 and 1.2. Contract: Professional Support Service 1; Award date: June 2006; Completion date: September 2010; Contract value: $163.7 million; Awarded to: IBM; Purpose: Business process analysis, training, organizational change management, and deployment and sustainment support. Contract: Professional Support Service 2; Award date: June 2006; Completion date: September 2010; Contract value: $69 million; Awarded to: General Dynamics Information Technology; Purpose: Support to the government in its oversight of the system integrators and other contractors, release management, and IV&V. Source: GAO analysis of DON data. [End of table] Overview of Navy ERP's Status: Template 1 of Navy ERP was originally planned to reach full operational capability (FOC) in fiscal year 2011, and its original estimated life- cycle cost was about $1.87 billion.[Footnote 8] The estimate was later baselined[Footnote 9] in August 2004 at about $2.0 billion.[Footnote 10] In December 2006 and again in September 2007, the program was rebaselined. FOC is now planned for fiscal year 2013, and the estimated life-cycle cost is about $2.4 billion (a 31 percent increase over the original estimate).[Footnote 11] The program is currently in the production and deployment phase of the defense acquisition system, having completed the system development and demonstration phase in September 2007.[Footnote 12] This was 17 months later than the program's original schedule set in August 2004, but on time according to the revised schedule set in December 2006. Changes in the program's acquisition phase timeline are depicted in figure 1, and life-cycle cost estimates are depicted in figure 2. Figure 1: Navy ERP Life-Cycle Cost Estimates in Fiscal Years 2003, 2004, and 2007: [Refer to PDF for image: illustration] Phase: Concept refinement and technology development; Fiscal year 2003-2004: Program established (activity prior to the 2004 plan); Phase: System development and demonstration; Fiscal year 2004-2006: 2004 plan; Fiscal year 2004-2008: 2007 plan; Phase: Production and deployment; Fiscal year 2006-2011: 2004 plan; * 2006: Initial Operational Capability; * 2011: Full Operational Capability. Fiscal year 2007-2013: 2007 plan; * 2008: Initial Operational Capability; * 2013: Full Operational Capability. Source: GAO analysis of DON data. [End of figure] Figure 2: Navy ERP Life-Cycle Cost Estimates in Fiscal Years 2003, 2004, and 2007: [Refer to PDF for image: vertical bar graph] Fiscal year: 2003; Navy ERP Cost estimate: $1.87 billion. Fiscal year: 2004; Navy ERP Cost estimate: $1.99 billion. Fiscal year: 2007; Navy ERP Cost estimate: $2.44 billion. Source: GAO analysis of DON data. [End of figure] Release 1.0 was deployed at NAVAIR in October 2007, after passing developmental testing and evaluation. Initial operational capability (IOC) was achieved in May 2008, 22 months later than the baseline established in August 2004, and 4 months later than the new baseline established in September 2007. According to program documentation, these delays were due, in part, to challenges experienced at NAVAIR in converting data from legacy systems to run on the new system and implementing new business procedures associated with the system. In light of the delays at NAVAIR in achieving IOC, the deployment schedules for the other commands were revised in 2008. Release 1.0 was deployed at NAVSUP in October 2008 as scheduled, but deployment at SPAWAR was rescheduled for October 2009, 18 months later than planned, and at NAVSEA General Fund in October 2010, and at Navy Working Capital Fund in October 2011, each 12 months later than planned. Release 1.1 is currently being developed and tested, and is planned to be deployed at NAVSUP in February 2010, 7 months later than planned, and at the Navy's Fleet and Industrial Supply Centers (FISC)[Footnote 13] starting in February 2011. Changes in the deployment schedule are depicted in figure 3. Figure 3: Navy ERP Deployment Schedule: [Refer to PDF for image: illustration] Release: 1.0 Financial and Acquisition; NAVAIR: Late FY 2007 (2007 plan); NAVSUP: Late FY 2008 (2007 and 2008 plan); SPAWAR: Mid FY 2007 (2007 plan); Early FY 2010 (2008 plan); NAVSEA for General Fund: Early FY 2010 (2007 plan); Early FY 2011 (2008 plan); NAVSEA for Working Capital Fund: Early FY 2011 (2007 plan); Early FY 2012 (2008 plan). Release: 1.1 Wholesale and Retail Supply; NAVSUP: Late FY 2009 (2007 plan); Mid FY 2010 (2008 plan); FISC (first of seven deployments): Mid FY 2011 (2008 plan). Source: GAO analysis of DON data. [End of figure] Prior GAO Reviews of DOD Business System Investments Have Identified IT Management Weaknesses: We have previously reported that DOD has not effectively managed key aspects of a number of business system investments,[Footnote 14] including Navy ERP. Among other things, our reviews have identified weaknesses in such areas as architectural alignment and informed investment decision making, which are the focus of the Fiscal Year 2005 Defense Authorization Act business system provisions.[Footnote 15] Our reviews have also identified weaknesses in other system acquisition and investment management areas, such as earned value management,[Footnote 16] economic justification, risk management, requirements management, test management, and IV&V practices. In September 2008, we reported that DOD had implemented key information technology (IT) management controls on Navy ERP to varying degrees of effectiveness.[Footnote 17] For example, the control associated with managing system requirements had been effectively implemented, and important aspects of other controls had been at least partially implemented, including those associated with economically justifying investment in the program and proactively managing program risks. However, other aspects of these controls, as well as the bulk of what was needed to effectively implement earned value management, had not been effectively implemented. As a result, the controls that were not effectively implemented had, in part, contributed to sizable cost and schedule shortfalls. Accordingly, we made recommendations aimed at improving cost and schedule estimating, earned value management, and risk management. DOD largely agreed with our recommendations. In July 2008, we reported that DOD had not implemented key aspects of its IT acquisition policies and related guidance on its Global Combat Support System-Marine Corps (GCSS-MC) program.[Footnote 18] For example, we reported that it had not economically justified its investment in GCSS-MC on the basis of reliable estimates of both benefits and costs and had not effectively implemented earned value management. Moreover, the program office had not adequately managed all program risks and had not used key system quality measures. We concluded that by not effectively implementing these IT management controls, the program was at risk of not delivering a system solution that optimally supports corporate mission needs, maximizes capability mission performance, and is delivered on time and within budget. Accordingly, we made recommendations aimed at strengthening cost estimating, schedule estimating, risk management, and system quality measurement. The department largely agreed with our recommendations. In July 2007, we reported that the Army's approach for investing about $5 billion in three related programs--the General Fund Enterprise Business System, Global Combat Support System-Army Field/Tactical, and Logistics Modernization Program--did not include alignment with the Army enterprise architecture or use of a portfolio-based business system investment review process.[Footnote 19] Further, the Logistics Modernization Program's testing was not adequate and had contributed to the Army's inability to resolve operational problems. In addition, the Army had not established an IV&V function for any of the three programs. Accordingly, we recommended, among other things, use of an independent test team and establishment of an IV&V function. DOD agreed with the recommendations. In December 2005, we reported that DON had not, among other things, economically justified its ongoing and planned investment in the Naval Tactical Command Support System (NTCSS) and had not adequately conducted requirements management and testing activities.[Footnote 20] Specifically, requirements were not traceable and developmental testing had not identified problems that, subsequently, twice prevented the system from passing operational testing. Moreover, DON had not effectively performed key measurement, reporting, budgeting, and oversight activities. We concluded that DON could not determine whether NTCSS, as defined and as being developed, was the right solution to meet its strategic business and technological needs. Accordingly, we recommended developing the analytical basis necessary to know if continued investment in NTCSS represented a prudent use of limited resources, and strengthening program management, conditional upon a decision to proceed with further investment in the program. The department largely agreed with our recommendations. In September 2005, we reported that while Navy ERP had the potential to address some of DON's financial management weaknesses, it faced significant challenges and risks, including developing and implementing system interfaces with other systems and converting data from legacy systems.[Footnote 21] Also, we reported that the program was not capturing quantitative data to assess effectiveness, and had not established an IV&V function. We made recommendations to address these areas, including having the IV&V agent report directly to program oversight bodies, as well as the program manager. DOD generally agreed with our recommendations, including that an IV&V function should be established. However, it stated that the IV&V team would report directly to program management who in turn would inform program oversight officials of any significant IV&V results. In response, we reiterated the need for the IV&V to be independent of the program and stated that performing IV&V activities independently of the development and management functions helps to ensure that the results are unbiased and based on objective evidence. We also reiterated our support for the recommendation that the IV&V reports be provided to the appropriate oversight body so that it can determine whether any of the IV&V results are significant. We noted that doing so would give added assurance that the results were objective and that those responsible for authorizing future investments in Navy ERP have the information needed to make informed decisions. Key Aspects of Navy ERP Testing Have Been Effectively Managed: To be effectively managed, testing should be planned and conducted in a structured and disciplined fashion. According to DOD and industry guidance,[Footnote 22] system testing should be progressive, meaning that it should consist of a series of test events that first focus on the performance of individual system components, then on the performance of integrated system components, followed by system-level tests that focus on whether the entire system (or major system increments) is acceptable, interoperable with related systems, and operationally suitable to users. For this series of related test events to be conducted effectively, all test events need to be, among other things, governed by a well-defined test management structure and adequately planned. Further, the results of each test event need to be captured and used to ensure that problems discovered are disclosed and corrected. Key aspects of Navy ERP testing have been effectively managed. Specifically, the program has established an effective test management structure, key development events were based on well-defined plans, the results of all executed test events were documented, and problems found during testing (i.e., test defects) were captured in a test management tool and subsequently analyzed, resolved, and disclosed to decision makers. Further, while we identified instances in which the tool did not contain key data about defects that are needed to ensure that unauthorized changes to the status of defects do not occur, the number of instances found are not sufficient to conclude that the controls were not operating effectively. Notwithstanding the missing data, this means that Navy ERP testing has been performed in a manner that increases the chances that the system will meet operational needs and perform as intended. A Well-defined Test Management Structure Has Been Established: The program office has established a test management structure that satisfies key elements of DOD and industry guidance.[Footnote 23] For example, the program has developed a Test and Evaluation Master Plan (TEMP) that defines the program's test strategy. As provided for in the guidance, this strategy consists of a sequence of tests in a simulated environment to verify first that individual system parts meet specified requirements (i.e., development testing) and then verify that these combined parts perform as intended in an operational environment (i.e., operational testing). As we have previously reported,[Footnote 24] such a sequencing of test events is an effective approach because it permits the source of defects to be isolated sooner, before it is more difficult and expensive to address. More specifically, the strategy includes a sequence of developmental tests for each release consisting of three cycles of integrated system testing (IST) followed by user acceptance testing (UAT). Following development testing, the sequence of operational tests includes the Navy's independent operational test agency conducting initial operational test and evaluation (IOT&E) and then follow-on operational test and evaluation (FOT&E), as needed, to validate the resolution of deficiencies found during IOT&E. See table 5 for a brief description of the purpose of each test activity, and figure 4 for the schedule of Release 1.0 and 1.1 test activities. Table 5: Description of the Purpose of Navy ERP Tests: Test: Developmental testing: IST; Purpose: To validate that the technical and functional components of the system work properly together and operate as specified by the requirements. Test: Developmental testing: Cycle 1 (Scenario Testing); Purpose: To validate chains of business process transactions using small scenarios, such as a standard sales order, delivery, and invoicing. Also, independent evaluators observe scenario testing in preparation for operational test and evaluation. Test: Developmental testing: Cycle 2 (Scenario Testing and Conversions and Interfaces); Purpose: To validate more complex sequences of transactions plus customized software. Test: Developmental testing: Cycle 3 (Final Integration Testing); Purpose: To validate the entire system, including external components. Test: Developmental testing: UAT; Purpose: To allow the customer to ensure Navy ERP works properly and operates as specified by the requirements. Test: Operational testing: IOT&E; Purpose: To evaluate the operational effectiveness and suitability of the system. Test: Operational testing: FOT&E; Purpose: To verify the correction of deficiencies identified during IOT&E. Source: GAO analysis of DON data. [End of table] Figure 4: Release 1.0 and 1.1 Test Activity Schedule: [Refer to PDF for image: illustration] Release 1.0: Developmental testing: IST Cycle 1: Duration FY 2007, October-November; IST Cycle 2: Duration FY 2007, December-January; IST Cycle 3: Duration FY 2007, February-June; UAT: Duration FY 2007, July-August; Release 1.0: IOT&E: at NAVAIR: Duration, FY 2008, November-March; Release 1.0: FOT&E: at NAVAIR and NAVSUP: Duration, FY 2009, January-April. Release 1.1: Developmental testing: IST Cycle 1: Duration, FY 2009, January-February; IST Cycle 2: Duration, FY 2009, March-April; IST Cycle 3: Duration, FY 2009, May-October, 2010; UAT: Duration, FY 2010, October-December; Release 1.1: IOT&E: at NAVSUP: Duration, FY 2010, May-August. Source: GAO analysis of DON data. [End of figure] The TEMP also clearly identifies the roles and responsibilities of key Navy ERP testing organizations, as provided for in DOD and industry guidance. For example, it describes specific responsibilities of the program manager, system integrator, quality assurance/test team lead, and independent operational test and evaluation organizations. Table 6 summarizes the responsibilities of these various test organizations. Table 6: Navy ERP Testing-Related Organizations and Respective Roles and Responsibilities: Testing-related organization: Program manager; Responsibilities: Provides overall management and direction of Navy ERP test and evaluation; Conducts test readiness reviews; Certifies that the program is ready to proceed from developmental to operational testing in a developmental test and evaluation report. Testing-related organization: System integrator; Responsibilities: Supports the execution of integration and user acceptance testing, including training system testers and users; Reports to the Navy ERP program manager. Testing-related organization: Quality assurance/test team lead; Responsibilities: Creates the test and evaluation strategy and developmental test and evaluation plan; Assists in planning, coordinating, and conducting developmental testing and evaluation, and reporting the results to the program manager; Conducts integration testing. Testing-related organization: Operational Test and Evaluation Force; Responsibilities: Plans and conducts Navy ERP operational test and evaluation (OT&E); Reports results and recommendations to DOD's Director, Operational Test and Evaluation; Performs follow-on OT&E to verify that deficiencies found during initial OT&E have been resolved. Testing-related organization: Joint Interoperability Test Command; Responsibilities: Certifies to the Joint Chiefs of Staff that interoperability requirements are met; Verifies readiness for interoperability to the responsible operational test agency during or prior to operational test readiness review. Testing-related organization: Office of Director, Operational Test and Evaluation; Responsibilities: Reviews and approves IOT&E and FOT&E plans; Analyzes OT&E results; Provides independent assessment to the MDA. Source: GAO analysis of DOD data. [End of table] Well-defined Plans for Developmental Test Events Were Developed: According to relevant guidance,[Footnote 25] test activities should be governed by well-defined and approved plans. Among other things, such plans are to include a defect triage process, metrics for measuring progress in resolving defects, test entrance and exit criteria, and test readiness reviews. Each developmental test event for Release 1.0 (i.e., each cycle of integrated systems testing and user acceptance testing) was based on a well-defined test plan. For example, each plan provided for conducting daily triage meetings to (1) assign new defects a criticality level using documented criteria,[Footnote 26] (2) record new defects and update the status of old defects in the test management tool, and (3) address other defect and testing issues. Further, each plan included defect metrics, such as the number of defects found and corrected and their age. In addition, each plan specified that testing was not complete until all major defects found during the cycle were resolved, and all unresolved defects' impact on the next test event were understood. Further, the plans provided for holding test readiness reviews to review test results as a condition for proceeding to the next event. By ensuring that plans for key development test activities include these aspects of effective test planning, the risk of test activities not being effectively and efficiently performed is reduced, thus increasing the chances that the system will meet operational requirements and perform as intended. Test Results Were Documented and Reported, but Key Information about Changes to the Status of Reported Defects Was Not Always Recorded: According to industry guidance[Footnote 27], effective system testing includes capturing, analyzing, resolving, and disclosing to decision makers the status of problems found during testing (i.e., test defects). Further, this guidance states that these results should be collected and stored according to defined procedures and placed under appropriate levels of control to ensure that any changes to the results are fully documented. To the program's credit, the relevant testing organizations have documented test defects in accordance with defined plans. For example, daily triage meetings involving the test team lead, testers, and functional experts were held to review each new defect, assign it a criticality level, and designate someone responsible for resolving it and for monitoring and updating its resolution in the test management tool. Further, test readiness reviews were conducted at which entrance and exit criteria for each key test event were evaluated before proceeding to the next event. As part of these reviews, the program office and oversight officials, command representatives, and test officials reviewed the results of test events to ensure, among other things, that significant defects were closed and that there were no unresolved defects that could affect execution of the next test event. However, the test management tool did not always contain key data for all recorded defects that are needed to ensure that unauthorized changes to the status of defects do not occur. According to information systems auditing guidelines,[Footnote 28] audit tools should be in place to monitor user access to systems to detect possible errors or unauthorized changes. For Navy ERP, this was not always the case. Specifically, while the tool has the capability to track changes to test defects in a history log,[Footnote 29] our analysis of 80 randomly selected defects in the tool disclosed two instances in which the tool did not record when a change in the defect's status was made or who made the change. In addition, our analysis of 12 additional defects that were potential anomalies[Footnote 30] disclosed two additional instances where the tool did not record when a change was made and who made it. While our sample size and results do not support any conclusions as to the overall effectiveness of the controls in place for recording and tracking test defect status changes, they do show that it is possible that changes can be made without a complete audit trail surrounding those changes. After we shared our results with program officials, they stated that they provided each instance for resolution to the vendor responsible for the tracking tool. These officials attributed these instances to vendor updates to the tool that caused the history settings to default to "off." To address this weakness, they added that they are now ensuring that the history logs are set correctly after any update to the tool. This addition is a positive step because without an effective information system access audit tool, the probability of test defect status errors or unauthorized changes is increased. System Changes Have Been Controlled, but Their Cost and Schedule Impacts Were Not Sufficiently Considered: Industry best practices and DOD guidance[Footnote 31] recognize the importance of system change control when developing and maintaining a system. Once the composition of a system is sufficiently defined, a baseline configuration is normally established, and changes to that baseline are placed under a disciplined change control process to ensure that unjustified and unauthorized changes are not introduced. Elements of disciplined change control include (1) formally documenting a change control process, (2) rigorously adhering to the documented process, and (3) adopting objective criteria for considering a proposed change, including its estimated cost and schedule impact. To its credit, the Navy ERP program has formally documented a change control process. Specifically, it has a plan and related procedures that include the purpose and scope of the process--to ensure that any changes made to the system are properly identified, developed, and implemented in a defined and controlled environment. It also is using an automated tool to capture and track the disposition of each change request. Further, it has defined roles and responsibilities and a related decision-making structure for reviewing and approving system changes. In this regard, the program has established a hierarchy of review and approval boards, including a Configuration Control Board to review all changes and a Configuration Management Board to further review changes estimated to require more than 100 hours or $25,000 to implement. Furthermore, a Navy ERP Senior Integration Board was recently established to review and approve requests to add, delete, or change the program's requirements. In addition, the change control process states that the decisions are to be based on, among others, the system engineering and earned value management (i.e., cost and schedule) impacts the change will introduce, such as the estimated number of work hours that will be required to effect the change. Table 7 provides a brief description of the decision-making authorities and boards and their respective roles and responsibilities. Table 7: Roles and Responsibilities for Change Review and Approval: Review and approval organizations: Navy ERP Senior Integration Board; Roles and responsibilities: Reviews and approves Engineering Change Proposals, which are proposed changes that would impact system scope, configuration, cost, or schedule by adding, deleting, or changing requirements. The board is chaired by the Principal Deputy Assistant Secretary of the Navy, Research, Development, and Acquisition. Review and approval organizations: Configuration Management Board; Roles and responsibilities: Reviews and approves change requests requiring more than 100 hours or $25,000 to implement. The board is chaired by the program manager and includes representatives from the earned value management team (i.e., cost and schedule). Review and approval organizations: Configuration Control Board; Roles and responsibilities: Reviews all change requests and approves those requiring less than 100 hours or $25,000 to implement. The board is chaired by the systems engineer and includes representatives from the earned value management team (i.e., cost and schedule). Review and approval organizations: Engineering Review Board; Roles and responsibilities: Ensures change requests are ready to proceed to the Configuration Control Board by reviewing and recommending changes. This board is facilitated and chaired by the systems engineer and the configuration manager to ensure the change request documentation is complete. Review and approval organizations: Technical Change Control Board; Roles and responsibilities: Approves or defers transport change requests, which are requests to release changes into the deployed system. The board is chaired by the production manager. Source: GAO analysis of DON documentation. [End of table] Navy ERP is largely adhering to its documented change control process. Specifically, our review of a random sample of 60 change requests and minutes of related board meetings held between May 2006 and April 2009 showed that the change requests were captured and tracked using an automated tool, and they were reviewed and approved by the designated decision-making authorities and boards, in accordance with the program's documented process. However, the program has not sufficiently or consistently considered the cost and schedule impacts of proposed changes. Our analysis of the random sample of 60 change requests, including our review of related board meeting minutes, showed no evidence that cost and schedule impacts were identified or that they were considered. Specifically, we did not see evidence that the cost and schedule impacts of these change requests were assessed. According to program officials, the cost and schedule impacts of each change were discussed at control board meetings. In addition, they provided two change requests to demonstrate this. However, while these change requests did include schedule impact, they did not include the anticipated cost impact of proposed changes. Rather, these two, as well as those in our random sample, included the estimated number of work hours required to implement the change. Because the cost of any proposed change depends on other factors besides work hours, such as labor rates, the estimated number of work hours is not sufficient for considering the cost impact of a change. In the absence of verifiable evidence that cost and schedule impacts were consistently considered, approval authorities do not appear to have been provided key information needed to fully inform their decisions on whether or not to approve a change. System changes that are approved without a full understanding of their cost and schedule impacts could result in unwarranted cost increases and schedule delays. Navy ERP IV&V Function Is Not Independent and Has Not Been Fully Performed: The purpose of IV&V is to independently ensure that program processes and products meet quality standards. The use of an IV&V function is recognized as an effective practice for large and complex system development and acquisition programs, like Navy ERP, as it provides objective insight into the program's processes and associated work products.[Footnote 32] To be effective, verification and validation activities should be performed by an entity that is managerially independent of the system development and management processes and products that are being reviewed.[Footnote 33] Among other things, such independence helps to ensure that the results are unbiased and based on objective evidence. The Navy has not effectively managed its IV&V function because it has not ensured that the contractor performing this function is independent of the products and processes that this contractor is reviewing and because it has not ensured that the contractor is meeting contractual requirements. In June 2006, DON awarded a professional support services contract to General Dynamics Information Technology (GDIT), to include responsibilities for, among other things, IV&V, program management support, and delivery of releases according to cost and schedule constraints. According to the program manager, the contractor's IV&V function is organizationally separate from, and thus independent of, the contractor's Navy ERP system development function. However, the subcontractor performing the IV&V function is also performing release management. According to the GDIT contract, the release manager is responsible for developing and deploying a system release that meets operational requirements within the program's cost and schedule constraints, but it also states that the IV&V function is responsible for supporting the government in its review, approval, and acceptance of Navy ERP products (e.g., releases). The contract also states that GDIT is eligible for an optional award fee payment based on its performance in meeting, among other things, these cost and schedule constraints. Because performance of the system development and management role makes the contractor potentially unable to render impartial assistance to the government in performing the IV&V function, the contractor has an inherent conflict of interest relative to meeting cost and schedule commitments and disclosing the results of verification and validation reviews that may affect its ability to do so. The IV&V function's lack of independence is amplified by the fact that it reports directly and solely to the program manager. As we have previously reported,[Footnote 34] the IV&V function should report the issues or weaknesses that increase the risks associated with the project to program oversight officials, as well as to program management, to better ensure that the verification and validation results are objective and that the officials responsible for making program investment decisions are fully informed. Furthermore, these officials, once informed, can ensure that the issues or weaknesses reported are promptly addressed. Without ensuring sufficient managerial independence, valuable information may not reach decision makers, potentially leading to the release of a system that does not adequately meet users' needs and operate as intended. Beyond the IV&V function's lack of independence, the program office has not ensured that the subcontractor has produced the range of deliverables that were contractually required and defined in the IV&V plan. For example, the contract and plan call for weekly and monthly reports identifying weaknesses in program processes and recommendations for improvement, a work plan for accomplishing IV&V tasks, and associated assessment reports that follow the System Engineering Plan and program schedule. However, the IV&V contractor has largely not delivered these products. Specifically, until recently, it did not produce a work plan and only monthly reports were delivered, and these reports only list meetings that the IV&V contactor attended and documents that it reviewed. They do not, for example, identify program weaknesses or provide recommendations for improvement. According to program officials, they have relied on oral reports from the subcontractor at weekly meetings, and these lessons learned have been incorporated into program guidance. According to the contractor, the Navy has expended about $1.8 million between June 2006 and September 2008 for IV&V activities, with an additional $249,000 planned to be spent in fiscal year 2009. Following our inquiries about an IV&V work plan, the IV&V contractor developed such a plan in October 2008, more than 2 years after the contract was awarded, that lists program activities and processes to be assessed, such as configuration management and testing. While this plan does not include time frames for starting and completing these assessments, meeting minutes show that the status of assessments has been discussed with the program manager during IV&V review meetings. The first planned assessment was delivered to the program in March 2009 and provides recommendations for improving the program's configuration management process, such as using the automated tool to produce certain reports and enhancing training to understand how the tool is used. Further, program officials stated that they have also recently begun requiring the contractor to provide formal quarterly reports, the first of which was delivered to the program manager in January 2009. Our review of this quarterly report shows that it provides recommendations for improving the program's risk management process and organizational change management strategy. Notwithstanding the recent steps that the program office has taken, it nevertheless lacks an independent perspective on the program's products and management processes. Conclusions: DOD's successes in delivering large-scale business systems, such as Navy ERP, are in large part determined by the extent to which it employs the kind of rigorous and disciplined IT management controls that are reflected in department policies and related guidance. While implementing these controls does not guarantee a successful program, it does minimize a program's exposure to risk and thus the likelihood that it will fall short of expectations. In the case of Navy ERP, living up to expectations is important because the program is large, complex, and critical to addressing the department's long-standing problems related to financial transparency and asset visibility. The Navy ERP program office has largely implemented a range of effective controls associated with system testing and change control, including acting quickly to address issues with the audit log for its test management tool, but more can be done to ensure that the cost and schedule impacts of proposed changes are explicitly documented and considered when decisions are reached. Moreover, while the program office has contracted for IV&V activities, it has not ensured that the contractor is independent of the products and processes that it is to review and has not held the contractor accountable for producing the full range of IV&V deliverables required under the contract. Moreover, it has not ensured that its IV&V contractor is accountable to a level of management above the program office, as we previously recommended. Notwithstanding the program office's considerable effectiveness in how it has managed both system testing and change control, these weaknesses increase the risk of investing in system changes that are not economically justified and unnecessarily limit the value that an IV&V agent can bring to a program like Navy ERP. By addressing these weaknesses, the department can better ensure that taxpayer dollars are wisely and prudently invested. Recommendations for Executive Action: To strengthen the management of Navy ERP's change control process, we recommend that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to (1) revise the Navy ERP procedures for controlling system changes to explicitly require that a proposed change's life-cycle cost impact be estimated and considered in making change request decisions and (2) capture the cost and schedule impacts of each proposed change in the Navy ERP automated change control tracking tool. To increase the value of Navy ERP IV&V, we recommend that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to (1) stop performance of the IV&V function under the existing contract and (2) engage the services of a new IV&V agent that is independent of all Navy ERP management, development, testing, and deployment activities that it may review. In addition, we reiterate our prior recommendation relative to ensuring that the Navy ERP IV&V agent report directly to program oversight officials, while concurrently sharing IV&V results with the program office. [Refer to PDF for image] [End of figure] Agency Comments: In written comments on a draft of this report, signed by the Assistant Deputy Chief Management Officer and reprinted in appendix II, the department concurred with our recommendations, and stated that it will take the appropriate corrective actions within the next 7 months. We are sending copies of this report to interested congressional committees; the Director, Office of Management and Budget; the Congressional Budget Office; and the Secretary of Defense. The report also is available at no charge on our Web site at [hyperlink, http://www.gao.gov]. If you or your staffs have any questions on matters discussed in this report, please contact me at (202) 512-3439 or hiter@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made major contributions to this report are listed in appendix III. Signed by: Randolph C. Hite: Director: Information Technology Architecture and Systems Issues: [End of section] Appendix I: Objectives, Scope, and Methodology: Our objectives were to determine whether (1) system testing is being effectively managed, (2) system changes are being effectively controlled, and (3) independent verification and validation (IV&V) activities are being effectively managed for the Navy Enterprise Resource Planning (ERP) program. To determine if Navy ERP testing is being effectively managed, we reviewed relevant documentation, such as the Test and Evaluation Master Plan and test reports and compared them with relevant federal and related guidance. Further, we reviewed development test plans and procedures for each test event and compared them with best practices to determine whether well-defined plans were developed. We also examined test results and reports, including test readiness review documentation and compared them against plans to determine whether they had been executed in accordance with the plans. Moreover, to determine the extent to which test defect data were being captured, analyzed, and reported, we inspected 80 randomly selected defects from a sample of 2,258 defects in the program's test management system. In addition, we reviewed the history logs associated with each of these 80 defects to determine whether appropriate levels of control were in place to ensure that any changes to the results were fully documented. This sample was designed with a 5 percent tolerable error rate at the 95 percent level of confidence, so that, if we found 0 problems in our sample, we could conclude statistically that the error rate was less than 4 percent. In addition, we interviewed cognizant officials, including the program's test lead and the Navy's independent operational testers, about their roles and responsibilities for test management. To determine if Navy ERP changes are being effectively controlled, we reviewed relevant program documentation, such as the change control policies, plans, and procedures, and compared them with relevant federal and industry guidance. Further, to determine the extent to which the program is reviewing and approving change requests according to its documented plans and procedures, we inspected 60 randomly selected change requests in the program's configuration management system. In addition, we reviewed the change request forms associated with these 60 change requests and related control board meeting minutes to determine whether objective criteria for considering a proposed change, including estimated cost or schedule impacts, were adopted. In addition, we interviewed cognizant officials, including the program manager and systems engineer, about their roles and responsibilities for reviewing, approving, and tracking change requests. To determine if IV&V activities are being effectively managed we reviewed Navy ERP's IV&V contract, strategy, and plans and compared them with relevant industry guidance. We also analyzed the contractual relationships relative to legal standards that govern organizational conflict of interest. In addition, we examined IV&V monthly status reports, work plans, an assessment report, and a quarterly report, to determine the extent to which contract requirements were met. We interviewed contractor and program officials about their roles and responsibilities for IV&V and to determine the extent to which the program's IV&V function is independent. We conducted this performance audit at Department of Defense offices in the Washington, D.C., metropolitan area; Annapolis, Maryland; and Norfolk, Virginia; from August 2008 to September 2009, in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Appendix II: Comments from the Department of Defense: Office Of The Deputy Chief Management Officer: 9010 Defense Pentagon: Washington, DC 20301-9010: August 21, 2009: Mr. Randolph C. Hite: Director, Information Technology Architecture and Systems Issues: U.S. Government Accountability Office: 441 G Street, NW: Washington, DC 20548: Dear Mr. Hite: This is the Department of Defense (DoD) response to the GAO draft report 09-841, "DOD Business Systems Modernization: Navy Implementing A Number Of Key Management Controls On Enterprise Resource Planning System, but Improvements Still Needed," dated July 21, 2009 (GAO Code 310666). The Department concurs with all four of GAO's recommendations. The Navy Enterprise Resource Planning (ERP) program office will take the appropriate corrective actions within the next seven months. Detailed responses to each recommendation are attached. We appreciate the support of GAO as the Department further advances in its business transformation efforts, and look forward to continuing our partnership in achieving our shared goals. Signed by: Elizabeth A. McGrath: Assistant Deputy Chief Management Officer: Attachment(s): As stated: [End of letter] GAO Draft Report Dated July 21, 2009: GAO-09-841 (GAO Code 310666): "DOD Business Systems Modernization: Navy Implementing A Number Of Key Management Controls On Enterprise Resource Planning System, But Improvements Still Needed" Department Of Defense Comments To The GAO Recommendations: Recommendation 1: The GAO recommends that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to revise the Navy Enterprise Resource Planning (ERP) procedures for controlling system changes to explicitly require that a proposed change's life cycle cost impact be estimated and considered in making change request decisions. (Page 31/GAO Draft Report). DOD Response: Concur. The Navy ERP program office will revise the Enterprise Change Request Process and Procedures document to require explicitly that a proposed change's life cycle cost impact be estimated as part of the change control decision process. Corrective actions to address GAO's recommendation will he taken by the beginning of Fiscal Year (FY) 2010. Recommendation 2: The GAO recommends that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to capture the cost and schedule impacts of each proposed change in the Navy ERP automated change control tracking tool. (Page 31/GAO Draft Report). DOD Response: Concur. The Navy ERP program office will update its automated change control tracking tool to capture "dollarized" cost impacts and better identify schedule impacts of future proposed changes. Corrective actions to address GAO's recommendation will he taken by the beginning of FY 2010. Recommendation 3: The GAO recommends that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command, to stop performance of the Independent Verification and Validation (IV&V) function under the existing contract. (Page 31/GAO Draft Report). DOD Response: Concur. The Navy ERP program office plans to stop IV&V functions under the existing contract at the end of the current fiscal year (2009). Recommendation 4: The GAO recommends that the Secretary of Defense direct the Secretary of the Navy, through the appropriate chain of command. to engage the services of a new IV&V agent that is independent of all Navy ERP management, development, testing, and deployment activities that it may review. (Page 31/GAO Draft Report). DOD Response: Concur. The Navy ERP program plans to execute future IV&V functions using contract support that is not associated with any of the other Navy ERP program activities to ensure there is no conflict of interest, real or perceived. Corrective actions to address GAO's recommendation will be taken no later than March 2010. [End of section] Appendix III: GAO Contact and Staff Acknowledgments: GAO Contact: Randolph C. Hite, (202) 512-3439, or hiter@gao.gov: Staff Acknowledgments: In addition to the individual named above, key contributors to this report were Neelaxi Lakhmani, Assistant Director; Monica Anatalio; Carl Barden; Neil Doherty; Cheryl Dottermusch; Lee McCracken; Karl Seifert; Adam Vodraska; Shaunyce Wallace; and Jeffrey Woodward. [End of section] Footnotes: [1] Business systems are information systems, including financial and nonfinancial systems that support DOD business operations, such as civilian personnel, finance, health, logistics, military personnel, procurement, and transportation. [2] GAO, High-Risk Series: An Update, [hyperlink, http://www.gao.gov/products/GAO-09-271] (Washington, D.C.: January 2009). [3] See, for example, GAO, DOD Business Systems Modernization: Important Management Controls Being Implemented on Major Navy Program, but Improvements Needed in Key Areas, [hyperlink, http://www.gao.gov/products/GAO-08-896] (Washington, D.C.: Sept. 8, 2008); DOD Business Systems Modernization: Key Marine Corps System Acquisition Needs to Be Better Justified, Defined, and Managed, [hyperlink, http://www.gao.gov/products/GAO-08-822] (Washington, D.C.: July 28, 2008); and DOD Business Transformation: Lack of an Integrated Strategy Puts the Army's Asset Visibility System Investments at Risk, [hyperlink, http://www.gao.gov/products/GAO-07-860] (Washington, D.C.: July 27, 2007). [4] [hyperlink, http://www.gao.gov/products/GAO-08-896]. [5] The four pilots are SIGMA, CABRILLO, NEMAIS, and SMART. [6] The Navy is considering deleting the third release, Release 1.2, from Template 1. [7] Release 1.0 accounts for about 56 percent of the requirements; Release 1.1, about 33 percent; and Release 1.2, about 10 percent. [8] This 2003 estimate, which was prepared to assist in budget development and support the Milestone A/B approval, was for development, deployment, and sustainment costs in fiscal years 2003 through 2021. [9] According to DOD's acquisition guidebook, an Acquisition Program Baseline is a program manager's estimated cost, schedule, and performance goals. Goals consist of objective values, which represent what the user desires and expects, and threshold values, which represent acceptable limits. When the program manager determines that a current cost, schedule, or performance threshold value will not be achieved, the MDA must be notified, and a new baseline developed, reviewed by decision makers and, if the program is to continue, approved by the MDA. [10] According to the August 2004 Acquisition Program Baseline, this estimate is for acquisition, operations, and support for fiscal years 2004 through 2021. [11] According to the September 2007 Acquisition Program Baseline, this estimate is for acquisition, operations, and support for fiscal years 2004 through 2023. [12] The defense acquisition system is a framework-based approach that is intended to translate mission needs and requirements into stable, affordable, and well-managed acquisition programs. It was updated in December 2008 and consists of five key program life-cycle phases and three related milestone decision points--(1) Materiel Solution Analysis (previously Concept Refinement), followed by Milestone A; (2) Technology Development, followed by Milestone B; (3) Engineering and Manufacturing Development (previously System Development and Demonstration), followed by Milestone C; (4) Production and Deployment; and (5) Operations and Support. [13] Fleet and Industrial Supply Centers are located in San Diego, California; Norfolk, Virginia; Jacksonville, Florida; Puget Sound, Washington; Pearl Harbor, Hawaii; Yokosuka, Japan; and Sigonella, Italy; and provide worldwide logistics services for the Navy. [14] See, for example, [hyperlink, http://www.gao.gov/products/GAO-08-896]; GAO, DOD Business Systems Modernization: Planned Investment in Navy Program to Create Cashless Shipboard Environment Needs to be Justified and Better Managed, [hyperlink, http://www.gao.gov/products/GAO-08-922] (Washington, D.C.: Sept. 8, 2008); [hyperlink, http://www.gao.gov/products/GAO-08-822]; [hyperlink, http://www.gao.gov/products/GAO-07-860]; Information Technology: DOD Needs to Ensure that Navy Marine Corps Intranet Program Is Meeting Goals and Satisfying Customers, [hyperlink, http://www.gao.gov/products/GAO-07-51] (Washington, D.C.: Dec. 8, 2006); DOD Systems Modernization: Planned Investment in the Navy Tactical Command Support System Needs to be Reassessed, [hyperlink, http://www.gao.gov/products/GAO-06-215] (Washington, D.C.: Dec. 5, 2005); and DOD Business Systems Modernization: Navy ERP Adherence to Best Business Practices Critical to Avoid Past Failures, [hyperlink, http://www.gao.gov/products/GAO-05-858] (Washington, D.C.: Sept. 29, 2005). [15] Ronald W. Reagan National Defense Authorization Act for Fiscal Year 2005, Pub. L. No. 108-375, Sec. 332 (2004) (codified at 10 U.S.C. Sections 186 and 2222). [16] Earned value management is a means for measuring actual program progress against cost and schedule estimates. [17] [hyperlink, http://www.gao.gov/products/GAO-08-896]. [18] [hyperlink, http://www.gao.gov/products/GAO-08-822]. [19] [hyperlink, http://www.gao.gov/products/GAO-07-860]. [20] [hyperlink, http://www.gao.gov/products/GAO-06-215]. [21] [hyperlink, http://www.gao.gov/products/GAO-05-858]. [22] See, for example, Office of the Under Secretary of Defense for Acquisition, Technology, and Logistics, Department of Defense Instruction 5000.02 (Arlington, VA: Dec. 2, 2008); Defense Acquisition University, Test and Evaluation Management Guide, 5th ed. (Fort Belvoir, VA: January 2005); Institute of Electrical and Electronics Engineers, Inc., Standard for Software Verification and Validation, IEEE Std 1012-2004 (New York, NY: June 8, 2005); Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2 (Pittsburgh, PA: May 2008); and GAO, Year 2000 Computing Crisis: A Testing Guide, [hyperlink, http://www.gao.gov/products/GAO/AIMD-10.1.21] (Washington, D.C.: November 1998). [23] Office of the Under Secretary of Defense for Acquisition, Technology, and Logistics, Department of Defense Instruction 5000.02 (Arlington, VA: Dec. 2, 2008); Defense Acquisition University, Test and Evaluation Management Guide, 5th ed. (Fort Belvoir, VA: January 2005); and Institute of Electrical and Electronics Engineers, Inc., Standard for Software and System Test Documentation, IEEE Std 829-2008 (New York, NY: 2008). [24] GAO, Secure Border Initiative: DHS Needs to Address Significant Risks in Delivering Key Technology Investment, [hyperlink, http://www.gao.gov/products/GAO-08-1086] (Washington, D.C.: Sept. 22, 2008). [25] See, for example, [hyperlink, http://www.gao.gov/products/GAO/AIMD-10.1.21]. [26] According to program documentation, criticality levels range from 1 to 5, as follows: 1 is a problem that prevents accomplishment of an operational or mission critical capability; 2 is a major technical problem with no work-around solution; 3 is a major technical problem with a work-around solution; 4 is a minor technical problem; and 5 is any other defect, such as a cosmetic problem. [27] Institute of Electrical and Electronics Engineers, Inc., Standard for Information Technology--Software Life Cycle Processes-- Implementation Considerations, IEEE/EIA Std 12207.2-1997 (New York, NY: April 1998) and Software Engineering Institute, Capability Maturity Model Integration for Acquisition, version 1.2 (Pittsburgh, PA: May 2008). [28] Information Systems Audit and Control Association, Inc., IS Standards, Guidelines and Procedures for Auditing and Control Professionals (Rolling Meadows, IL: Jan. 15, 2009). [29] According to program documentation, the date a defect is entered into the system and the date the status of the defect is changed to "closed" are automatically populated. Further, changes to a defect's status, including from "new" to "open" and from "open" to "closed;" changes to the criticality level; and the user who makes the changes are tracked in a defect's history log. [30] These anomalies are defects that we found that (1) were attributed to integrated system test events, but were not detected until after the system was deployed; (2) had a criticality level that was different from the level that was reported at a test readiness review; (3) were deferred to a later test event or to post-deployment to be verified as resolved; or (4) had no criticality level. [31] See, for example, Electronics Industries Alliance, National Consensus Standard for Configuration Management , ANSI/EIA-649-1998 (Arlington, VA: August 1998) and Department of Defense, Military Handbook: Configuration Management Guidance, MIL-HDBK-61A(SE) (Washington, D.C.: Feb. 7, 2001). [32] GAO, Homeland Security: U.S. Visitor and Immigration Status Indicator Technology Program Planning and Execution Improvements Needed, [hyperlink, http://www.gao.gov/products/GAO-09-96] (Washington, D.C.: Dec. 12, 2008); Homeland Security: Recommendations to Improve Management of Key Border Security Program Need to Be Implemented, [hyperlink, http://www.gao.gov/products/GAO-06-296] (Washington, D.C.: Feb. 14, 2006); and [hyperlink, http://www.gao.gov/products/GAO-05-858]. [33] Institute of Electrical and Electronics Engineers, Inc., Standard for Software Verification and Validation, IEEE Std 1012-2004 (New York, NY: June 8, 2005). [34] See [hyperlink, http://www.gao.gov/products/GAO-07-860] and [hyperlink, http://www.gao.gov/products/GAO-05-858]. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO‘s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO‘s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.