Information Technology Management

Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved Gao ID: GAO-04-49 January 12, 2004

Over the years, the Congress has promulgated laws and the Office of Management and Budget and GAO have issued policies and guidance, respectively, on (1) information technology (IT) strategic planning/performance measurement (which defines what an organization seeks to accomplish, identifies the strategies it will use to achieve desired results, and then determines how well it is succeeding in reaching resultsoriented goals and achieving objectives) and (2) investment management (which involves selecting, controlling, and evaluating investments). To obtain an understanding of the government's implementation of these key IT management policies, congressional requesters asked GAO to determine the extent to which 26 major agencies have in place practices associated with key legislative and other requirements for (1) IT strategic planning/ performance measurement and (2) IT investment management.

Agencies' use of 12 IT strategic planning/performance measurement practices--identified based on legislation, policy, and guidance--is uneven. For example, agencies generally have IT strategic plans and goals, but these goals are not always linked to specific performance measures that are tracked. Without enterprisewide performance measures that are tracked against actual results, agencies lack critical information about whether their overall IT activities are achieving expected goals. Agencies' use of 18 IT investment management practices that GAO identified is also mixed. For example, the agencies largely have IT investment management boards, but no agency had the practices associated with the control phase fully in place. Executive-level oversight of project-level management activities provides organizations with increased assurance that each investment will achieve the desired cost, benefit, and schedule results. Agencies cited a variety of reasons for not having practices fully in place, such as that the chief information officer position had been vacant, that not including a requirement in guidance was an oversight, and that the process was being revised, although they could not always provide an explanation. Regardless of the reason, these practices are important ingredients for ensuring effective strategic planning, performance measurement, and investment management, which, in turn, make it more likely that the billions of dollars in government IT investments are wisely spent.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-04-49, Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved This is the accessible text file for GAO report number GAO-04-49 entitled 'Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved' which was released on February 11, 2004. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: January 2004: INFORMATION TECHNOLOGY MANAGEMENT: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved: GAO-04-49: GAO Highlights: Highlights of GAO-04-49, a report to congressional requesters Why GAO Did This Study: Over the years, the Congress has promulgated laws and the Office of Management and Budget and GAO have issued policies and guidance, respectively, on (1) information technology (IT) strategic planning/ performance measurement (which defines what an organization seeks to accomplish, identifies the strategies it will use to achieve desired results, and then determines how well it is succeeding in reaching results-oriented goals and achieving objectives) and (2) investment management (which involves selecting, controlling, and evaluating investments). To obtain an understanding of the government‘s implementation of these key IT management policies, congressional requesters asked GAO to determine the extent to which 26 major agencies have in place practices associated with key legislative and other requirements for (1) IT strategic planning/performance measurement and (2) IT investment management. What GAO Found: Agencies‘ use of 12 IT strategic planning/performance measurement practices”identified based on legislation, policy, and guidance”is uneven (see figure, below left). For example, agencies generally have IT strategic plans and goals, but these goals are not always linked to specific performance measures that are tracked. Without enterprisewide performance measures that are tracked against actual results, agencies lack critical information about whether their overall IT activities are achieving expected goals. Agencies‘ use of 18 IT investment management practices that GAO identified is also mixed (see figure, below right). For example, the agencies largely have IT investment management boards, but no agency had the practices associated with the control phase fully in place. Executive-level oversight of project-level management activities provides organizations with increased assurance that each investment will achieve the desired cost, benefit, and schedule results. Agencies cited a variety of reasons for not having practices fully in place, such as that the chief information officer position had been vacant, that not including a requirement in guidance was an oversight, and that the process was being revised, although they could not always provide an explanation. Regardless of the reason, these practices are important ingredients for ensuring effective strategic planning, performance measurement, and investment management, which, in turn, make it more likely that the billions of dollars in government IT investments are wisely spent. What GAO Recommends: GAO is making a number of recommendations, including that each agency take action to address IT strategic planning, performance measurement, and investment management practices that are not fully in place. In commenting on a draft of the report, most agencies generally agreed with our findings and recommendations. www.gao.gov/cgi-bin/getrpt?GAO-04-49. To view the full product, including the scope and methodology, click on the link above. For more information, contact David Powner at (202) 512-9286 or pownerd@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: Agencies' Use of IT Strategic Planning/Performance Measurement Practices Is Uneven: Agencies' Use of IT Investment Management Practices Is Mixed: Conclusions: Recommendations: Agency Comments and Our Evaluation: Appendixes: Appendix I: Recommendations to Departments and Agencies: Appendix II: Comments from the Department of Agriculture: Appendix III: Comments from the Department of Commerce: Appendix IV: Comments from the Department of Defense (including comments from the Departments of the Air Force, Army, and Navy): GAO Comments: Appendix V: Comments from the Department of Education: GAO Comments: Appendix VI: Comments from the Environmental Protection Agency: GAO Comments: Appendix VII: Comments from the General Services Administration: GAO Comments: Appendix VIII: Comments from the Department of Health and Human Services: Appendix IX: Comments from the Department of Housing and Urban Development: Appendix X: Comments from the Department of the Interior: Appendix XI: Comments from the Department of Justice: GAO Comments: Appendix XII: Comments from the Department of Labor: GAO Comments: Appendix XIII: Comments from the National Aeronautics and Space Administration: GAO Comments: Appendix XIV: Comments from the Nuclear Regulatory Commission: Appendix XV: Comments from the Social Security Administration: GAO Comments: Appendix XVI: Comments from the Department of State: GAO Comments: Appendix XVII: Comments from the U.S. Agency for International Development: GAO Comments: Appendix XVIII: Comments from the Department of Veterans Affairs: GAO Comments: Appendix XIX: GAO Contacts and Staff Acknowledgments: GAO Contacts: Staff Acknowledgments: Tables: Table 1: IT Strategic Planning/Performance Measurement Practices: Table 2: IT Investment Management Practices: Figures: Figure 1: Percentage of Agencies' Use of IT Strategic Planning/ Performance Measurement Practices: Figure 2: Percentage of Agencies' Use of IT Investment Management Practices: Abbreviations: CFO: chief financial officer: CIO: chief information officer: COTS: commercial-off-the-shelf: DHS: Department of Homeland Security: DOD: Department of Defense: EPA: Environmental Protection Agency: FISMA: Federal Information Security Management Act: GISRA: Government Information Security Reform Act: GPRA: Government Performance and Results Act: GSA: General Services Administration: HHS: Department of Health and Human Services: HUD: Department of Housing and Urban Development: IRM: information resources management: IT: information technology: IV&V: independent verification and validation: NARA: National Archives and Records Administration: NASA: National Aeronautics and Space Administration: NRC: Nuclear Regulatory Commission: NSF: National Science Foundation: OMB: Office of Management and Budget: OPM: Office of Personnel Management: SBA: Small Business Administration: SSA: Social Security Administration: USAID: U.S. Agency for International Development: VA: Department of Veterans Affairs: Letter January 12, 2004: The Honorable Susan M. Collins Chairman Committee on Governmental Affairs United States Senate: The Honorable Tom Davis Chairman Committee on Government Reform House of Representatives: The Honorable Adam H. Putnam Chairman Subcommittee on Technology, Information Policy, Intergovernmental Relations and the Census Committee on Government Reform House of Representatives: According to the President's most recent budget, the federal government spends billions of dollars annually on information technology (IT)-- reportedly investing about $50 billion in fiscal year 2002 and expecting to invest about $60 billion in fiscal year 2004.[Footnote 1] Despite this substantial investment, the government's management of information resources has produced mixed results. Although agencies have taken constructive steps to implement modern strategies, systems, and management policies and practices, our most recent high-risk and performance and accountability series identified continuing high-risk system modernization efforts and governmentwide information and technology management challenges.[Footnote 2] For years, the Congress has been working to increase the effectiveness of information and technology management in the federal government by passing legislation and providing oversight. For example, the Paperwork Reduction Act of 1995 applied life-cycle management principles to information and technology management and required that agencies indicate in strategic information resources management (IRM) plans how they are applying information resources to improve the productivity, efficiency, and effectiveness of government programs.[Footnote 3] The Clinger-Cohen Act of 1996 amended the Paperwork Reduction Act, establishing agency chief information officers (CIO) who report directly to the agency head and are responsible for information resources management activities. Among other things, the Clinger-Cohen Act also (1) required senior executive involvement in IT decision making and (2) imposed much-needed discipline in acquiring and managing technology resources. To obtain a broad view of the government's implementation of key IT management, you requested that we determine the extent to which agencies have in place practices associated with key legislative and other requirements for (1) IT strategic planning/performance measurement and (2) IT investment management. To address these objectives, we identified and reviewed major legislative requirements and executive orders pertaining to IT strategic planning/performance measurement, which defines what an organization seeks to accomplish, identifies the strategies it will use to achieve desired results, and then determines--through measurement--how well it is succeeding in reaching results-oriented goals and achieving objectives; and IT investment management, which involves selecting, controlling, and evaluating investments. Specifically, we identified 30 important IT management practices in these areas using legislative requirements, such as the Paperwork Reduction Act and the Clinger-Cohen Act, and policy and guidance issued by the Office of Management and Budget (OMB)[Footnote 4] and GAO.[Footnote 5] We selected 26 organizations for our review (23 major departments and agencies identified in 31 U.S.C. 901[Footnote 6] and the 3 military services). Results in Brief: Agencies' use of IT strategic planning/performance measurement practices is uneven--46 percent of the practices are in place, 41 percent are partially in place, and 7 percent are not in place.[Footnote 7] The lack of full implementation of these practices is of concern because effective strategic planning is important to ensure that agencies' IT goals are aligned with the strategic goals of the agency. Also important is having measures in place to monitor whether, or the extent to which, IT is supporting the agency. The agencies generally have IRM plans or IT strategic plans, but these plans do not always address important IRM elements, such as information collection, records management, or privacy. In addition, although agencies generally have goals associated with IT, these goals are not always linked to specific performance measures. Moreover, many agencies do not monitor actual-versus-expected performance against enterprisewide IT performance measures in their IRM plans. Agencies cited a variety of reasons why the strategic planning/performance measurement practices are not in place, including that there was a lack of support from agency leadership, that the agency had not been developing IRM plans until recently and recognized that the plans needed further refinement, or that the process is being revised. In addition, the agencies in our review could not always identify why the practices were not fully in place. Regardless of the reason, these practices were generally derived from legislative requirements and governmentwide policies and are fundamental ingredients to effective IT planning and performance measurement; therefore, it is important that they be implemented. Agencies' use of IT investment management practices is also mixed in that 44 percent of the practices are in place, 37 percent are partially in place, and 17 percent are not in place.[Footnote 8] Only by effectively and efficiently managing their IT resources through a robust investment management process can agencies gain opportunities to make better allocation decisions among many investment alternatives and further leverage their investments. As part of their investment management process, the agencies largely have IT investment management boards in place that are responsible for making decisions on selecting investments. However, many of these boards do not have written policies and procedures covering oversight or control of projects that cover such critical areas as corrective action plans and the tracking of such actions to resolution. Having these policies and procedures is a critical element of the control phase of a comprehensive IT investment management process, which helps ensure that investments are on track and are continuing to meet mission needs. As in the strategic planning/ performance measurement area, agencies were not always able to explain why certain IT investment management practices were not in place. However, among the reasons cited were that the CIO position had been vacant, that not including a given requirement in an investment management guide was an oversight, and that the investment management process was being revised. Nevertheless, the full implementation of the investment management practices would bring more rigor and structure to how agencies select and manage their IT investments. We are making a number of recommendations, including that each agency take action to address IT strategic planning, performance measurement, and investment management practices that are not fully in place. We received written or oral comments on a draft of this report from 25 of the agencies[Footnote 9] in our review. Most agencies generally agreed with our findings and recommendations, and some provided additional documentation and information that we incorporated into the report, as appropriate. Background: Advances in the use of IT and the Internet are continuing to change the way that federal agencies communicate, use, and disseminate information; deliver services; and conduct business. For example, electronic government (e-government) has the potential to help build better relationships between government and the public by facilitating timely and efficient interaction with citizens. To help the agencies more effectively manage IT, the Congress has established a statutory framework of requirements and roles and responsibilities relating to information and technology management. Nevertheless, the agencies face significant challenges in effectively planning for and managing their IT. Such challenges can be overcome through the use of a systematic and robust management approach that addresses critical elements, such as IT strategic planning and investment management. Federal Government's Statutory Framework for Information and Technology Management: The Congress established a statutory framework to help address the information and technology management challenges that agencies face. Under this framework, agencies are accountable for effectively and efficiently developing, acquiring, and using IT in their organizations. In particular, the Paperwork Reduction Act of 1995 and the Clinger- Cohen Act of 1996 require agency heads, acting through agency CIOs, to, among other things, * better link their IT planning and investment decisions to program missions and goals; * develop and maintain a strategic IRM plan that describes how IRM activities help accomplish agency missions; * develop and maintain an ongoing process to establish goals for improving IRM's contribution to program productivity, efficiency, and effectiveness; methods for measuring progress toward these goals; and clear roles and responsibilities for achieving these goals; * develop and implement a sound IT architecture; * implement and enforce IT management policies, procedures, standards, and guidelines; * establish policies and procedures for ensuring that IT systems provide reliable, consistent, and timely financial or program performance data; and: * implement and enforce applicable policies, procedures, standards, and guidelines on privacy, security, disclosure, and information sharing. Moreover, under the government's current legislative framework, OMB has important responsibilities for providing direction on governmentwide information and technology management and overseeing agency activities in these areas. Among OMB's responsibilities are: * ensuring agency integration of IRM plans, program plans, and budgets for the acquisition and use of IT and the efficiency and effectiveness of interagency IT initiatives; * developing and maintaining a governmentwide strategic IRM plan; * developing, as part of the budget process, a mechanism for analyzing, tracking, and evaluating the risks and results of all major capital investments made by an executive agency for information systems;[Footnote 10] * directing and overseeing the implementation of policy, principles, standards, and guidelines for the dissemination of and access to public information; * encouraging agency heads to develop and use best practices in IT acquisitions; and: * developing and overseeing the implementation of privacy and security policies, principles, standards, and guidelines. Further, in 2002, the Congress passed, and the President signed, legislation intended to improve the collection, use, and dissemination of government information and to strengthen information security. Specifically, Public Law 107-347, the E-Government Act of 2002, which was enacted in December 2002, includes provisions to promote the use of the Internet and other information technologies to provide government services electronically. The E-Government Act also contains the Federal Information Security Management Act (FISMA) of 2002, which replaced and strengthened the Government Information Security Reform legislative provisions (commonly referred to as "GISRA").[Footnote 11] Among other provisions, FISMA requires each agency, including national security agencies, to (1) establish an agencywide risk-based information security program to be overseen by the agency CIO and ensure that information security is practiced throughout the life cycle of each agency system; and (2) develop, maintain, and annually update an inventory of major information systems (including major national security systems) operated by the agency or under its control. Federal IT Challenges: Even with the framework laid out by the Congress, the federal government faces enduring IT challenges. Specifically, in January 2003, we reported on a variety of challenges facing federal agencies in continuing to take advantage of the opportunities presented by IT.[Footnote 12] Unless and until the challenges outlined below are overcome, federal agencies are unlikely to optimize their use of IT, which can affect an organization's ability to effectively and efficiently implement its programs and missions. * Pursuing opportunities for e-government. E-government offers many opportunities to better serve the public, make government more efficient and effective, and reduce costs. Federal agencies have implemented a wide array of e-government applications, including using the Internet to collect and disseminate information and forms; buy and pay for goods and services; submit bids and proposals; and apply for licenses, grants, and benefits. Although substantial progress has been made, the government has not yet fully reached its potential in this area. Recognizing this, a key element of the President's Management Agenda is the expansion of e-government to enhance access to information and services, particularly through the Internet. In response, OMB established a task force that selected a strategic set of initiatives to lead this expansion. Our review of the initial planning projects associated with these initiatives found that important aspects--such as collaboration and customer focus--had not been thought out for all of the projects and that major uncertainties in funding and milestones were not uncommon. Accordingly, we recommended that OMB take: steps as overseer of the e-government initiatives to reduce the risk that the projects would not meet their objectives.[Footnote 13] * Improving the collection, use, and dissemination of government information. The rapid evolution of IT is creating challenges in managing and preserving electronic records. Complex electronic records are increasingly being created in a decentralized environment and in volumes that make it difficult to organize them and make them accessible. Further, storage media themselves are affected by the dual problems of obsolescence and deterioration. These problems are compounded as computer hardware and application software become obsolete, since they may leave behind electronic records that can no longer be read. Overall responsibility for the government's electronic records lies with the National Archives and Records Administration (NARA). Our past work has shown that while NARA has taken some action to respond to the challenges associated with managing and preserving electronic records, most electronic records remain unscheduled; that is, their value had not been assessed and their disposition had not been determined.[Footnote 14] In addition, records of historical value were not being identified and provided to NARA; as a result, they were at risk of being lost. We recommended that NARA develop strategies for raising agency management's awareness of the importance of records management and for performing systematic inspections. In July 2003 we testified that although NARA has made progress in addressing these issues, more work remains to be done.[Footnote 15] The growth of electronic information--as well as the security threats facing our nation--are also highlighting privacy issues. For example, online privacy has emerged as one of the key--and most contentious-- issues surrounding the continued evolution of the Internet. In addition, our survey of 25 departments and agencies about their implementation of the Privacy Act--which regulates how federal agencies may use the personal information that individuals supply when obtaining government services or fulfilling obligations--found that a key characteristic of the agencies' 2,400 systems of records is that an: estimated 70 percent contained electronic records.[Footnote 16] Our survey also found that although compliance with Privacy Act provisions and related OMB guidance was generally high in many areas, according to agency reports, it was uneven across the federal government. To improve agency compliance and address issues reported by the agencies, we made recommendations to OMB, such as to direct agencies to correct compliance deficiencies, to monitor agency compliance, and to reassess its guidance. * Strengthening information security. Since September 1996, we have reported that poor information security is a high-risk area across the federal government with potentially devastating consequences.[Footnote 17] Although agencies have taken steps to redesign and strengthen their information system security programs, our analyses of information security at major federal agencies have shown that federal systems were not being adequately protected from computer-based threats. Our latest analyses of audit reports published from October 2001 through October 2002 continue to show significant weaknesses in federal computer systems that put critical operations and assets at risk.[Footnote 18] In addition, in June 2003 we testified that agencies' fiscal year 2002 reports and evaluations required by GISRA found that many agencies have not implemented security requirements for most of their systems, such as performing risk assessments and testing controls.[Footnote 19] In addition, the usefulness of agency corrective action plans may be limited when they do not identify all weaknesses or contain realistic completion dates. One of the most serious problems currently facing the government is cyber critical infrastructure protection, which is protecting the information systems that support the nation's critical infrastructures, such as national defense and power distribution. Since the September 11 attacks, warnings of the potential for terrorist cyber attacks against our critical infrastructures have increased. In addition, as greater amounts of money are transferred through computer systems, as more sensitive economic and commercial information is exchanged electronically, and as the nation's defense and intelligence communities increasingly rely on commercially available information technology, the likelihood increases that information attacks will threaten vital national interests. Among the critical infrastructure protection challenges the government faces are (1) developing a national critical infrastructure protection strategy, (2) improving analysis and warning capabilities, and (3) improving information sharing on threats and vulnerabilities. For each of the challenges, improvements have been made and continuing efforts are in progress, but much more is needed to address them. In particular, we have identified and made numerous recommendations over the last several years concerning critical infrastructure challenges that still need to be addressed. As a result of our concerns in this area, we have expanded our information security high-risk area to include cyber critical infrastructure protection.[Footnote 20] * Constructing and enforcing sound enterprise architectures. Our experience with federal agencies has shown that attempts to modernize IT environments without blueprints--models simplifying the complexities of how agencies operate today, how they want to operate in the future, and how they will get there--often result in unconstrained investment and systems that are duplicative and ineffective. Enterprise architectures offer such blueprints. Our reports on the federal government's use of enterprise architectures in both February 2002 and November 2003 found that agencies' use of enterprise architectures was a work in progress, with much to be accomplished.[Footnote 21] Nevertheless, opportunities exist to significantly improve this outlook if OMB were to adopt a governmentwide, structured, and systematic approach to promoting enterprise architecture use, measuring agency progress, and identifying and pursuing governmentwide solutions to common enterprise architecture challenges that agencies face. Accordingly, we made recommendations to OMB to address these areas. * Employing IT system and service management practices. Our work and other best-practice research have shown that applying rigorous practices to the acquisition or development of IT systems or the acquisition of IT services improves the likelihood of success. In other words, the quality of IT systems and services is governed largely by the quality of the processes involved in developing or acquiring each. For example, using models and methods that define and determine organizations' software-intensive systems process maturity that were developed by Carnegie Mellon University's Software Engineering Institute, which is recognized for its expertise in software processes, we evaluated several agencies' software development or acquisition processes. We found that agencies are not consistently using rigorous or disciplined system management practices. We have made numerous recommendations to agencies to improve their management processes, and they have taken, or plan to take, actions to improve.[Footnote 22] Regarding IT services acquisition, we identified leading commercial practices for outsourcing IT services that government entities could use to enhance their acquisition of IT services.[Footnote 23] * Using effective agency IT investment management practices. Investments in IT can have a dramatic impact on an organization's performance. If managed effectively, these investments can vastly improve government performance and accountability. If not, however, they can result in wasteful spending and lost opportunities for improving delivery of services to the public. Using our information: technology investment management maturity framework,[Footnote 24] we evaluated selected agencies and found that while some processes have been put in place to help them effectively manage their planned and ongoing IT investments, more work remains.[Footnote 25] IT Challenges Are Interdependent: Complicating the government's ability to overcome these IT management challenges are these challenges' interdependencies. As a result, the inability of an organization to successfully address one IT management area can reduce the effectiveness of its success in addressing another management function. For example, a critical aspect of implementing effective e-government solutions and developing and deploying major systems development projects is ensuring that robust information security is built into these endeavors early and is periodically revisited. The government's many IT challenges can be addressed by the use of effective planning and execution, which can be achieved, in part, through strategic planning/performance measurement, and investment management. For example, strong strategic planning is focused on using IT to help accomplish the highest priority customer needs and mission goals, while effective performance measurement helps determine the success or failure of IT activities. Finally, IT investment management provides a systematic method for minimizing risks while maximizing the return on investments and involves a process for selecting, controlling, and evaluating investments. These processes, too, are interdependent. For example, the investment management process is a principal mechanism to ensure the effective execution of an agency's IT strategic plan. Objectives, Scope, and Methodology: Our objectives were to determine the extent to which federal agencies are following practices associated with key legislative and other requirements for (1) IT strategic planning/performance measurement and (2) IT investment management. To address these objectives, we identified and reviewed major legislative requirements and executive orders pertaining to IT strategic planning, performance measurement, and investment management. Specifically, we reviewed: * the Paperwork Reduction Act of 1995; * the Clinger-Cohen Act of 1996; * the E-Government Act of 2002; * the Federal Information Security Management Act of 2002; * Executive Order 13011, Federal Information Technology; and: * Executive Order 13103, Computer Software Piracy. Using these requirements and policy and guidance issued by OMB[Footnote 26] and GAO,[Footnote 27] we identified 30 IT management practices that (1) can be applied at the enterprise level and (2) were verifiable through documentation and interviews. These 30 practices focused on various critical aspects of IT strategic management, performance measurement, and investment management, including the development of IRM plans, the identification of goals and related measures, and the selection and control of IT investments, respectively. We selected 26 major departments and agencies for our review (23 entities identified in 31 U.S.C. 901 and the 3 military services).[Footnote 28] At our request, each agency completed a self- assessment on whether and how it had implemented the 30 IT management practices. We reviewed the completed agency self-assessments and accompanying documentation, including agency and IT strategic plans, agency performance plans and reports required by the Government Performance and Results Act, and IT investment management policy and guidance, and interviewed applicable agency IT officials to corroborate whether the practices were in place. We did not evaluate the effectiveness of agencies' implementation of the practices. For example, we did not review specific IT investments to determine whether they were selected, controlled, and reviewed in accordance with agency policy and guidance. However, we reviewed applicable prior GAO and agency inspector general reports and discussed whether agency policies had been fully implemented with applicable agency IT officials. On the basis of the above information, we assessed whether the practices were in place, using the following definitions: * Yes--the practice was in place. * Partially--the agency has some, but not all, aspects of the practice in place. Examples of circumstances in which the agency would receive this designation include when (1) some, but not all, of the elements of the practice were in place; (2) the agency documented that it has the information or process in place but it was not in the prescribed form (e.g., in a specific document as required by law or OMB); (3) the agency's documentation was in draft form; or (4) the agency had a policy related to the practice but evidence supported that it had not been completely or consistently implemented. * No--the practice was not in place. * Not applicable--the practice was not relevant to the agency's particular circumstances. We also collected information from the Department of Homeland Security (DHS) but found that since it had been established so recently, it was too early to judge its IT strategic planning, performance measurement, and investment management. As a result, although we provided information on what DHS was doing with respect to these areas, we did not include it in our assessment. We also interviewed officials from OMB's Office of Information and Regulatory Affairs regarding OMB's role in establishing policies and overseeing agencies' implementation of the identified practices. We performed our work at the agencies' offices in greater Washington, D.C. We conducted our review between April and mid-December 2003 in accordance with generally accepted government auditing standards. Agencies' Use of IT Strategic Planning/Performance Measurement Practices Is Uneven: The use of IT strategic planning/performance measurement practices is uneven (see fig. 1), which is of concern because a well-defined strategic planning process helps ensure that an agency's IT goals are aligned with that agency's strategic goals. Moreover, establishing performance measures and monitoring actual-versus-expected performance of those measures can help determine whether IT is making a difference in improving performance. Among the practices or elements of practices that agencies largely have in place were those pertaining to establishing goals and performance measures. On the other hand, agencies are less likely to have fully documented their IT strategic planning processes, developed comprehensive IRM plans, linked performance measures to their enterprisewide IT goals, or monitored actual-versus-expected performance for these enterprisewide goals. Agencies cited various reasons, such as the lack of support from agency leadership, for not having strategic practices/performance measurement practices in place. Without strong strategic management practices, it is less likely that IT is being used to maximize improvement in mission performance. Moreover, without enterprisewide performance measures that are being tracked against actual results, agencies lack critical information about whether their overall IT activities, at a governmentwide cost of billions of dollars annually, are achieving expected goals. Figure 1: Percentage of Agencies' Use of IT Strategic Planning/ Performance Measurement Practices: [See PDF for image] Note: Yes--the practice was in place. Partially--the agency has some, but not all, aspects of the practice in place. Examples of circumstances in which the agency would receive this designation include when (1) some, but not all, of the elements of the practice were in place; (2) the agency documented that it has the information or process in place but it was not in the prescribed form (e.g., in a specific document as required by law or OMB); (3) the agency's documentation was in draft form; or (4) the agency had a policy related to the practice but evidence supported that it had not been completely or consistently implemented. No--the practice was not in place. Not applicable--the practice was not relevant to the agency's particular circumstances. [End of figure] Governmentwide Progress Demonstrated, but More Work Remains: Critical aspects of the strategic planning/performance measurement area include documenting the agency's IT strategic planning processes, developing IRM plans, establishing goals, and measuring performance to evaluate whether goals are being met. Although the agencies often have these practices, or elements of these practices, in place, additional work remains, as demonstrated by the following examples: * Strategic planning process. Strategic planning defines what an organization seeks to accomplish and identifies the strategies it will use to achieve desired results. A defined strategic planning process allows an agency to clearly articulate its strategic direction and to establish linkages among planning elements such as goals, objectives, and strategies. About half of the agencies fully documented their strategic planning processes. For example, the General Services Administration (GSA) documented an IT governance structure that addresses the roles and responsibilities of various organizations in strategic planning and investment management. In addition, in its IT strategic plan, GSA describes how it developed the plan, including its vision, business- related priorities, and goals. In contrast, the Department of Agriculture has not completely documented its IT strategic planning process or integrated its IT management operations and decisions with other agency processes. According to Agriculture IT officials, the department's ongoing budget and performance integration initiative is expected to result in a more clearly defined and integrated IT strategic management planning process. Such a process provides the essential foundation for ensuring that IT resources are effectively managed. * Strategic IRM plans. The Paperwork Reduction Act requires that agencies indicate in strategic IRM plans how they are applying information resources to improve the productivity, efficiency, and effectiveness of government programs. An important element of a strategic plan is that it presents an integrated system of high-level decisions that are reached through a formal, visible process. The plan is thus an effective tool with which to communicate the mission and direction to stakeholders. In addition, a strategic IRM plan that communicates a clear and comprehensive vision for how the agency will use information resources to improve agency performance is important because IRM encompasses virtually all aspects of an agency's information activities. Although the Paperwork Reduction Act also requires agencies to develop IRM plans in accordance with OMB's guidance, OMB does not provide cohesive guidance on the specific contents of IRM plans. OMB Circular A-130 directs that agencies have IRM plans that support agency strategic plans, provide a description of how IRM helps accomplish agency missions, and ensure that IRM decisions are integrated with organizational planning, budgets, procurement, financial management, human resources management, and program decisions. However, Circular A- 130 does not provide overall guidance on the plan's contents. As a result, although agencies generally provided OMB with a variety of planning documents to meet its requirement that they submit an IRM plan, these plans were generally limited to IT strategic or e- government issues and did not address other elements of IRM, as defined by the Paperwork Reduction Act. Specifically, these plans generally include individual IT projects and initiatives, security, and enterprise architecture elements but do not often address other information functions, such as information collection, records management, and privacy, or the coordinated management of all information functions. OMB IT staff agreed that the agency has not set forth guidance on the contents of agency IRM plans in a single place, stating that its focus has been on looking at agencies' cumulative results and not on planning documents. In addition, these staff also noted that agencies account for their IRM activities through multiple documents (e.g., Information Collection Budgets[Footnote 29] and Government Paperwork Elimination Act[Footnote 30] plans). However, the OMB IT staff stated that they would look at whether more guidance is needed to help agencies in their development of IRM plans, but have not yet made a commitment to provide such guidance. Half the agencies indicated a need for OMB to provide additional guidance on the development and content of IRM plans. Strong agency strategic IRM plans could also provide valuable input to a governmentwide IRM plan, which is also required by the Paperwork Reduction Act. As we reported last year, although OMB designated the CIO Council's strategic plan for fiscal years 2001-2002 as the governmentwide strategic IRM plan, it does not constitute an effective and comprehensive strategic vision.[Footnote 31] Accordingly, we recommended that OMB develop and implement a governmentwide strategic IRM plan that articulates a comprehensive federal vision and plan for all aspects of government information. In April 2003, we testified that OMB had taken a number of actions that demonstrate progress in fulfilling the Paperwork Reduction Act's requirement of providing a unifying IRM vision.[Footnote 32] However, more remains to be done. In particular, we reported that although OMB's strategies and models are promising, their ability to reduce paperwork burden and accomplish other objectives depends on how OMB implements them. One element required by the Clinger-Cohen Act to be included in agency IRM plans is the identification of a major IT acquisition program(s), or any phase or increment of that program, that significantly deviated from cost, performance, or schedule goals established by the program. However, few agencies met this requirement. In these cases, a common reason cited for not including this information was that it was not appropriate to have such detailed information in a strategic plan because such plans should be forward thinking and may not be developed every year. Agencies also identified other mechanisms that they use to track and report cost, schedule, and performance deviations. Because agencies generally do not address this Clinger-Cohen Act requirement in their IRM plans, they may benefit from additional guidance from OMB on how to address this requirement. * IT goals. The Paperwork Reduction Act and the Clinger-Cohen Act require agencies to establish goals that address how IT contributes to program productivity, efficiency, effectiveness, and service delivery to the public. We have previously reported that leading organizations define specific goals, objectives, and measures, use a diversity of measure types, and describe how IT outputs and outcomes impact operational customer and agency program delivery requirements.[Footnote 33] The agencies generally have the types of goals outlined in the Paperwork Reduction Act and the Clinger-Cohen Act. For example, the Social Security Administration (SSA) set a goal of achieving an average of at least a 2 percent per year improvement in productivity, and it expects that advances in automation will be a key to achieving this goal along with process and regulation changes. In addition, the Department of Veterans Affairs' (VA) latest departmental strategic plan has a goal that includes using business process reengineering and technology integration to speed up delivery of benefit payments, improve the quality of health care provided in its medical centers, and administer programs more efficiently. The VA goal includes strategies such as using its enterprise architecture as a continuous improvement process, implementing e-government solutions to transform paper-based electronic collections to electronic-based mechanisms, and establishing a single, high-performance wide area data network. Five agencies do not have one or more of the goals required by the Paperwork Reduction Act and the Clinger-Cohen Act. For example, the Department of Labor's single IT strategic goal--to provide better and more secure service to citizens, businesses, government, and Labor employees to improve mission performance--which it included in its fiscal year 2004 performance plan, does not address all required goals. Further, in contrast to other agencies, Labor does not have goals in its IRM plan. It is important that agencies specify clear goals and objectives to set the focus and direction of IT performance. * IT performance measures. The Paperwork Reduction Act, the Clinger- Cohen Act, and Executive Order 13103 require agencies to establish a variety of IT performance measures, such as those related to how IT contributes to program productivity, efficiency, and effectiveness, and to monitor the actual-versus-expected performance of those measures. As we have previously reported, an effective performance management system offers a variety of benefits, including serving as an early warning indicator of problems and the effectiveness of corrective actions, providing input to resource allocation and planning, and providing periodic feedback to employees, customers, stakeholders, and the general public about the quality, quantity, cost, and timeliness of products and services.[Footnote 34] Although the agencies largely have one or more of the required performance measures, these measures are not always linked to the agencies' enterprisewide IT goals. For example, the Department of Defense (DOD), Air Force, and Navy have a variety of enterprisewide IT goals but do not have performance measures associated with these goals. Each of these organizations are in the process of developing such measures. To illustrate, the Air Force's August 2002 information strategy includes nine goals, such as providing decision makers and all Air Force personnel with on-demand access to authoritative, relevant, and sufficient information to perform their duties efficiently and effectively, but does not have performance measures for these goals. The Air Force recognizes the importance of linking performance measures to its goals and is developing such measures, which it expects to complete by the fourth quarter of fiscal year 2004. Leading organizations use performance measures to objectively evaluate mission, business, and project outcomes. Such organizations also focus on performance measures for gauging service to key management processes and tailoring performance measures to determine whether IT is making a difference in improving performance. Few agencies monitored actual- versus-expected performance for all of their enterprisewide IT goals. Specifically, although some agencies tracked actual-versus-expected outcomes for the IT performance measures in their performance plans or accountability reports and/or for specific IT projects, they generally did not track the performance measures specified in their IRM plans. For example, although the Department of Health and Human Services' (HHS) IT strategic plan identifies enterprisewide goals and performance measures, these measures generally do not identify quantified outcomes (e.g., the measures indicate that the outcome will be a percentage transaction increase or cost decrease in certain areas but do not provide a baseline or target). In addition, the HHS plan does not describe how the department will monitor actual-versus-expected performance for these measures. HHS's Director of Business Operations in its IRM office reported that the department recognizes the need to develop an integrated program for monitoring performance against the enterprisewide measures in the IT strategic plan. He stated that HHS has recently begun an initiative to establish such a process. By not measuring actual-versus-expected performance, agencies lack the information to determine where to target agency resources to improve overall mission accomplishment. * Benchmarking. The Clinger-Cohen Act requires agencies to quantitatively benchmark agency process performance against public-and private-sector organizations, where comparable processes and organizations exist. Benchmarking is used by entities because there may be external organizations that have more innovative or more efficient processes than their own processes. Our previous study of IT performance measurement at leading organizations found that they had spent considerable time and effort comparing their performance information with that of other organizations.[Footnote 35] Seven agencies have mechanisms--such as policies and strategies--in place related to benchmarking their IT processes. For example, DOD's information resources and IT directive states that DOD components shall routinely and systematically benchmark their functional processes against models of excellence in the public and private sector and use these and other analyses to develop, simplify, or refine the processes before IT solutions are applied. In general, however, agencies' benchmarking decisions are ad hoc. Few agencies have developed a mechanism to identify comparable external private-or public-sector organizations and processes and/or have policies related to benchmarking; however, all but 10 of the agencies provided examples of benchmarking that had been performed. For example, the Small Business Administration (SBA) does not have benchmarking policies in place, but the agency provided an example of a benchmarking study performed by a contractor that compared SBA's IT operations and processes against industry cost and performance benchmarks and best practices and resulted in recommendations for improvement. Practice-Specific Analysis: Table 1 provides additional detail on each strategic planning/ performance measurement practice and our evaluation of whether each agency had the practice in place. The table indicates that work remains for the agencies to have each of the practices fully in place as well as that several agencies reported that they were taking, or planned to take, actions to address the practices or elements of practices. Table 1: IT Strategic Planning/Performance Measurement PracticesA: Practice 1.1: The agency has documented its IT strategic management process, including, at a minimum: * the responsibilities and accountability for IT resources across the agency, including the relationship between the chief information officer (CIO), chief financial officer (CFO), and mission/program officials; and; * the method by which the agency defines program information needs and develops strategies, systems, and capabilities to meet those needs. Results; Yes: 12; Partially: 11; No: 1; NA: 2. Comments: * Yes--the Departments of the Air Force, Army, Commerce, Defense (DOD), Education, Energy, Labor, Navy, and Veterans Affairs (VA) and the General Services Administration (GSA), the Office of Personnel Management (OPM), and the Social Security Administration (SSA) have this practice in place; * Partially--the Departments of Agriculture[C], Health and Human Services (HHS)[C], Interior, Justice, and Transportation, and the Environmental Protection Agency (EPA), the National Aeronautics and Space Administration (NASA)[C], and the Small Business Administration (SBA) do not have a completely documented IT strategic planning process. The Department of Housing and Urban Development (HUD)[C] does not clearly describe the roles and responsibilities of the CFO and program managers in IT strategic planning. The Nuclear Regulatory Commission's (NRC) roles and responsibilities in its IT strategic management process are not clearly defined. The Department of the Treasury's[C] documentation supporting this practice is in draft form; * No--the National Science Foundation (NSF) does not have this practice in place; * NA (not applicable)--the Department of State and the U.S. Agency for International Development (USAID) are transitioning to a joint strategic planning process that will support their common policy objectives. The first step in this process was the August 2003 issuance of a State/USAID strategic plan. Because a new joint IT strategic planning process is also being implemented, it is too early to evaluate whether the new process will address this practice. Practice 1.2: The agency has documented its process to integrate IT management operations and decisions with organizational planning, budget, financial management, human resources management, and program decisions. Results; Yes: 13; Partially: 10; No: 1; NA: 2. Comments: * Yes--Air Force, Army, Commerce, DOD, Education, GSA, Labor, Navy, NSF, OPM, SBA, SSA, and VA have this practice in place; * Partially--Agriculture[C] and EPA have not completely documented the integration of their IT management operations and decisions with other agency processes. Energy[C], HUD, NASA[C], and Justice have not documented how their IT management operations and decisions are integrated with human resources management. HHS[C] has not documented how its IT management operations and decisions are integrated with its budget processes. NRC reported that improvement is needed in how IT planning is integrated with the budget and human resources management. Transportation's[C] IT human capital planning is not yet integrated with the agency's human capital planning. Treasury's[C] documentation pertaining to this practice is in draft form; * No--Interior does not have this practice in place; * NA--this practice is not applicable to State and USAID for reasons outlined in practice 1.1. Practice 1.3: The agency requires that information security management processes be integrated with strategic and operational planning processes. Results; Yes: 24; Partially: 2; No: 0; NA: 0. Comments: * Yes--Agriculture, Air Force, Army, Commerce, DOD, Education, Energy, EPA, GSA, HHS, HUD, Interior, Justice, Labor, NASA, Navy, NSF, OPM, SBA, SSA, State, Transportation, USAID, and VA have this practice in place; * Partially--NRC and Treasury's[C] documentation supporting this practice is in draft form. Practice 1.4: The agency has a process that involves the CFO, or comparable official, to develop and maintain a full and accurate accounting of IT-related expenditures, expenses, and results. Results; Yes: 15; Partially: 11; No: 0; NA: 0. Comments: * Yes--Agriculture, Commerce, Energy, GSA, HUD, Interior, Justice, NASA, NRC, NSF, OPM, SSA, Transportation, Treasury, and VA reported that they have this practice in place.[B]; * Partially--prior GAO or inspector general work indicates that Army, Air Force, DOD, EPA, and Navy do not capture and report on the full costs of their programs. State and USAID reported that IT internal costs are not consistently captured. HHS reported that not all internal costs are captured and that the CFO is not involved in the process used to derive its IT costs. Education and Labor's CFOs are not involved in the process used to derive their IT costs. SBA reported that not all costs are captured for nonmajor systems. Practice 1.5: The agency prepares an enterprisewide strategic information resources management (IRM) plan that, at a minimum: * describes how IT activities will be used to help accomplish agency missions and operations, including related resources; and; * identifies a major IT acquisition program(s) or any phase or increment of that program that has significantly deviated from the cost, performance, or schedule goals established for the program. Results; Yes: 2; Partially: 22; No: 0; NA: 2. Comments: * Yes--Commerce and NSF have this practice in place; * Partially--Agriculture, Air Force, Army, EPA, GSA, HHS, HUD, Interior, Justice, Labor, NASA, OPM, and SBA's IRM plans do not include resources and major IT acquisition programs that deviated from cost, schedule, or performance goals. Education, Energy, Navy, SSA, and Transportation's IRM plans do not include major IT acquisition programs that deviated from cost, schedule, or performance goals. DOD and NRC's draft IRM plans do not include resources and major IT acquisition programs that deviated from cost, schedule, or performance goals in their IRM plans. Treasury and VA's draft IRM plans do not include resources or major IT acquisition programs that deviated from cost, schedule, or performance goals in their IRM plans; * NA--this practice is not applicable to State and USAID for reasons outlined in practice 1.1. Practice 1.6: The agency's performance plan required under Government Performance and Results Act (GPRA) includes; * a description of how IT supports strategic and program goals: * the resources and time periods required to implement the information security program plan required by the Federal Information Security Management Act (FISMA), and; * a description of major IT acquisitions contained in the capital asset plan that will bear significantly on the achievement of a performance goal. Results; Yes: 0; Partially: 23; No: 0; NA: 3. Comments: * Partially--no agency's performance plan, except VA's, includes time periods, and none includes resources required to implement the information security program plan required by FISMA. In addition, Agriculture, DOD, HHS, and Interior's plans also do not include a description of major IT acquisitions contained in their capital asset plans that bear significantly on the achievement of a performance goal; * NA--this practice is not applicable to Air Force, Army, and Navy because they are not required to produce such plans. Practice 1.7: The agency has a documented process to; * develop IT goals in support of agency needs: * measure progress against these goals, and; * assign roles and responsibilities for achieving these goals. Results; Yes: 4; Partially: 12; No: 8; NA: 2: Comments: * Yes--Army, GSA, OPM, and SSA have this practice in place; * Partially--Agriculture[C], NRC, and NSF do not have a documented process for assigning roles and responsibilities for achieving their enterprisewide IT goals. DOD[C] and HHS[C] have not established a documented process for measuring progress against their enterprisewide IT goals. Energy has this process in place for some, but not all, of its IT goals and performance measures. Air Force[C], Education, and Navy[C] do not have a documented process to measure against their enterprisewide IT goals or to assign roles and responsibilities for achieving these goals. Treasury's[C] documentation in support of this practice is in draft form. Transportation is piloting a process. VA's[C] documentation supporting this practice does not explicitly address how IT goals are developed and roles and responsibilities assigned; * No--Commerce[C], EPA, HUD[C], Interior, Justice[C], Labor, NASA, and SBA do not have this practice in place; * NA--this practice is not applicable to State and USAID for reasons outlined in practice 1.1. Practice 1.8: The agency has established goals that, at a minimum, address how IT contributes to; * program productivity: * efficiency: * effectiveness, and; * service delivery to the public (if applicable). Results; Yes: 19; Partially: 5; No: 0; NA: 2. Comments: * Yes--Agriculture, Air Force, Army, Commerce, DOD, Education, EPA, GSA, HHS, HUD, Interior, Justice, NASA, NSF, OPM, SBA, SSA, Treasury, and VA have this practice in place; * Partially--Navy does not have an IT goal associated with service delivery to the public. Energy, Labor, and Transportation do not have a goal associated with how IT contributes to program productivity. NRC's documentation in support of this practice is in draft form; * NA--this practice is not applicable to State and USAID for reasons outlined in practice 1.1. Practice 1.9: The agency has established IT performance measures and monitors actual-versus-expected performance that at least addresses; * how IT contributes to program productivity: * how IT contributes to the efficiency of agency operations: * how IT contributes to the effectiveness of agency operations: * service delivery to the public (if applicable): * how electronic government initiatives enable progress toward agency goals and statutory mandates: * the performance of IT programs (e.g., system development and acquisition projects), and; * agency compliance with federal software piracy policy. Results; Yes: 0; Partially: 23; No: 1; NA: 2. Comments: * Partially--Agriculture[C], HHS[C], Interior, NASA, OPM, and VA[C] generally do not track actual-versus-expected performance for enterprisewide measures in their IRM plans. Commerce[C], EPA, Justice, SBA, and Treasury have some enterprisewide IT performance measures in their performance plans or accountability reports in which actual- versus-expected performance is tracked but do not have measures for the enterprisewide IT goals in their IRM plans. SBA also does not have performance measures associated with program productivity, efficiency, effectiveness, and performance of IT programs. Moreover, Treasury's[C] IRM plan is in draft form. Air Force[C] has not developed measures for the enterprisewide goals in its information strategy and does not have measures associated with program productivity, electronic government, and service delivery to the public. Army[C] has neither performance measures for all of the objectives related to its enterprise IT goals nor measures associated with service delivery to the public. Navy[C] has not developed measures for the enterprisewide goals in its IRM plan and does not have measures related to how IT contributes to the effectiveness and efficiency of agency operations, service delivery to the public, or e-government. Education does not have measures related to how IT contributes to program productivity and the effectiveness and efficiency of agency operations and does not track actual-versus- expected performance of measures identified in its IRM plan. GSA did not provide evidence that it tracked actual versus expected performance for one of its IT goals in its IRM plan. HUD[C] does not have performance measures related to how IT contributed to program productivity and does not track actual-versus-expected performance for enterprisewide measures in its IRM plan. Labor does not have performance measures associated with program productivity and efficiency. Energy and NRC's performance measures are not linked to the enterprisewide IT goals contained in their IRM plans. In addition, Energy does not have a measure associated with program productivity. Transportation's[C] performance measures are generally not linked to the goals contained in its IRM plan, and it does not track actual- versus-expected performance for its enterprisewide measures. SSA reported that it has performance measures associated with the overall performance of its IT programs but provided no supporting documentation. Finally, no agency has performance measures related to the effectiveness of controls to prevent software piracy; * No--DOD[C] does not have this practice in place but is working on developing such measures; * NA--this practice is not applicable to State and USAID for reasons outlined in practice 1.1. Practice 1.10: The agency has developed IT performance measures that align with and support the goals in the GPRA performance plan. Results; Yes: 22; Partially: 0; No: 1; NA: 3. Comments: * Yes--Agriculture, Commerce, Education, Energy, EPA, GSA, HHS, HUD, Interior, Justice, Labor, NASA, NRC, NSF, OPM, SBA, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * No--DOD does not have this practice in place; * NA--this practice is not applicable to the Air Force, Army, and Navy because they are not required to produce such plans. Practice 1.11: The agency developed an annual report, included as part of its budget submission, that describes progress in achieving goals for improving the efficiency and effectiveness of agency operations and, as appropriate, the delivery of services to the public through the effective use of IT. Results; Yes: 25; Partially: 1; No: 0; NA: 0. Comments: * Yes--Agriculture, Air Force, Army, Commerce, DOD, Education, Energy, EPA, GSA, HHS, HUD, Interior, Justice, Labor, NASA, Navy, NRC, NSF, OPM, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * Partially--SBA has not reported progress on achieving its goals for improving the efficiency and effectiveness of agency operations. Practice 1.12: The agency requires that its IT management processes be benchmarked against appropriate processes and/or organizations from the public and private sectors in terms of cost, speed, productivity, and quality of outputs and outcomes where comparable processes and organizations in the public or private sectors exist. Results; Yes: 7; Partially: 9; No: 10; NA: 0. Comments: * Yes--Air Force, Army, DOD, Education, Navy, NRC, and VA have this practice in place; * Partially--Agriculture, Commerce, Energy, GSA, Interior, NASA, SBA, SSA, and Transportation provided an example of a process that they have benchmarked, but benchmarking is being performed on an ad hoc basis; * No--EPA, HHS[C], HUD[C], Justice, Labor, NSF, OPM, State, Treasury[C], and USAID do not have this practice in place. Source: GAO. [A] Due to its recent establishment, we did not include DHS as a part of this analysis. [B] We have previously reported that agencies are making progress to address financial management system weaknesses but that agency management does not yet have the full range of information needed for accountability, performance reporting, and decision making. In addition, for fiscal year 2002, auditors reported that 19 agency systems were not compliant with the Federal Financial Management Improvement Act, including Agriculture, Commerce, Education, HUD, Interior, and NASA. (Financial Management: Sustained Efforts Needed to Achieve FFMIA Accountability, [Hyperlink, http://www.gao.gov/cgi-bin/ getrpt?GAO-03-1062] GAO-03-1062, Sept. 30, 2002). [C] The agency reported that it was taking, or planned to take, action to address this practice or elements of the practice. Note: Yes--the practice was in place. Partially--the agency has some, but not all, aspects of the practice in place. Examples of circumstances in which the agency would receive this designation include when (1) some, but not all, of the elements of the practice were in place; (2) the agency documented that it has the information or process in place but it was not in the prescribed form (e.g., in a specific document as required by law or OMB); (3) the agency's documentation was in draft form; or (4) the agency had a policy related to the practice, but evidence supported that it had not been completely or consistently implemented. No--the practice was not in place. NA (not applicable)--the practice was not relevant to the agency's particular circumstances. [End of table] Agency IT officials could not identify why practices were not in place in all cases, but in those instances in which reasons were identified, a variety of explanations were provided. For example, reasons cited by agency IT officials included that they lacked the support from agency leadership, that the agency had not been developing IRM plans until recently and recognized that the plan needed further refinement, that the process was being revised (in at least one case because of changes that are needed to reflect a loss of component organizations to the new DHS), and that requirements were evolving. In other cases, the agency reported that it had the information but it was not in the format required by legislation. For instance, FISMA requires agencies to include in the performance plans required by the Government Performance and Results Act the resources, including budget, staffing, and training, and time periods to implement its information security program. None of the agencies included this information in their performance plans.[Footnote 36] However, the agencies commonly reported that they had this information but that it was in another document. Nevertheless, this does not negate the need for having the agency report to the Congress in the required form. This is particularly important since, as in the example of the FISMA requirement, the reporting requirement involves a public document, whereas other reports may not be publicly available. In the case of DHS, while we did not include the department in our assessment and in table 1, the department is in the process of developing its first IT strategic plan. According to DHS, it expects to complete this plan by mid-February 2004. Agencies' Use of IT Investment Management Practices Is Mixed: The use of IT investment management practices is mixed (as shown in fig. 2), which demonstrates that agencies do not have all the processes in place to effectively select, control, and evaluate investments. An IT investment management process is an integrated approach to managing investments that provides for the continuous identification, selection, control, life-cycle management, and evaluation of IT investments. Among the investment management practices that are most frequently in place are having investment management boards and requiring that projects demonstrate that they are economically beneficial. Practices less commonly in place are those requiring that IT investments be performed in a modular, or incremental, manner and that they be effectively controlled. Only by effectively and efficiently managing their IT resources through a robust investment management process can agencies gain opportunities to make better allocation decisions among many investment alternatives and further leverage their IT investments. Figure 2: Percentage of Agencies' Use of IT Investment Management Practices[A]: [See PDF for image] [A] Percentages do not add up to 100 percent due to rounding. Note: Yes--the practice was in place. Partially--the agency has some, but not all, aspects of the practice in place. Examples of circumstances in which the agency would receive this designation include when (1) some, but not all, of the elements of the practice were in place; (2) the agency documented that it has the information or process in place but it was not in the prescribed form (e.g., in a specific document as required by law or OMB); (3) the agency's documentation was in draft form; or (4) the agency had a policy related to the practice, but evidence supported that it had not been completely or consistently implemented. No--the practice was not in place. Not applicable--the practice was not relevant to the agency's particular circumstances. [End of figure] Governmentwide Progress Demonstrated, but More Work Remains: Critical aspects of IT investment management include developing well- supported proposals, establishing investment management boards, and selecting and controlling IT investments. The agencies' use of practices associated with these aspects of investment management is wide-ranging, as follows: * IT investment proposals. Various legislative requirements, an executive order, and OMB policies provide minimum standards that govern agencies' consideration of IT investments. In addition, we have issued guidance to agencies for selecting, controlling, and evaluating IT investments.[Footnote 37] Such processes help ensure, for example, that investments are cost-beneficial and meet mission needs and that the most appropriate development or acquisition approach is chosen. The agencies in our review have mixed results when evaluated against these various criteria. For example, the agencies almost always require that proposed investments demonstrate that they support the agency's business needs, are cost-beneficial, address security issues, and consider alternatives. To demonstrate, the Department of Transportation requires that proposed projects complete a business case to indicate that the project (1) will meet basic requirements in areas such as mission need, affordability, technical standards, and disabled access requirements, (2) is economically beneficial, and (3) has considered alternatives. One element in this area that agencies were not as likely to have fully in place was the Clinger-Cohen Act requirement that agencies follow, to the maximum extent practicable, a modular, or incremental, approach when investing in IT projects. Incremental investment helps to mitigate the risks inherent in large IT acquisitions/developments by breaking apart a single large project into smaller, independently useful components with known and defined relationships and dependencies. An example of such an approach is DOD's policy stating that IT acquisition decisions should be based on phased, evolutionary segments that are as brief and narrow in scope as possible and that each segment should solve a specific part of an overall mission problem and deliver a measurable net benefit independent of future segments.[Footnote 38] However, 14 agencies do not have a policy that calls for investments to be done in a modular manner. For example, although the Environmental Protection Agency (EPA) reported that it worked with program offices to try to segment work so that the scope and size of each project is manageable, it does not have a policy that calls for investments to be done in a modular manner. The absence of a policy calls into question whether EPA is implementing incremental investment in a consistent and effective manner. * Investment management boards. Our investment management guide states that establishing one or more IT investment boards is a key component of the investment management process. According to our guide, the membership of this board should include key business executives and should be responsible for final project funding decisions or should provide recommendations for the projects under its scope of authority. Such executive-level boards, made up of business-unit executives, concentrate management's attention on assessing and managing risks and regulating the trade-offs between continued funding of existing operations and developing new performance capabilities. Almost all of the agencies in our review have one or more enterprise- level investment management boards. For example, HUD's Technology Investment Board Executive Committee and supporting boards have responsibility for selecting, controlling, and evaluating the department's IT investments. HUD's contractor-performed maturity audits also have helped the department validate its board structure and its related investment management processes. However, the investment management boards for six agencies are not involved, or the agency did not document the board's involvement, in the control phase. For example, the National Science Foundation (NSF) has a CIO advisory group that addresses only the select phase of the IT investment management process. NSF's CIO explained that the agency reviews the progress of its major information system projects through other means, such as meetings with management. In providing comments on a draft of this report, the CIO stated that he believes that NSF has a comprehensive set of management processes and review structures to select, control, and evaluate IT investments and cited various groups and committees used as part of this process. However, NSF's summary of its investment management process and memo establishing the CIO advisory group include only general statements related to the oversight of IT investments, and NSF provided no additional documentation demonstrating that its investment management board plays a role in the control and evaluation phases. Our investment management guidance identifies having an IT investment management board(s) be responsible for project oversight as a critical process. Maintaining responsibility for oversight with the same body that selected the investment is crucial to fostering a culture of accountability by holding the investment board that initially selected an investment responsible for its ongoing success. In addition, 17 agencies do not fully address the practice that calls for processes to be in place that address the coordination and alignment of multiple investment review boards. For example, we recently reported that the Department of the Interior has established three department-level IT investment boards and begun to take steps to ensure that investment boards are established at the bureau level.[Footnote 39] However, at the time of our review, the department (1) could not assert that department-level board members exhibited core competencies in using Interior's IT investment approach and (2) had limited ability to oversee investments in its bureaus. We made recommendations to Interior to strengthen both the activities of the department-level boards and the department's ability to oversee investment management activities at the bureaus. * Selection of IT investments. During the selection phase of an IT investment management process, the organization (1) selects projects that will best support its mission needs and (2) identifies and analyzes each project's risks and returns before committing significant funds. To achieve desired results, it is important that agencies have a selection process that, for example, uses selection criteria to choose the IT investments that best support the organization's mission and prioritizes proposals. Twenty-two agencies use selection criteria in choosing their IT investments. In addition, about half the agencies use scoring models[Footnote 40] to help choose their investments. For example, the working group and CIO office officials that support the Department of Education's investment review board used a scoring model as part of deciding which IT investments to recommend for the board's consideration and approval. This model contained two main categories of criteria: (1) value criteria that measured the impact and significance of the initiative, given project goals and the strategic objectives of the department; and (2) health criteria that measured the potential for the success of the initiative and helped to assess both the performance and the associated risks that are involved in project and contract management. In the case of DOD, in February 2003 we reported that it had established some, and was establishing other IT investment criteria, but these criteria had not been finalized.[Footnote 41] Accordingly, we recommended, and DOD concurred, that DOD establish a standard set of criteria. In September we reported that this recommendation had not been implemented.[Footnote 42] DOD officials stated that the department was developing the criteria but that the proposed governance structure had not yet been adopted. * Control over IT investments. During the control phase of the IT investment management process, the organization ensures that, as projects develop and as funds are spent, the project is continuing to meet mission needs at the expected levels of cost and risk. If the project is not meeting expectations or if problems have arisen, steps are quickly taken to address the deficiencies. Executive level oversight of project-level management activities provides the organization with increased assurance that each investment will achieve the desired cost, benefit, and schedule results. Although no agencies had the practices associated with the control phase fully in place, some have implemented important aspects of this phase. For example, Labor requires project managers to prepare a control status report based on a review schedule established during the selection phase, which is reviewed by the Office of the CIO and its technical review board as part of determining whether to continue, modify, or cancel the initiative.[Footnote 43] For initiatives meeting certain criteria, the technical review board makes recommendations to the management council, which serves as the department's top tier executive investment review council, is chaired by the Assistant Secretary of Administration and Management, and consists of component agency heads. Nevertheless, in general, the agencies are weaker in the practices pertaining to the control phase of the investment management process than in the selection phase. In particular, the agencies did not always have important mechanisms in place for agencywide investment management boards to effectively control investments, including decision-making rules for project oversight, early warning mechanisms, and/or requirements that corrective actions for under-performing projects be agreed upon and tracked. For example, the Department of the Treasury does not have a department-level control process; instead, each bureau may conduct its own reviews that address the performance of its IT investments and corrective actions for under-performing projects. In a multitiered organization like Treasury, the department is responsible for providing leadership and oversight for foundational critical processes by ensuring that written policies and procedures are established, repositories of information are created that support IT investment decision making, resources are allocated, responsibilities are assigned, and all of the activities are properly carried out where they may be most effectively executed. In such an organization, the CIO is specifically responsible for ensuring that the organization is effectively managing its IT investments at every level. Treasury IT officials recognize the department's weaknesses in this area and informed us that they are working on developing a new capital planning and investment control process that is expected to address these weaknesses. Similarly, the Department of Energy is planning on implementing the investment control process outlined in its September 2003 capital planning and investment control guide in fiscal year 2004, which addresses important elements such as corrective action plans. However, this guide does not document the role of Energy's investment management boards in this process. Practice-Specific Analysis: Table 2 provides additional detail on each investment management practice and our evaluation of whether each agency had the practice in place. The table indicates those practices in which improvement is needed as well as which agencies reported that they were taking, or planned to take, actions to address the practices or elements of practices. Table 2: IT Investment Management Practices[A]: Practice 2.1: The agency has a documented IT investment management process that, at a minimum; * specifies the roles of key people (including the CIO) and groups within the IT investment management process; * outlines significant events and decision points; * identifies external and environmental factors that influence the process; * explains how the IT investment management process is coordinated with other organizational plans and processes, and; * describes the relationship between the investment management process and the agency's enterprise architecture. Results; Yes: 12; Partially: 14; No: 0; NA: 0. Comments: * Yes--Commerce, Education, Energy, GSA, HUD, Interior, Justice, OPM, SBA, SSA, State, and USAID have this practice in place; * Partially--Agriculture and Labor do not describe the relationship between their investment management processes, and their enterprise architectures in their IT capital planning and investment control guide. Air Force, EPA, and VA documentation related to this practice is in draft form. In addition, Air Force's[C] draft portfolio management document does not identify external and environmental factors or describe the relationship between the investment management process and the enterprise architecture. DOD[C] is piloting a draft IT portfolio management policy, but this policy does not address how this process relates to its other organizational plans and processes and its enterprise architecture or identify external and environmental factors. HHS[C] does not address how this process relates to its other organizational plans and processes and its enterprise architecture or identify external and environmental factors. NRC's current and draft capital planning and investment control policies do not address how this process relates to its other organizational plans and processes and its enterprise architecture or identify external and environmental factors. Army and NASA's[C] investment management policies and guidance do not describe the relationship of this process to its enterprise architecture. Navy[C] recognizes the need to clarify roles and responsibilities related to IT investment management, and its IT capital planning guide does not identify external and environmental factors. NSF does not have an IT investment management guide, and its summary of its policy does not address how this process relates to its other organizational plans and processes and its enterprise architecture or identify external and environmental factors. Transportation reported that there was little integration between its capital planning and investment control process and the budget. Treasury[C] does not have a capital planning and investment control guide, and its documentation supporting this practice is in draft form. Practice 2.2: The agency established one or more agencywide IT investment management boards responsible for selecting, controlling, and evaluating IT investments that, at a minimum; * have final project funding decision authority (or provide recommendations) over projects within their scope of authority, and; * are composed of key business unit executives. Results; Yes: 14; Partially: 10; No: 2; NA: 0. Comments: * Yes--Agriculture, Commerce, Education, GSA, HHS, HUD, Interior, Labor, OPM, SBA, SSA, State, Transportation, and VA have this practice in place; * Partially--Treasury[C] and USAID have not completely implemented this practice. Air Force, Army[C], Energy, NASA, NRC, and NSF's IT investment management boards are not responsible for controlling and evaluating IT investments, or this role has not been fully documented. EPA's documentation in support of this practice is in draft form. Navy's[C] IT investment management board governance process is not completely implemented; * No--DOD[B,C] does not have this practice in place. Justice[C] reported that it is piloting an IT investment management board, but did not provide documentation on the responsibilities, processes, or makeup of this board. Practice 2.3: The agencywide board(s) work processes and decision- making processes are described and documented. Results; Yes: 9; Partially: 6; No: 11; NA: 0. Comments: * Yes--Agriculture, Commerce, Education, HUD, Interior, Labor, SBA, State, and Transportation have this practice in place; * Partially--Army has not consistently implemented this practice. GSA did not have policies and procedures for each of its IT investment management boards. HHS has not established procedures for the development, documentation, and review of IT investments. EPA and VA's documentation related to this practice is in draft form. USAID has not completely implemented this practice; * No--Air Force, DOD[B,C], Energy, Justice[C], NASA, Navy[C], NRC, NSF, OPM, SSA[C], and Treasury[C] do not have this practice in place. Practice 2.4: If more than one IT investment management board exists in the organization (e.g., at the component level), the organization has; * documented policies and procedures that describe the processes for aligning and coordinating IT investment decision making; * criteria for determining where in the organization different types of IT investment decisions are made, and; * processes that describe how cross-functional investments and decisions (e.g., common applications) are handled. Results; Yes: 2; Partially: 10; No: 7; NA: 7. Comments: * Yes--GSA and Labor have this practice in place; * Partially--Agriculture does not have documented policies and processes for aligning and coordinating IT investment decision making or processes for describing how cross-functional investments and decisions are made. Air Force, Army, Commerce, Education, HHS[C], and Transportation[C] do not have documented policies and procedures for aligning and coordinating investment decision making among their investment management boards. Interior[B] has not fully implemented its governance process for aligning and coordinating its IT investment decision making. OPM did not describe its criteria for determining major systems or describe how cross-functional investments and decisions are handled. SBA did not address whether its enterprisewide board can invoke final decision-making authority over its program office boards; * No--DOD,[B,C] Energy, NASA, Navy[C], Treasury[C], and VA[C] do not have this practice in place. Justice[C] reported that it is piloting an IT investment management board but did not provide supporting documentation; * NA--EPA, HUD, NRC, NSF, SSA, State, and USAID do not have multiple IT investment management boards. Practice 2.5: As part of its investment management process, the agency has available an annually updated comprehensive inventory of its major information systems that includes major national security systems and interfaces. Results; Yes: 21; Partially: 5; No: 0; NA: 0. Comments: * Yes--Agriculture, Air Force, Commerce, Education, EPA, GSA, HUD, Interior, Justice, Labor, Navy, NRC, NSF, OPM, SBA, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * Partially--Army's[C] inventory is not complete and does not include interfaces. A DOD inspector general report stated that DOD's inventory may not capture the universe of current DOD business management systems. Energy and NASA's inventories do not include interfaces. HHS reported that its Exhibit 300s fulfill the requirements of this practice but did not provide supporting documentation. Practice 2.6: A standard, documented procedure is used so that developing and maintaining the inventory is a repeatable event, which produces inventory data that are timely, sufficient, complete, and compatible. Results; Yes: 21; Partially: 1; No: 4; NA: 0. Comments: * Yes--Agriculture, Air Force, Commerce, DOD, Education, EPA, GSA, HUD, Interior, Justice, Labor, Navy, NSF, OPM, SBA, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * Partially--Army's documentation is in draft form; * No--Energy[C], HHS, NASA, and NRC do not have this practice in place. Practice 2.7: The IT asset inventory is used as part of managerial decision making. Results; Yes: 12; Partially: 11; No: 3; NA: 0. Comments: * Yes--Agriculture, Army, Commerce, Education, GSA, HUD, Labor, Navy, SSA, State, Transportation, and VA have this practice in place; * Partially--DOD, Energy, EPA, Interior,[B] NRC[C], NSF, OPM, SBA, and USAID do not explicitly document how their IT asset inventory is used to identify asset duplication. Air Force reported that its inventory is not being consistently used to identify asset duplication. Justice[C] reported that it has begun to use its IT asset inventory to identify asset duplication as part of a pilot of its new IT investment management process; * No--HHS, NASA, and Treasury[C] do not have this practice in place. Practice 2.8: Proposed IT investments are required to document that they have addressed the following items during project planning:; * that the project supports the organization's business and mission needs and meets users' needs; * whether the function should be performed by the public or private sector; * whether the function or project should be performed or is being performed by another agency; * that alternatives have been considered, and; * how security will be addressed. Results; Yes: 25; Partially: 1; No: 0; NA: 0. Comments: * Yes--Agriculture, Air Force, Army, Commerce, DOD, Education, Energy, EPA, GSA, HUD, Interior, Justice, Labor, NASA, Navy, NRC, NSF, OPM, SBA, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * Partially--the HHS policy addressing the element related to whether the function or project should be performed by the private sector or another government agency is in draft form. This information is normally contained in the Exhibit 300s, but HHS did not provide us with this documentation. Practice 2.9: In considering a proposed IT project, the agency requires that the project demonstrate that it is economically beneficial through the development of a business case that at least addresses costs, benefits, schedule, and risks. Results; Yes: 25; Partially: 1; No: 0; NA: 0. Comments: * Yes--Agriculture, Air Force, Army, Commerce, Education, Energy, EPA, GSA, HHS, HUD, Interior, Justice, Labor, NASA, Navy, NRC, NSF, OPM, SBA, SSA, State, Transportation, Treasury, USAID, and VA have this practice in place; * Partially--DOD has not consistently implemented this practice. Practice 2.10: In considering a proposed IT project, the agency requires that the project demonstrate that it is consistent with federal and agency enterprise architectures. Results; Yes: 20; Partially: 6; No: 0; NA: 0. Comments: * Yes--Air Force, Army, Commerce, Education, Energy, EPA, GSA, HHS, HUD, Interior, Labor, Navy, NRC, NSF, OPM, SSA, State, Transportation, Treasury, and VA have this practice in place; * Partially--the agencies are required to include how major IT investments align with the agency's enterprise architecture and the federal enterprise architecture in their budget Exhibit 300s. However, the following agencies do not have policies requiring compliance with the agency enterprise architectures, which is a core element in our Enterprise Architecture Management Framework. Agriculture and NASA[B,C] do not have approved policies related to this practice but require compliance as part of their IT investment management reviews for the fiscal year 2005 budget cycle. Justice and USAID do not have a policy requiring that IT projects comply with their enterprise architecture. SBA's policy requiring compliance with its enterprise architecture is in draft form. DOD does not have a policy requiring compliance with its business enterprise architecture. Practice 2.11: The agency requires that the proposed IT investment, at a minimum; * support work processes that it has simplified or redesigned to reduce costs and improve effectiveness, and; * make maximum use of commercial-off-the-shelf (COTS) software. Results; Yes: 11; Partially: 7; No: 8; NA: 0: Comments: * Yes--Air Force, Army, DOD, GSA, Justice, Labor, NASA, Navy, NSF, SSA, and VA have this practice in place; * Partially--Education, HHS, Interior, and SBA do not require that proposed IT investments support work processes that have been simplified or redesigned. NRC has policies related to this practice but reported that they have not been fully integrated into its investment decision making. Energy's business case guidelines address this practice, but Energy reported that consideration of these factors is not required for selection and approval. EPA's policy related to COTS is in draft form; * No--Agriculture[C], Commerce, HUD, OPM, State, Transportation, Treasury[C], and USAID do not have this practice in place. Practice 2.12: The agency has established project selection criteria distributed throughout the organization that include, at a minimum; * cost, benefit, schedule, and risk elements; * measures such as net benefits, net risks, and risk-adjusted return on investment; and; * qualitative criteria for comparing and prioritizing alternative information systems investment projects. Results; Yes: 6; Partially: 16; No: 4; NA: 0: Comments: * Yes--Agriculture, GSA, Energy, NASA, Transportation, and VA have this practice in place; * Partially--Commerce[C], Education, HUD[C], Justice, Labor, Navy, SBA[B], State, and USAID have project selection criteria that do not include net risk and risk-adjusted return on investment. DOD[B] has established some IT investment criteria, but these criteria are not finalized or part of an investment review process. EPA has project selection criteria that do not include net risks, risk-adjusted return on investment, or qualitative criteria. EPA's documentation in support of this practice is also in draft form. Interior's project selection criteria do not include cost and schedule. Air Force[C] and Army's[C] project selection criteria do not include cost, benefit, schedule, and risk elements or measures such as net benefits, net risks, and risk- adjusted return on investment. OPM has not consistently implemented this practice. SSA's[B] criteria is high-level and not explicit; * No--HHS[C], NRC, NSF, and Treasury[C] do not have this practice in place. Practice 2.13: The agency has established a structured selection process that, at a minimum; * selects IT proposals using selection criteria; * identifies and addresses possible IT investments and proposals that are conflicting, overlapping, strategically unlinked, or redundant; * prioritizes proposals; and; * is integrated with budget, financial, and program management decisions. Results; Yes: 8; Partially: 15; No: 3; NA: 0. Comments: * Yes--Agriculture, Commerce, Education, GSA, HUD, Labor, SBA, and State have this practice in place; * Partially--Air Force's[C] documentation in support of this practice is in draft form and does not include prioritizing proposals across the enterprise or the use of a scoring model. Army's prioritized list is limited to investments to address capability shortfalls. DOD[C] is piloting a draft IT portfolio management policy that includes a selection process. EPA's documentation of its selection processes is in draft form. Energy, Interior, and Transportation do not prioritize their IT proposals. Justice[C] does not use a scoring model or prioritize or rank its IT proposals. NASA[C] does not have a process for identifying possible conflicting, overlapping, strategically unlinked, or redundant proposals; does not use a scoring model; and does not prioritize or rank its IT proposals. Navy generally does not use its IT investment management boards outlined in its governance process as part of its IT investment selection process and does not use a scoring model or prioritize or rank its IT proposals. NRC does not select IT proposals using selection criteria, prioritize proposals, or document how its selection process is integrated with budget, financial, and program management decisions. OPM has not consistently implemented this practice. SSA[B] does not use a scoring model. USAID[C] does not have a process for identifying possible conflicting, overlapping, strategically unlinked, or redundant proposals. VA does not have a process to identify and address possible conflicting, overlapping, strategically unlinked, or redundant IT investments and does not prioritize IT proposals for selection; * No--HHS[C], Treasury[C], and NSF do not have this practice in place. Practice 2.14: Agency policy calls for investments to be modularized (e.g., managed and procured in well-defined useful segments or modules that are short in duration and small in scope) to the maximum extent achievable. Results; Yes: 9; Partially: 3; No: 14; NA: 0. Comments: * Yes--Air Force, Army, Education, Justice, NASA, Navy, NRC, SBA, and VA have this practice in place; * Partially--DOD had not consistently implemented this practice. HHS and NSF's documentation supporting this practice is in draft form; * No--Agriculture, Commerce[C], Energy, EPA, GSA[C], HUD, Interior, Labor, OPM, SSA, State, Transportation[C], Treasury, and USAID do not have this practice in place. Practice 2.15: The agencywide investment management board(s) has written policies and procedures for management oversight of IT projects that cover, at a minimum; * decision-making rules for project oversight that allow for terminating projects, when appropriate; * current project data, including expected and actual cost, schedule, and performance data, to be provided to senior management periodically and at major milestones; * criteria or thresholds related to deviations in cost, schedule, or system capability actuals versus expected project performance; and; * the generation of an action plan to address a project's problem(s) and track resolution. Results; Yes: 0; Partially: 20; No: 6; NA: 0. Comments: * Partially--Agriculture[C] reported that it has not implemented the corrective action plan element in a consistent manner. Air Force[C], NASA[C], and SSA[C] have control processes but do not explicitly document the role, responsibility, and authority of their enterprisewide IT investment management boards in the control phase. Army[C], DOD[B,C], and Navy's control processes do not involve enterprisewide IT investment management boards. Commerce[C] does not have decision-making rules to guide oversight of IT investments and projects are not required to submit reports of deviations in system capability. Education has not consistently required corrective actions or tracked corrective actions related to control phase reviews. GSA does not have clear decision-making rules, require projects to report on deviations in system capability, or require that corrective actions be tracked to resolution. HHS[C] does not have decision-making rules to guide oversight of IT investments, review projects at major milestones, or systematically track corrective actions. HUD[C] does not require reports of deviations of system capability or monitor projects at key milestones. Interior[B] does not have decision-making rules for oversight of IT investments, require reports of deviations of system capability, or require corrective action plans. Justice[B,C], reported that it is piloting an IT investment management board that includes the control phase but has not provided documentation supporting that all of the practice elements are addressed. Labor and Transportation have evaluation criteria to assess investments during the control phase, but do not have decision-making rules to guide their investment management boards' decisions. OPM has not consistently implemented this practice. State's draft documentation does not require projects to be reviewed at key milestones. USAID[C] does not have decision-making rules, require reports on deviations in system capability, and review projects at major milestones, and its policy for requiring action plans is in draft form. VA's[C] policies and procedures on decision-making rules, criteria or thresholds for system capability, and the generation of action plans have not been fully documented; * No--SBA[B,C] and Treasury[C] do not have this practice in place. Energy plans to implement a control process in fiscal year 2004, but its new capital planning and investment review guide does not address the role of its investment management boards in the process. EPA[C] is implementing its control process in fiscal year 2004. NRC's current and draft capital planning and investment control documentation do not address the elements of this practice and do not explicitly document the role, responsibility, and authority of its enterprisewide IT investment management board in this process. NSF's investment management board is not responsible for the control process. NSF reported that it uses other mechanisms to implement this practice but provided no supporting documentation. Practice 2.16: The agencywide investment management board(s) established an oversight mechanism of funded investments that, at a minimum; * determines whether mission requirements have changed; * determines whether the investment continues to fulfill ongoing and anticipated mission requirements; * determines whether the investment is proceeding in a timely manner toward agreed-upon milestones; * employs early warning mechanisms that enable it to take corrective action at the first sign of cost, schedule, or performance slippages; and; * includes the use of independent verification and validation (IV&V) reviews of under-performing projects, where appropriate. Results; Yes: 2; Partially: 19; No: 5; NA: 0. Comments: * Yes--GSA and VA have this practice in place; * Partially--Agriculture[C] reported that its oversight of IT investments has not been consistently implemented. Air Force[C], NASA[C], and SSA[C] have control processes but did not explicitly document the role, responsibility, and authority of their enterprisewide IT investment management boards in this process. Army[C], DOD[B,C], and Navy's control processes do not involve enterprisewide IT investment management boards. Commerce and Labor do not employ an early warning mechanism. State[C] has procedures for control phase reviews, but they are not fully implemented. Education, HHS, and HUD do not have a process for using IV&V reviews. Interior[B,C], does not have a process to determine whether investments are proceeding in a timely manner toward agreed-upon milestones, employ an early warning mechanism, or use IV&V reviews. Justice[B,C], reported that it is piloting an IT investment management board that includes the control phase but did not provide documentation supporting that all of the practice elements are addressed. OPM has not consistently implemented this practice. SBA[B] did not provide evidence that it had implemented all of the oversight mechanisms in its investment management guide and did not use IV&V reviews. Transportation and USAID do not employ an early warning system or have a process for using IV&V reviews; * No--Treasury[C] does not have this practice in place. Energy plans to implement a control process in fiscal year 2004, but its new capital planning and investment review guide does not address the role of its investment management boards in the process. EPA[C] is implementing its control process in fiscal year 2004. NRC's current and draft capital planning and investment control documentation does not address the elements of this practice and does not explicitly document the role, responsibility, and authority of its enterprisewide IT investment management board in this process. NSF's investment management board is not responsible for the control process. NSF reported that it uses other mechanisms to implement this practice but provided no supporting documentation. Practice 2.17: Corrective actions for under-performing projects are agreed upon, documented, and tracked by the agencywide investment management board(s). Results; Yes: 5; Partially: 12; No: 9; NA: 0. Comments: * Yes--Commerce, HUD, Labor, Transportation, and VA have this practice in place; * Partially--Agriculture[C] and SBA[B] reported that they have not consistently implemented this practice. Air Force[C], NASA[C], and SSA have control processes but did not explicitly document the role, responsibility, and authority of their enterprisewide IT investment management boards in this process. SSA[C] also did not provide support that it was tracking corrective actions. Army[C], DOD[B], and Navy's control processes do not involve enterprisewide IT investment management boards. Education has not consistently required corrective actions or tracked corrective actions related to control phase reviews. GSA and HHS[C] do not systematically track corrective actions. State[C] has procedures for control phase reviews, but they are not fully implemented; * No--Interior[C], Justice, OPM, Treasury[C], and USAID do not have this practice in place. Energy plans to implement a control process in fiscal year 2004, but its new capital planning and investment review guide does not address the role of its investment management boards in the process. EPA[C] is implementing its control process in fiscal year 2004. NRC's current and draft capital planning and investment control documentation does not address the elements of this practice and does not explicitly document the role, responsibility, and authority of its enterprisewide IT investment management board in this process. NSF's investment management board is not responsible for the control process. NSF reported that it uses other mechanisms to implement this practice, but provided no supporting documentation. Practice 2.18: The agencywide investment management board(s) requires that postimplementation reviews be conducted to; * validate expected benefits and costs, and; * document and disseminate lessons learned. Results; Yes: 6; Partially: 17; No: 3; NA: 3. Comments: * Yes--Agriculture, GSA, HUD, Labor, OPM, and VA have this practice in place; * Partially--Army, DOD, NASA[C], Navy, NRC, NSF, and SSA's[C] evaluation processes do not involve an enterprisewide IT investment management board. NSF also does not define what is to be included in a postimplementation review and SSA[B] reported that such reviews are not done regularly. Commerce[C] reported that postimplementation reviews have not been consistently completed and are not required to be reported to its investment management board. Air Force's[C] documentation in support of this practice is in draft form and does not document the role of its IT investment management boards in this process. Education[C] reported that postimplementation reviews were not always performed. Energy[C], Justice[B], Transportation[C], and USAID have a policy related to this practice, but it has not been implemented. Also, Energy's processes do not involve an enterprisewide IT investment management board. HHS, SBA[B,C], and State[C] have a policy related to this practice but did not provide evidence that it has been completely implemented. In addition, HHS's policy does not specifically address validating expected benefits and costs; * No--EPA[C] is implementing its evaluation process in fiscal year 2004. Interior[C] and Treasury[C] do not have this practice in place. Source: GAO. [A] Due to its recent establishment, we did not include DHS as a part of this analysis. [B] We have an outstanding recommendation related to this practice. [C] The agency reported that it was taking, or planned to take, action to address this practice, or elements of the practice. Note: Yes--the practice was in place. Partially--the agency has some, but not all, aspects of the practice in place. Examples of circumstances in which the agency would receive this designation include when (1) some, but not all, of the elements of the practice were in place; (2) the agency documented that it has the information or process in place but it was not in the prescribed form (e.g., in a specific document as required by law or OMB); (3) the agency's documentation was in draft form; or (4) the agency had a policy related to the practice, but evidence supported that it had not been completely or consistently implemented. No--the practice was not in place. Not applicable--the practice was not relevant to the agency's particular circumstances. [End of table] Among the variety of reasons cited for practices not being fully in place were that the CIO position had been vacant, that not including a requirement in the IT investment management guide was an oversight, and that the process was being revised. However, in some cases the agencies could not identify why certain practices were not in place. Regarding DHS, although we did not include the department in our assessment or table 2, the department has investment management processes that it has put in place or is in the process of putting in place. Conclusions: Federal agencies did not always have in place important practices associated with IT laws, policies, and guidance. At the governmentwide level, agencies generally have IT strategic plans or information resources management (IRM) plans that address IT elements, such as security and enterprise architecture, but do not cover other aspects of IRM that are part of the Paperwork Reduction Act, such as information collection, records management, and privacy. This may be attributed, in part, to OMB not establishing comprehensive guidance for the agencies detailing the elements that should be included in such a plan. There were also numerous instances of individual agencies that do not have specific IT strategic planning, performance measurement, or investment management practices fully in place. Agencies cited a variety of reasons for not having these practices in place, such as that the CIO position had been vacant, not including a requirement in guidance was an oversight, or that the process was being revised. Nevertheless, not only are these practices based on law, executive orders, OMB policies, and our guidance, but they are also important ingredients for ensuring effective strategic planning, performance measurement, and investment management, which, in turn, make it more likely that the billions of dollars in government IT investments will be wisely spent. Accordingly, we believe that it is important that they be expeditiously implemented by individual agencies. Recommendations: To help agencies in developing strategic IRM plans that fully comply with the Paperwork Reduction Act of 1995, we recommend that the Director, OMB, develop and disseminate to agencies guidance on developing such plans. At a minimum, such guidance should address all elements of IRM, as defined by the Paperwork Reduction Act. As part of this guidance, OMB should also consider the most effective means for agencies to communicate information about any major IT acquisition program(s) or phase or increment of that program that significantly deviated from cost, performance, or schedule goals established by the program. One option for communicating this information, for example, could be through the annual agency performance reports that are required by the Government Performance and Results Act. We are also generally making recommendations to the agencies in our review regarding those practices that are not fully in place unless, for example, (1) we have outstanding recommendations related to the practice, (2) the agency has a draft document addressing the practice, or (3) implementation of the practice was ongoing. Appendix I contains these recommendations. Agency Comments and Our Evaluation: We received written or oral comments on a draft of this report from OMB and 25 of the agencies in our review.[Footnote 44] We also requested comments from the Department of Homeland Security and the Office of Personnel Management, but none were provided. Regarding OMB, in oral comments on a draft of this report, representatives from OMB's Office of Information and Regulatory Affairs and Office of the General Counsel questioned the need for additional IRM plan guidance because they do not want to be prescriptive in terms of what agencies include in their plans. We continue to believe that agencies need additional guidance from OMB on the development and content of their IRM plans because OMB Circular A-130 does not provide overall guidance on the contents of agency IRM plans and half the agencies indicated a need for OMB to provide additional guidance on the development and content of IRM plans. Further, additional guidance would help to ensure that agency plans address all elements of IRM, as defined by the Paperwork Reduction Act. A strategic IRM plan that communicates a clear and comprehensive vision for how the agency will use information resources to improve agency performance is important because IRM encompasses virtually all aspects of an agency's information activities. In commenting on a draft of the report, most of the agencies in our review generally agreed with our findings and recommendations. The agencies' specific comments are as follows: * Agriculture's CIO stated that the department concurred with the findings in this report and provided information on action it was taking, or planned to take, to implement the recommendations. Agriculture's written comments are reproduced in appendix II. * The Secretary of Commerce concurred with the recommendations in this report and stated that, in response, the department is updating its policies and procedures. Commerce's written comments are reproduced in appendix III. * DOD's Deputy Assistant Secretary of Defense (Deputy CIO) stated that the department concurred or partially concurred with the recommendations in this report. DOD also provided additional documentation and information on actions that it is taking, or planned to take, to address these recommendations. We modified our report based on these comments and documentation, as appropriate. DOD's written comments, along with our responses, are reproduced in appendix IV. * Education's Assistant Secretary for Management/CIO stated that the agency generally agreed with our assessment of the department's use of IT strategic planning/performance measurement and investment management practices. Education provided additional comments and documentation related to two of our practices. We modified our report on the basis of these comments and documentation, as appropriate. Education's written comments, along with our responses, are reproduced in appendix V. * Energy's Director of Architecture and Standards provided e-mail comments stating that the department believes that GAO fairly depicted where the department currently stands in the IT investment management process. The director also provided other comments that were technical in nature and that we addressed, as appropriate. * EPA's Assistant Administrator/CIO generally agreed with our findings and recommendations on the need to complete work currently under way to formalize the documentation of IT management practices. However, EPA questioned our characterization of the agency's IT management and strategic planning and provided other comments, which we addressed, as appropriate. EPA's written comments, along with our responses, are reproduced in appendix VI. * GSA's CIO stated that the agency generally agreed with the findings and recommendations in the report. GSA provided suggested changes and additional information and documentation related to nine of our practices and two recommendations. We modified our report on the basis of these comments and documentation, as appropriate. GSA's written comments, along with our responses, are reproduced in appendix VII. * HHS's Acting Principal Deputy Inspector General stated that the department concurred with the findings and recommendations of the report. HHS's written comments are reproduced in appendix VIII. * HUD's Assistant Secretary for Administration/CIO stated that the department was in agreement with the recommendations in this report. HUD's written comments are reproduced in appendix IX. * Interior's Acting Assistant Secretary for Policy, Management and Budget stated that the recommendations in our report would further improve the department's IT investment management. Interior's written comments are reproduced in appendix X. * Justice's CIO stated that, overall, the department concurred with the findings and recommendations in this report, noting that our recommendations will assist in further defining IT strategic planning, performance measurement, and investment management practices. Justice's written comments, along with our response, are reproduced in appendix XI. * Labor's Assistant Secretary for Administration and Management/CIO reported that the department generally concurred with this report and provided suggested changes in two areas, which we addressed, as appropriate. Labor's written comments, along with our responses, are reproduced in appendix XII. * NASA's Deputy Administrator reported that the agency generally concurred with the recommendations in this report and provided additional information on actions that it is taking, or planned to take, to address these recommendations. NASA's written comments, along with our response, are reproduced in appendix XIII. * NSF's CIO provided e-mail comments disagreeing with three areas of this report. First, NSF did not agree with our assessment of practice 1.1, stating that the agency has a comprehensive agency-level planning framework that includes a suite of planning documents and internal and external oversight activities that it believes addresses IT planning requirements. However, our review of the planning documents cited by NSF in its self-assessment found that it did not address the elements of the practice. In particular, the agency did not describe the responsibility and accountability for IT resources or the method that it uses to define program information needs and how such needs will be met. Moreover, in our exit conference with NSF officials, the CIO indicated agreement with our assessment. Since NSF provided no additional documentation, we did not modify the report. Second, the CIO disagreed with our characterization of the agency's enterprisewide investment management board. We modified the report to reflect the CIO's comments; however, we did not change our overall assessment of the role of the board because NSF's summary of its investment management process and memo establishing the CIO advisory group include only general statements related to the oversight of IT investments, and NSF provided no additional documentation demonstrating that its investment management board plays a role in the control and evaluation phases. Third, the CIO stated that NSF has established processes, management, and oversight controls over IT investments. However, NSF provided limited documentation on the control phase of its investment management process. In particular, NSF's summary of its investment management process and memo establishing the CIO advisory group include only general statements related to the oversight of IT investments, and NSF provided no additional documentation demonstrating that its investment management board plays a role in the control and evaluation phases. Accordingly, we did not modify the report. * NRC's Executive Director for Operations stated that this report provides useful information and agreed that the practices are important for ensuring effective use of government IT investments but had no specific comments. NRC's written comments are reproduced in appendix XIV. * SBA's GAO liaison provided e-mail comments questioning the need to have its enterprise investment management board have final decision- making authority over IT investments. Our IT investment management guidance states that enterprise-level IT investment boards be capable of reviewing lower-level board actions and invoking final decision- making authority over all IT investments.[Footnote 45] In particular, if disputes or disagreements arise over decision-making jurisdiction about a specific IT investment project, the enterprise board must be able to resolve the issue. Accordingly, we did not modify the report. SBA also provided technical comments that we incorporated, as appropriate. * SSA's Commissioner generally agreed with the recommendations in the report and provided comments on each recommendation that we addressed, as appropriate. SSA's written comments, along with our responses, are reproduced in appendix XV. * State's Assistant Secretary/Chief Financial Officer stated that the findings in the report are consistent with discussions held with its IT staff and provided additional information on four practices. On the basis of this additional information, we modified our report, as appropriate. State's written comments, along with our response, are reproduced in appendix XVI. * A program analyst in the Department of Transportation's Office of the CIO provided oral comments that were technical in nature that we addressed, as appropriate. * The Acting Director, Budget and Administrative Management in Treasury's Office of the CIO, provided oral comments stating that the department concurred with our findings and recommendations. The official further stated that the department recognized its shortcomings and was working to correct them. * USAID's Assistant Administrator, Bureau for Management, did not address whether the agency agreed or disagreed with our overall findings or recommendations but commented on our evaluation of two practices, which we addressed, as appropriate. USAID's written comments, along with our response, are reproduced in appendix XVII. * The Secretary of VA stated that the department concurred with the recommendations in the report and provided comments on actions that it has taken, or planned to take, in response. We modified the report based on these comments, as appropriate. VA's written comments, along with our responses, are reproduced in appendix XVIII. As agreed with your offices, unless you publicly announce the contents of this report earlier, we plan no further distribution until 30 days from the report date. At that time, we will send copies of this report to the secretaries of the Departments of Agriculture, the Air Force, the Army, Commerce, Defense, Education, Energy, Health and Human Services, Homeland Security, Housing and Urban Development, the Interior, Justice, Labor, the Navy, State, Transportation, the Treasury, and Veterans Affairs; the administrators of the Environmental Protection Agency, General Services Administration, National Aeronautics and Space Administration, Small Business Administration, and U.S. Agency for International Development; the commissioners of the Nuclear Regulatory Commission and the Social Security Administration; and the directors of the National Science Foundation, Office of Management and Budget, and Office of Personnel Management. We will also make copies available to others upon request. In addition, this report will be available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov]. If you have any questions on matters discussed in this report, please contact me at (202) 512-9286 or Linda J. Lambert, Assistant Director, at (202) 512-9556. We can also be reached by e-mail at [Hyperlink, pownerd@gao.gov] and [Hyperlink, lambertl@gao.gov], respectively. Other contacts and key contributors to this report are listed in appendix XIX. Signed by: David A. Powner: Director, Information Technology Management Issues: [End of section] Appendixes: Appendix I: Recommendations to Departments and Agencies: Agriculture: To improve the department's information technology (IT) strategic planning/performance measurement processes, we recommend that the Secretary of Agriculture take the following six actions: * document the department's IT strategic management processes and how they are integrated with other major departmental processes, such as the budget and human resources management; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by the Federal Information Security Management Act (FISMA) and include a description of major IT acquisitions contained in its capital asset plan that bear significantly on its performance goals; * implement a process for assigning roles and responsibilities for achieving the department's IT goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the department's enterprisewide IT performance measures in its information resources management (IRM) plan; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Agriculture take the following four actions: * include a description of the relationship between the IT investment management process and the department's enterprise architecture in its IT capital planning and investment control guide and require that IT investments be in compliance with the agency's enterprise architecture; * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments, including cross-cutting investments; * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of commercial-off- the-shelf (COTS) software; and: * establish a policy requiring modularized IT investments. Air Force: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of the Air Force take the following two actions: * establish a documented process for measuring progress against the department's IT goals and assign roles and responsibilities for achieving these goals; and: * develop IT performance measures related to the IT goals in the department's information strategy, including measures such as those contained in practice 1.9 in our report, and track actual-versus- expected performance. To improve the department's IT investment management processes, we recommend that the Secretary of the Air Force take the following four actions: * include a description of the relationship between the IT investment management process and the department's enterprise architecture, and an identification of external and environmental factors in its portfolio management guide; * include costs, benefits, schedule, and risk elements as well as measures such as net benefits, net risks, and risk-adjusted return-on- investment in the department's project selection criteria; * implement a scoring model and develop a prioritized list of IT investments as part of its project selection process; and: * document the role, responsibility, and authority of its IT investment management boards, including work processes, alignment, and coordination of decision making among its various boards, and document processes for controlling and evaluating IT investments, such as those outlined in practices 2.15, 2.16, 2.17, and 2.18. Army: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of the Army take the following action: * complete the development of IT performance measures related to the Army's enterprisewide IT goals, including measures such as those in practice 1.9 in our report, and track actual-versus-expected performance. To improve the department's IT investment management processes, we recommend that the Secretary of the Army take the following four actions: * include a description of the relationship between the IT investment management process and the department's enterprise architecture in the department's IT capital planning and investment control guide; * document the alignment and coordination of responsibilities of its various IT investment management boards for decision making related to IT investments; * include costs, benefits, schedule, and risk elements as well as measures such as net benefits, net risks, and risk-adjusted return-on- investment in the department's project selection criteria; and: * involve the department's IT investment management boards in controlling and evaluating IT investments, including the development and documentation of oversight processes such as those in practices 2.15, 2.16, 2.17, and 2.18. Commerce: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Commerce take the following four actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * document its process of developing IT goals in support of agency needs, measuring progress against these goals, and assigning roles and responsibilities for achieving these goals; * develop performance measures related to the department's IT goals in its IRM plan, and track actual-versus-expected performance for these IT performance measures; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Commerce take the following eight actions: * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments; * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * include net risks and risk-adjusted return-on-investment in the department's project selection criteria; * establish a policy requiring modularized IT investments; * develop decision-making rules to help guide the investment management board's oversight of IT investments during the control phase; * require that reports of deviations in systems capability in a project be submitted to the IT investment management board; * develop an early warning mechanism that enables the investment management board to take corrective action at the first sign of cost, schedule, or performance slippages; and: * require postimplementation reviews be completed and the results reported to its investment management board. Defense: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Defense take the following three actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA, align its performance measures with the goals in the plan, and include a description of major IT acquisitions contained in its capital asset plan that bear significantly on its performance goals; * establish a documented process for measuring progress against the department's IT goals; * develop IT performance measures related to its IT goals, including, for example, the measures contained in practice 1.9 in our report and track actual-versus-expected performance. To improve the department's IT investment management processes, we recommend that the Secretary of Defense take the following action: * document, as part of its planned IT portfolio management process, how this process relates to other departmental processes and the department's enterprise architecture, and document the external and environmental factors that influence the process. Education: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Education take the following four actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * establish and document a process for measuring progress against the department's IT goals in its IRM plan and for assigning roles and responsibilities for achieving these goals; * develop performance measures related to how IT contributes to program productivity, the effectiveness and efficiency of agency operations, and the effectiveness of controls to prevent software piracy; and: * track actual-versus-expected performance for the department's enterprisewide IT performance measures in its IRM plan. To improve the department's IT investment management processes, we recommend that the Secretary of Education take the following five actions: * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments; * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs; * include net risks and risk-adjusted return-on-investment in the department's project selection criteria; * develop a process to use independent verification and validation reviews, when appropriate; and: * track the resolution of corrective actions for under-performing projects and report the results to the investment management board. Energy: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Energy take the following six actions: * document how its IT management operations and decisions are integrated with human resources management; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a goal related to how IT contributes to program productivity; * develop performance measures related to how IT contributes to program productivity and the effectiveness of controls to prevent software piracy; * develop and link performance measures to the department's enterprisewide goals in its IRM plan and track actual-versus-expected performance for these measures; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Energy take the following four actions: * include interfaces in its inventory of the agency's major information systems, implement a standard, documented procedure to maintain this inventory, and develop a mechanism to use the inventory as part of managerial decision making; * prioritize the department's IT proposals; * establish a policy requiring modularized IT investments; and: * document the role, responsibility, and authority of its IT investment management boards, including work processes, alignment, and coordination of decision making among its various boards, and document the processes for controlling and evaluating IT investments, such as those in practices 2.15, 2.16, 2.17, and 2.18. Environmental Protection Agency: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Administrator of the Environmental Protection Agency take the following six actions: * document the agency's IT strategic management processes and how they are integrated with other major departmental processes, such as the budget and human resources management; * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the agency's measures associated with the IT goals in its IRM plan; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Administrator of the Environmental Protection Agency take the following three actions: * include net risks, risk-adjusted return-on-investment, and qualitative criteria in the agency's project selection criteria; * establish a policy requiring modularized IT investments; and: * fully implement an IT investment management control phase, including the elements contained in practices 2.15, 2.16, and 2.17. General Services Administration: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Administrator of the General Services Administration take the following four actions: * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for each of the agency's measures associated with the IT goals in its IRM plan; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Administrator of the General Services Administration take the following four actions: * develop work processes and decision-making processes for the agency's investment management boards; * establish a policy requiring modularized IT investments; * help guide the oversight of IT investments by developing clear decision-making rules for its IT investment management board and by requiring that IT projects report on deviations in system capability; and: * track the resolution of corrective actions for under-performing projects and report the results to the investment management board. Health and Human Services: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Health and Human Services take the following six actions: * document the department's IT strategic management processes and how they are integrated with its budget processes; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA and include a description of major IT acquisitions contained in its capital asset plan that bear significantly on its performance goals; * establish a documented process for measuring progress against the department's IT goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for its enterprisewide IT performance measures in its IRM plan; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Health and Human Services take the following 10 actions: * revise the department's IT investment management policy to include (1) how this process relates to other agency processes, (2) an identification of external and environmental factors, (3) a description of the relationship between the process and the department's enterprise architecture, and (4) the use of independent verification and validation reviews, when appropriate. * develop procedures for the department's enterprisewide investment management board to document and review IT investments; * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments; * implement a standard, documented procedure to maintain the department's inventory of major information systems and develop a mechanism to use the inventory as part of managerial decision making; * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness; * implement a structured IT selection process that includes processes and criteria such as those in practices 2.12 and 2.13; * develop decision-making rules to help guide the investment management board's oversight of IT investments during the control phase; * require the investment management board to review projects at major milestones; * track the resolution of corrective actions for under-performing projects and report the results to the investment management board; and: * revise the department's investment management policy to require postimplementation reviews to address validating benefits and costs, and conduct such reviews. Housing and Urban Development: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Housing and Urban Development take the following six actions: * document the roles and responsibilities of the chief financial officer and program managers in IT strategic planning and how the department's IT management operations and decisions are integrated with human resources management; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to how IT contributes to program productivity and the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the department's enterprisewide IT performance measures in its IRM plan; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Housing and Urban Development take the following five actions: * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * include net risks and risk-adjusted return-on-investment in the department's project selection criteria; * establish a policy requiring modularized IT investments; * require IT projects to report on deviations in system capability and monitor IT projects at key milestones; and: * develop a process to use independent verification and validation reviews, when appropriate. Interior: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of the Interior take the following six actions: * document the department's IT strategic management processes and how they are integrated with other major departmental processes, including organizational planning, budget, financial management, human resources management, and program decisions; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA and include a description of major IT acquisitions contained in its capital asset plan that bear significantly on its performance goals; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the department's enterprisewide IT performance measures in its IRM plan; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of the Interior take the following five actions: * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness; * include cost and schedule in the department's project selection criteria and prioritize its IT proposals; * establish a policy requiring modularized IT investments; * require that corrective actions be undertaken, tracked, and reported to the investment management board for under-performing projects; and: * implement an evaluation process for IT investments that addresses the elements of practice 2.18. Justice: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Attorney General take the following six actions: * document the department's IT strategic management processes; * document how the department's IT management operations and decisions are integrated with human resources management processes; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the department's IT goals in its IRM plan, and track actual-versus-expected performance for these IT performance measures; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Attorney General take the following five actions: * develop work processes and procedures for the department's investment management boards, including aligning and coordinating IT investment decision making among its various boards; * establish a policy requiring that IT investments be in compliance with the agency's enterprise architecture; * include net risks and risk-adjusted return-on-investment in the department's project selection criteria; * implement a scoring model and develop a prioritized list of investments as part of the department's project selection process; and: * require that corrective actions be undertaken, tracked, and reported to the investment management board for under-performing projects. Labor: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Labor take the following five actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop a goal related to how IT contributes to program productivity; * develop performance measures related to how IT contributes to program productivity, efficiency, and the effectiveness of controls to prevent software piracy, and track actual-versus-expected performance; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Labor take the following five actions: * include a description of the relationship between the IT investment management process and the department's enterprise architecture in the department's IT capital planning and investment control guide; * include net risks and risk-adjusted return-on-investment in its project selection criteria; * establish a policy requiring modularized IT investments; * develop decision-making rules to help guide the investment management board's oversight of IT investments during the control phase; and: * develop an early warning mechanism that enables the investment management board to take corrective action at the first sign of cost, schedule, or performance slippages. National Aeronautics and Space Administration: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Administrator of the National Aeronautics and Space Administration take the following seven actions: * document the agency's IT strategic management processes; * document how the agency's IT management operations and decisions are integrated with human resources management processes; * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the agency's enterprisewide IT performance measures in its IRM plan; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Administrator of the National Aeronautics and Space Administration take the following four actions: * revise the agency's IT investment management policy and guidance to describe the relationship of this process to the agency's enterprise architecture; * include interfaces in its inventory of the agency's major information systems, implement a standard, documented procedure to maintain this inventory, and develop a mechanism to use the inventory as part of managerial decision making; * within the agency's IT investment selection process, implement a mechanism to identify possible conflicting, overlapping, strategically unlinked, or redundant proposals; implement a scoring model; and develop a prioritized list of investments; and: * document the role, responsibility, and authority of its IT investment management boards, including work processes, alignment, and coordination of decision making among its various boards, and document the processes for controlling and evaluating IT investments, such as those in practices 2.15, 2.16, 2.17, and 2.18. National Science Foundation: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Director of the National Science Foundation take the following five actions: * document the agency's IT strategic management processes; * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * implement a process for assigning roles and responsibilities for achieving its IT goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Director of the National Science Foundation take the following four actions: * develop an IT investment management guide that includes a description of the relationship between the IT investment management process and the agency's other organizational plans and processes and its enterprise architecture, and identify external and environmental factors that influence the process in the agency's IT capital planning and investment control policy; * implement a structured IT selection process that includes the elements of practices 2.12 and 2.13; * involve the department's IT investment management board in controlling and evaluating IT investments, including the development and documentation of oversight processes such as those in practices 2.15, 2.16, 2.17, and 2.18; and: * define and document the elements of the agency's postimplementation reviews. Navy: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of the Navy take the following three actions: * develop a documented process to measure progress against the department's enterprisewide IT goals and assign roles and responsibilities for achieving these goals; * develop an IT goal related to service delivery to the public; and: * develop IT performance measures related to the department's IT goals, including, at a minimum, measures contained in practice 1.9 in our report, and track actual-versus-expected performance. To improve the department's IT investment management processes, we recommend that the Secretary of the Navy take the following four actions: * include net risks and risk-adjusted return-on-investment in the department's project selection criteria; * implement a structured IT selection process that includes the elements of practice 2.13; * involve all elements of the department's IT investment management board governance process in selecting, controlling, and evaluating IT investments; and: * document the role, responsibility, and authority of its IT investment management boards, including work processes, alignment, and coordination of decision making among its various boards, and document the processes for controlling and evaluating IT investments, such as those outlined in practices 2.15, 2.16, 2.17, and 2.18. Nuclear Regulatory Commission: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Commissioner of the Nuclear Regulatory Commission take the following five actions: * document the agency's roles and responsibilities for its IT strategic management processes and how IT planning is integrated with its budget and human resources planning; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to assign roles and responsibilities for achieving its enterprisewide IT goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; and: * develop performance measures for the agency's enterprisewide goals in its IRM plan, and track actual-versus-expected performance for these measures. To improve the agency's IT investment management processes, we recommend that the Commissioner of the Nuclear Regulatory Commission take the following five actions: * include a description of the relationship between the IT investment management process and the department's other organizational plans and processes and its enterprise architecture, and identify external and environmental factors that influence the process in the agency's IT capital planning and investment control policy; * develop work processes and procedures for the agency's investment management boards; * implement a standard, documented procedure to maintain its IT asset inventory, and develop a mechanism to use the inventory as part of managerial decision making; * develop a structured IT investment management selection process that includes project selection criteria, a scoring model, and prioritization of proposed investments; and: * document the role, responsibility, and authority of its IT investment management boards, including work processes and control, and evaluate processes that address the oversight of IT investments, such as what is outlined in practices 2.15, 2.16, 2.17, and 2.18. Office of Personnel Management: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Director of the Office of Personnel Management take the following four actions: * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop performance measures related to the effectiveness of controls to prevent software piracy; * track actual-versus-expected performance for the agency's enterprisewide IT performance measures in its IRM plan; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Director of the Office of Personnel Management take the following four actions: * develop work processes and procedures for the agency's investment management board, including establishing criteria for defining major systems and documenting a process for handling cross-functional investments; * implement a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * establish a policy requiring modularized IT investments; and: * require that corrective actions be undertaken, tracked, and reported to the investment management board for under-performing projects. Small Business Administration: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Administrator of the Small Business Administration take the following five actions: * document the agency's IT strategic management processes; * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a documented process to develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the agency's IT goals in its IRM plan, including, at a minimum, measures related to how IT contributes to program productivity, efficiency, effectiveness, the overall performance of its IT programs, and the effectiveness of controls to prevent software piracy, and track actual-versus-expected performance for these IT performance measures; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Administrator of the Small Business Administration take the following two actions: * document a process that the investment management board can invoke final decision-making authority over IT investments addressed by lower- level boards; and: * implement a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs. Social Security Administration: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Commissioner of the Social Security Administration take the following three actions: * include in its annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop performance measures related to the performance of the agency's IT programs and the effectiveness of controls to prevent software piracy; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Commissioner of the Social Security Administration take the following four actions: * develop work processes and procedures for the agency's investment management board; * establish a policy requiring modularized IT investments; * document the role, responsibility, and authority of its IT investment management board for the oversight of IT investments, such as what is outlined in practices 2.15, 2.16, and 2.18; and: * require that corrective actions be tracked and reported to the investment management board for under-performing projects. State: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of State take the following two actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of State take the following five actions: * implement a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * establish a policy requiring modularized IT investments; * include risk-adjusted return-on-investment in the department's project selection criteria; * revise the department's draft IT investment management policy to include reviewing projects at major milestones; and: * fully implement an IT investment management control phase, including the elements contained in practices 2.16 and 2.17. Transportation: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Transportation take the following five actions: * document its IT strategic planning process; * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop a goal related to how IT contributes to program productivity; * develop performance measures related to the department's IT goals in its IRM plan, and track actual-versus-expected performance for these IT performance measures; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of Transportation take the following six actions: * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments; * implement a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * prioritize the department's IT proposals; * establish a policy requiring modularized IT investments; * develop and document decision-making rules to help guide the investment management board's oversight of IT investments during the control phase; and: * as part of the department's control phase, employ an early warning mechanism, and use independent verification and validation reviews, when appropriate. Treasury: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of the Treasury take the following four actions: * include in the department's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; * develop performance measures related to the effectiveness of controls to prevent software piracy; * develop performance measures related to the department's IT goals in its IRM plan, and track actual-versus-expected performance for these IT performance measures; and: * develop a mechanism for benchmarking the department's IT management processes, when appropriate. To improve the department's IT investment management processes, we recommend that the Secretary of the Treasury take the following eight actions: * develop a capital planning and investment control guide that includes, for example, the elements of practice 2.1; * develop work processes and procedures for the agency's IT investment management board, and document the alignment and coordination of responsibilities of its various boards for decision making related to investments, including the criteria for which investments--including cross-cutting investments--will be reviewed by the enterprisewide board; * use the department's IT asset inventory as part of managerial decision making, including using it to identify the potential for asset duplication; * establish a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * implement a structured IT selection process that includes the elements of practices 2.12 and 2.13; * establish a policy requiring modularized IT investments; * implement an IT investment management process that includes a control phase that addresses, for example, the elements of practices 2.15, 2.16, and 2.17; and: * implement an IT investment management process that includes an evaluation phase that addresses, for example, the elements of practice 2.18. U.S. Agency for International Development: To improve the agency's IT strategic planning/performance measurement processes, we recommend that the Administrator of the U.S. Agency for International Development take the following two actions: * include in the agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; and: * develop a mechanism for benchmarking the agency's IT management processes, when appropriate. To improve the agency's IT investment management processes, we recommend that the Administrator of the U.S. Agency for International Development take the following nine actions: * develop work processes and procedures for the agency's IT investment management board; * establish a policy requiring that IT investments be in compliance with the agency's enterprise architecture; * develop a policy requiring that proposed IT investments support work processes that have been simplified or redesigned to reduce costs and improve effectiveness and that makes maximum use of COTS software; * include net risks, risk-adjusted return-on-investment, and qualitative criteria in the agency's project selection criteria; * within the agency's IT investment selection process, implement a mechanism to identify possible conflicting, overlapping, strategically unlinked, or redundant proposals; * develop a policy requiring modularized IT investments; * develop decision-making rules, review projects at major milestones, and require projects to report on deviations in system capability to help guide the oversight of IT investments by the agency's investment management board during the control phase; * as part of the agency's control phase, employ an early warning mechanism, and use independent verification and validation reviews, when appropriate; and: * require that corrective actions be undertaken, tracked, and reported to the investment management board for under-performing projects. Veterans Affairs: To improve the department's IT strategic planning/performance measurement processes, we recommend that the Secretary of Veterans Affairs take the following four actions: * include in the department's annual performance plan the resources required to implement the information security program plan required by FISMA; * develop a documented process to measure progress against the department's IT goals, and assign roles and responsibilities for achieving these goals; * develop performance measures related to the effectiveness of controls to prevent software piracy; and: * track actual-versus-expected performance for the department's enterprisewide IT performance measures in its IRM plan. To improve the department's IT investment management processes, we recommend that the Secretary of Veterans Affairs take the following two actions: * document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments, including cross-cutting investments; and: * within the agency's IT investment selection process, implement a mechanism to identify possible conflicting, overlapping, strategically unlinked, or redundant proposals, and prioritize its IT investments. [End of section] Appendix II: Comments from the Department of Agriculture: USDA: United States Department of Agriculture: Office of the Chief Information Officer: 1400 Independence Avenue SW: Washington, DC 20250: DEC 3 2003: David A. Powner, Director: Information Technology Management Issues: General Accounting Office: Dear Mr. Powner: The United States Department of Agriculture (USDA) has reviewed draft report number GAO-04-49 entitled "Information Technology Management - Government-wide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" and is in agreement with the findings. USDA is fully committed to meeting the Federal information technology challenges that are outlined in the report. USDA is currently taking steps to implement the recommendations made in the report. For example, we are in the process of updating our Capital Planning and Investment Control guidance to reflect the improvements suggested by the GAO. Additional measures are being put in place to 1) improve the Department's IT strategic management/performance measurement processes; and 2) improve the Department's information technology (IT) investment management processes. We currently estimate that the six strategic management/performance measurement recommendations will be fully implemented by the end of fiscal year 2005 and the four investment management recommendations will be fully implemented by the end of fiscal year 2004. Thank you for the opportunity to review and provide feedback on the draft report. If additional information is needed, please contact Dr. Gregory Parham of my staff on (202) 720-5865. Sincerely, Signed for: Scott Charbo: Chief Information Officer: [End of section] Appendix III: Comments from the Department of Commerce: THE SECRETARY OF COMMERCE: Washington, D.C. 20230: December 11, 2003: Mr. David A. Powner: Director, Information Technology Management Issues: United States General Accounting Office: Washington, DC 20548: Dear Mr. Powner: Thank you for the opportunity to comment on the GAO draft report "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved." We appreciate the thoroughness of your review and believe that our information technology strategic planning and capital investment processes will benefit from your insights. We concur with the report's recommendations, and are updating the documentation for our relevant policies and procedures. Sincerely, Signed by: Donald L. Evans: [End of section] Appendix IV: Comments from the Department of Defense (including comments from the Departments of the Air Force, Army, and Navy): [See PDF for image] [End of figure] The following are GAO's comments on the Department of Defense's (DOD) letter dated December 5, 2003. GAO Comments: 1. DOD provided its annual report to the President and the Congress, which included its fiscal year 2004 performance plan. Based on a review of this plan, we modified our report. 2. We disagree that the cited objective fully addresses this issue. Specifically, although this objective addresses e-government, the wording of the objective, its description, and the discussion of related initiatives do not explicitly address service delivery to the public. Accordingly, we did not modify our report. 3. Our review of the acquisition management process documentation provided by the Navy did not support that the department's selection criteria include net risks and risk-adjusted return-on-investment. Accordingly, we did not modify our report. [End of section] Appendix V: Comments from the Department of Education: UNITED STATES DEPARTMENT OF EDUCATION: OFFICE OF MANAGEMENT: ASSISTANT SECRETARY: December 10, 2003: To: David A. Powner: Director, Information Technology Management Issues: From: William Leidinger: Assistant Secretary for Management and Chief Information Officer: Signed by: William Leidinger: Re: Draft Report: Information Technology Management: Government Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved: We thank you for the opportunity to review the Information Technology Management Report. In general, we agree with your assessment of our implementation of the various strategic planning and performance measurement practices as well as our investment management practices. We have taken the position that IRM planning is subsumed in the Department's Strategic and Annual Plans (Goal 6.3), and occurs in the development and maintenance of the enterprise architecture. Our architecture is business driven and our IT investments are aligned with the architecture. We believe this approach meets the intent of both the Paperwork Reduction Act and the Clinger-Cohen Act. You specifically recommended that we develop measures related to how IT contributes to program productivity and the effectiveness and efficiency of agency operations. Currently, we require all of our major and significant IT investments to collect and track against business process performance measures, IT performance measures, and customer- based performance measures. We assess progress that is reviewed by the Department's senior management quarterly. We believe that this approach satisfies the requirement that we develop the measure you stipulate. Finally, specifically regarding IT Management practices 2.5 and 2.6 on pages 46 and 47, the Department does have a repeatable process for developing and maintaining an inventory of its major information systems. The Department has documented this procedure in Handbook OCIO- 09, Handbook for Information Technology Security General Support Systems and Major Applications Inventory Procedures. This process is formally completed twice a year and includes documented Critical Infrastructure Protection questionnaires and Systems Inventory forms to support the contents of the Department's IT systems inventory. The following are GAO's comments on the Department of Education's letter dated December 10, 2003. GAO Comments: 1. We agree that Education requires IT investments to have performance measures. However, our practice dealt with enterprise-level measures, such as those found in the department's IRM plan, not project-specific measures. Education reported that the performance measures in its IRM plan do not measure how IT contributes to program productivity and the efficiency and effectiveness of agency operations. Accordingly, we did not modify our report. 2. We modified our assessment of practice 2.6 in this report and deleted the related recommendation based on our evaluation of additional documentation provided by Education. [End of section] Appendix VI: Comments from the Environmental Protection Agency: UNITED STATES ENVIRONMENTAL PROTECTION AGENCY: WASHINGTON, D.C. 20460: DEC 9 2003: OFFICE OF ENVIRONMENTAL INFORMATION: Mr. David Powner: Director, Information Technology Management Issues: U.S. General Accounting Office: 441 G Street, NW Washington, D.C. 20548: Dear Mr. Powner: Thank you for the opportunity to respond to the U.S. General Accounting Office (GAO) draft Report to Congressional Requesters "Information Technology Management: Government-wide Strategic Planning, Performance Measurement, and Investment Management Can be Further Improved" (GAO- 04-49). The U.S. Environmental Protection Agency (EPA) is pleased to note that the reviewers acknowledge that the Agency has established and employs best-practices for IT management for the majority of key dimensions of the report. We accept the reviewers' findings and recommendations that in some cases the Agency does need to complete work currently underway to formalize the documentation of those practices. We expect to promulgate a final version of new Agency-wide official procedures in the near future. At the same time, EPA believes that the report does not adequately reflect both key dimensions of EPA's IT management and strategic planning, and current best-practices in general. The draft report includes reference to the importance of enterprise architectures ("Constructing and enforcing sound enterprise architectures," p. 13). However, the sections regarding IRM strategic planning do not reflect how EPA's Enterprise Architecture (EA) constitutes the multi year, business driven, enterprise-wide comprehensive IRM strategic plan. The EA should, and for EPA does, derive first from strategic, line of business, performance and outcome-based needs. Those needs tie to the comprehensive range of IT assets (applications, data, services, hardware, software, security) from a strategic perspective (the multi year target), leading to the business-based priorities to move toward those goals (baseline and sequencing plans). As Chief Information Officer, I govern EPA's EA and the IT Capital Planning and Investment Control (CPIC) process to implement the architecture. This occurs through a formal integrated process in partnership with Agency senior executives, including the Chief Financial Officer. EPA requests that the reviewers reassess this dimension of their findings to acknowledge how EPA fulfills strategic planning mandates via our implementation of enterprise architecture and IT investment management practices. This also affects the recommendations regarding tracking actual-versus-expected performance as well as including net risks in project selection criteria and modularized IT investments which are all part of our EA and CPIC practice. In addition, for Practice 1.7 ("The agency has a documented process to: develop IT goals in support of agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals," p. 32) the reviewers state EPA "does not have this practice in place." As noted above, EPA has a very rigorous process to accomplish this through the integration of our EA and CPIC processes. The finding should be corrected to include EPA in the "partially" category to acknowledge we do have the practice in place and are working on documentation. Again, thank you for the opportunity to respond to this important report. EPA is proud of the successful implementation of many of the best practices in IT investment and infrastructure. If you have any questions relating to this information, please contact Steve Tiber, EPA-GAO Liaison, at 202-564-5184. Sincerely, Kimberly T. Nelson: Assistant Administrator and Chief Information Officer: Signed by: Kimberly T. Nelson: cc: Ramona Trovato; Michael W.S. Ryan; Mark Day; Maggie Mitchell: The following are GAO's comments on the Environmental Protection Agency's (EPA) letter dated December 9, 2003. GAO Comments: 1. As we reported and EPA acknowledged, its documentation on IT strategic planning and investment management was not complete or finalized. For example, the partial rating we gave EPA for its IT management and strategic planning practices--practices 1.1 and 1.2-- matched the agency's own self-assessment in these areas. Specifically, our review of planning documents cited by EPA in its self-assessment found that while the agency had documented agencywide roles and responsibilities for planning and managing IT resources and had documented its process to integrate the IT investment management process with the budget, EPA had not addressed other key elements of the practices. As an example, EPA had not fully documented the method by which it defines program information needs and develops strategies, systems, and capabilities to meet those needs. Since EPA provided no additional documentation, our practice assessment and our related recommendations remain unchanged. 2. As stated in our report, practice 1.7 refers to the documentation of the process used to develop IT goals and measures and the responsibility for achieving them. As EPA states in its comments, it is currently working on documenting this process. Accordingly, we did not modify our report. [End of section] Appendix VII: Comments from the General Services Administration: GSA: GSA Office of the Chief Information Officer: December 9, 2003: The Honorable David M. Walker: Comptroller General of the United States: General Accounting Office: Washington, DC 20548: Dear Mr. Walker: This is to provide comments on the General Accounting Office (GAO) findings and recommendations included in the Governmentwide Strategic Planning, Performance Measurement, and Investment Management Practices draft report. We generally agree with the findings and recommendations and are pleased that General Services Administration (GSA) was identified as fully meeting 21 of the 29 practice areas included in the report. There were eight (8) areas identified as GSA partially meeting and one (1) as GSA not meeting. We are concentrating our comments on the areas identified as partially or not meeting where we feel additional information or clarification needs to be provided to GAO. We appreciate the opportunity to review the draft report. Should you require additional information, please contact Ms. L. Diane Savoy, Director, Office of Policy and Plans, at (202) 501-3535. Sincerely, Signed by: Michael W. Carleton: Chief Information Officer: Enclosures: GAO GOVERNMENTWIDE STRATEGIC PLANNING, PERFORMANCE MEASUREMENT, AND INVESTMENT MANAGEMENT REVIEW: Table 1: IT Strategic Planning/Performance Measurement Practices: Practice 1.4: The agency has a process that involves the CFO or comparable official, to develop and maintain a full and accurate accounting of IT-related expenditures, expenses, and results. GAO Assessment: Partially - GSA has reported that not all costs may have been captured. GSA OCIO Response: To the best of our knowledge, all IT costs are captured in the GSA IT Capital Plan and submitted with the agency budget annually. We have been and continue to be very diligent about ensuring that all IT investments, agency-wide, are identified and captured in our IT Capital Plan. Practice 1.5: The agency prepares an enterprisewide strategic information resources management (IRM) plan that, at a minimum: * describes how IT activities will be used to help accomplish agency missions and operations, including related resources; * identifies major IT acquisition program(s) or any phase or increment of that program that has significantly deviated from the cost, performance, or schedule goals established for the program. GAO Assessment: Partially-GSA ... IRM plan(s) do not include resources and major IT acquisition programs that deviated from cost, schedule, or performance goals. GSA OCIO Response: The GSA IT Strategic Plan is a strategic document that sets the direction and focus of the GSA's IT program over a five-year period. It is not considered the appropriate place nor is it intended to track deviations from planned cost, performance or schedule goals. This information is tracked as part of the Control Phase of our IT Capital Planning and Investment Control process. GSA has provided GAO with significant documentation from our Project Summary Control database that currently tracks major investment deviations from planned cost, performance or schedule goals on a monthly basis. In addition, the project managers track these same items closely as they manage their investments day-to-day. We request that GAO reconsider this statement and modify it to at least acknowledge the processes that GSA has in place to monitor the planned and actual cost, performance and schedule. In support of this, we are attaching copies of additional monitoring documents that are utilized and that identify deviations cost, schedule or performance. (Attachments 1 and 2): Practice 1.6: The agency's performance plan required by GPR includes: * A description of how IT supports strategic and program goals: * The resources and time periods required to implement the information security program plan required by the Federal Information Security Management Act (FISMA), and: * A description of major IT acquisitions contained in the capital asset plan that will bear significantly on the achievement of a performance goal. GAO Assessment: Partially-No agency's performance plan, except VA's, includes time periods, and none includes resources required to implement the information security program plan required by FISMA. GSA OCIO Response: This will be reviewed by, coordinated with the Office of the Chief Financial Officer and the Senior Agency Information Security Official, and corrected appropriately. We note however, that GSA has developed and update quarterly to OMB, as required, the FISMA security report and Plan of Actions and Milestones (POA&M) which includes specifics and details regarding the resources and time periods to accomplish necessary corrective actions. Practice 1.9: The agency has established IT performance measures and monitors actual-Versus-expected performance that at least addresses: * How IT contributes to program productivity, * How IT contributes to the efficiency of agency operations, * How IT contributes to the effectiveness of agency operations, * Service delivery to the public (if applicable, * How electronic government initiatives enable progress toward agency goals and statutory mandates, * The performance of IT programs (e.g., system development and acquisition projects), and: * Agency compliance with federal software piracy policy. GAO Assessment: Partially-GSA does not have measures for one of its IT goals and did not have performance measures on service delivery to the public. GSA OCIO Response: GSA has performance measures that map to each of the IT goals. We have attached a revised copy of the GSA Goals to Performance Measures Mapping document that was submitted earlier. (Attachment 3): Please note that GSA's mission is to "help Federal agencies better serve the public by offering, at best value, superior workplaces, expert solutions, acquisition services and management policies." As clearly defined in our mission statement, GSA's primary role is in support of other Federal agencies. We have a lesser role in providing information to citizens through our Office of Citizen Services and Communications that was established on June 30, 2002. This Office has the following performance goals that are related providing service to citizens: 1. Increase access to government for all audiences through the use of web sites, call centers, e-mail, publications, and all forms of media. 2. Develop single government fact to citizens to enable the Federal government to become more citizen centric by presenting a "front door" for citizens who need timely, accurate, consist responses about government programs. The attached a revised copy of the GSA Goals to Performance Measures Mapping includes the specific performance measures to accomplish these goals. In further support of this, we have also attached copies of the performance information and results from the GSA Performance Measurement Tool for these goals. (Attachment 4): Practice 1.12: The agency requires that its IT management process be benchmarked against appropriate processes and/or organizations from the public and private sectors in terms of cost, speed, productivity, and quality of outputs and outcomes where comparable process and organizations in the public or private sectors exist. GAO Assessment: Partially - . . . GSA ... provided an example of a process that they have benchmarked, but benchmarking if being performed on an ad hoc basis. GSA OCIO Response: GSA has identified the need for and used benchmarking when it has been determined that comparable process and organizations in the public or private sector exist and that it will provide additional value in management decision-making. GSA recognizes the benefits of benchmarking and will continue to utilize it, as appropriate. Table 2: IT Investment Management Practices: Practice 2.3: The agencywide board(s) work processes and decision- making processes are described and documented. GAO Assessment: Partially - GSA did not have policies and procedures for each of its IT investment management boards. GSA OCIO Response: GSA was advised by GAO that because the pending Business Systems Council (BSC) charter, for the executive tier of our process is in draft that we would be rated "partially" on this practice. We are still planning to have the BSC Charter finalized in place before the end of the year. Charters exist for all other tiers of our investment management boards, and overall policies exist in the GSA IT Capital Planning and Investment Control policy and Guide. Copies of these documents were provided to GAO earlier. Practice 2.14: Agency policy calls for investments to be "modularized" (e.g., managed and procured in well-defined useful segments or "modules" that are short in duration and small in scope) to the maximum extent achievable. GAO Assessment: No-_ GSA ... do not have this practice in place. GSA OCIO Response: In earlier versions of the GSA IT Capital Planning and Investment Control Guide, that GAO was provided, included this requirement. This requierment was inadvertently not included in the August 2002 version of the Guide, which was reviewed by GAO. The GSA Capital Planning and IT Investment Guide will be updated and will include modularization. Practice 2.15: The agency wide investment management board(s) has written policies and procedures for management oversight of IT projects that cover at a minimum: * decision-making rules for project oversight that allow for the termination of projects, when appropriate; * current project data, including expected and actual cost, schedule, and performance data, to be provided to senior management periodically and at a major milestones; * criteria or thresholds related to deviations in cost, schedule, or system capability actual vs. expected project performance; and: * the generation of an action plans to address a project's problem(s) and track resolution. GAO Assessment: Partially - GSA does not have clear decision-making rules, require projects to report on deviations in system capability, or require that corrective actions be tracked to resolution. GSA OCIO Response: The GSA IT Capital Planning and Investment Control Guide identifies the requirements for reporting on major investments in the Control Phase of the IT Capital Planning and Investment Control process. The reporting is accomplished through the Summary Project Control database. Samples of reports from this database have been provided to GAO and provide additional samples are included in Attachments 1 and 2. The GSA IT Capital Planning and Investment Control guide requires the Information Technology Resources Boards (ITRBs) and the Technical Review Boards (TRBs), managed by sub-agency CIO's with the participation of the agency CIO, to closely monitor investments, and to develop and monitor correction actions when deviations occur. While this is a decentralized approach, it is fully integrated into our enterprise-wide process. In the upcoming revision of our IT Capital Planning and Investment Control Guide, we will add more details to further clarify this process. Practice 2.17: Corrective actions for under performing projects are agreed upon, documented, and tracked by the agencywide investment management board(s). GAO Assessment. Partially - GSA ... do not systematically track corrective actions. GSA OCIO Response: GSA Response: As stated in reference to 2.15 above, the GSA IT Capital Planning and Investment Control guide requires the Information Technology Resources Boards (ITRBs) and the Technical Review Boards (TRBs), managed by sub-agency CIO's with the participation of the agency CIO, to closely monitor investments, and to develop and monitor corrective actions when deviations occur. Appendix I, Recommendations to Department and Agencies: General Services Administration: GAO Recommendation: To improve the agency's IT strategic management/ performance measurement processes, we recommend that the Administrator or the General Services Administration: * develop performance measures for each of its IT goals in its ITM plan as well as measures related to how IT contributes to service delivery to the public and the effectiveness of controls to prevent software piracy: GSA Response: As noted in our response to Practice 1.9 and documented in attachment 3, GSA does have performance measures for each IT goal in the IT Strategic Plan. In addition, the response to Practice 1.9 also identified measures in place to address how IT contributes to service delivery to the public. We have also attached a copy of the GSA Order, 2104.1, GSA IT General Rules of Behavior, dated July 3, 2003 that outlines the GSA policy and requirements designed to prevent software piracy (Attachment 5). The specifics are included in Section 10. Software acceptable use. In addition, Section 6. Penalties for non-compliance stated "users who do not comply with the IT General Rules of Behavior may incur disciplinary action and/or criminal prosecution.": The GAO stated recommendation should be removed from the report. GAO Recommendation: To improve the agency's IT investment management processes, we recommend that the Administrator or the General Services Administration: * require corrective actions be undertaken, tracked, and reported to the investment management board for underperforming projects. GSA Response: As stated in our response to Practice 2.15 and 2.17, the GSA IT Capital Planning and Investment Control Guide does require corrective actions to be undertaken, tracked, and reported to the investment management board for underperforming projects. Specifically, the Guide, in Section 8, The IT Capital Planning and Investment Control process, (b) Control Phase, (1). Reporting, includes the following excerpts with critical parts highlighted in bold to draw your attention to these sections: The SSO CIOs will submit monthly stop light reports on all major IT projects that are ongoing developments, acquisitions, or enhancements. The OCIO will use these stop light reports to monitor projects through their life cycle. In addition to alerting the CIO of projects that vary significantly from planned cost, schedule and performance estimates, stop light reports will allow intermittent updates of baseline information on the agency's overall IT investments. For any reports which indicate a yellow or red light, SSOs must submit an update that indicates the change in status and the actions that are being accomplished to address conditions underlying the yellow or red light. Reports with red light changes must provide get-well plans. The SSO CIO's will synthesize project status data and submit summary control reports to the ITC and Executive Committee, as necessary. Variances in project schedule or cost goals of 10 percent or more must be reported to the ITC. Any variance or slippage in actual performance from established goals must be included in monthly stoplight reports to the CIO. In addition, projects which, based upon monthly reports or other indicators, consistently fail to meet requirements may be, in consultation with the SSO CIO, subject to a special independent review by an Independent Verification and Validation (IV&V) contractor or Governmentwide ITRB. The results of the review may be presented to the ITC and/or Executive Committee, as appropriate. When a TRB, ITRB, IV&V or Governmentwide ITRB finds a project has significant deviation from planned performance, cost and schedule, it may result in the project being modified or terminated. The GSA IT Capital Planning and Investment Control Guide, section (d) Executive and technical oversight, includes the following: (1). Executive Committee. The Executive Committee approves the IT Strategy and IT Capital Plan and acts on projects that significantly deviate from investment controls. (NOTE; This Committee is being replaced by the Business Systems Council being established): The Executive Committee makes the decision for final approval of the investment portfolio and decisions on systems that have significant deviations from planned performance, cost and schedule. The GAO stated recommendation should be removed from the report. The following are GAO's comments on the General Services Administration's (GSA) letter dated December 9, 2003. GAO Comments: 1. We based our evaluation on the agency's self-assessment and comments made by GSA's Director, Office of Policy and Plans. However, based on GSA's representation in commenting on our draft, we changed our evaluation of the referenced practice. 2. The Clinger-Cohen Act requires agencies to include in its information resources management (IRM) plan the identification of a major IT acquisition program(s), or any phase or increment of that program, that significantly deviated from cost, performance, or schedule goals established by the program. As we acknowledge in this report, agencies, which would include GSA, identified other mechanisms that they use to track and report cost, schedule, and performance deviations. Moreover, we evaluated agencies as a "partially" instead of a "no" in this practice to take into account that the agency had the required information, although it was not in the prescribed format. Accordingly, we did not modify our report. 3. The Federal Information Security Management Act of 2002 requires agencies to include in the performance plans required by the Government Performance and Results Act the resources and time periods to implement their information security program. As we noted in this report, agencies, which would include GSA, commonly stated that they had this information but that it was in another document. Nevertheless, this does not negate the need for having the agency report to the Congress in the form that it requires. This is particularly important since performance plans are public documents. Accordingly, we did not modify our report. 4. GSA's new documentation illustrates that it has performance measures for each of the IT goals in its IRM plan. However, GSA did not provide evidence that it was tracking actual versus expected performance for measures associated with one of its goals. We revised our report to reflect GSA's new documentation and our evaluation. 5. We revised our report on the basis of this new documentation. 6. GSA's highest-level IT investment management board is its Executive Committee. GSA did not provide a charter or any other evidence of policies and procedures for this committee. We therefore did not modify our report. 7. The additional documentation provided by GSA (1) does not address decision-making rules and (2) illustrates that GSA uses a monthly project control report on cost, schedule, and performance status, but the report does not explicitly address deviations in system capability. In addition, according to GSA's capital planning and investment control order, the format of the report is left to the applicable organization, thereby making it less likely that the investment management boards are obtaining consistent information. We therefore did not modify our report. 8. We agree that GSA's capital planning and investment control order requires that projects that have significant variances are to provide "get well" plans and that monthly control reports are used to report on project cost, schedule, and performance status. However, it is not clear that these status reports can be used to systemically track corrective actions. Moreover, according to GSA's capital planning and investment control order, the format of the monthly control report is left to the applicable organization, thereby making it less likely that the status of corrective actions is being consistently reported. We therefore did not modify our report. 9. See comment 8. 10. We modified our recommendations based on our evaluation of GSA's documentation. See comment 4 for our assessment. 11. Executive Order 13103 requires agencies to use software piracy performance measures that comply with guidance issued by the federal CIO Council.[Footnote 46] The Council, in turn, called on the agencies to develop such measures. The additional documentation that GSA provided was an order requiring agency employees to use properly licensed software, but it does not include performance measures that would demonstrate that this requirement is being honored. Measuring how well agencies are combating software piracy is important because it can verify that the controls that they have put in place are working. Accordingly, we did not change this part of the recommendation. 12. We modified our recommendation to reflect that GSA requires projects that have significant variances to develop corrective action plans. However, the other elements of the recommendation pertaining to the tracking and reporting on corrective actions remain outstanding. See comment 8 for additional information. [End of section] Appendix VIII: Comments from the Department of Health and Human Services: DEPARTMENT OF HEALTH & HUMAN SERVICES: Office of Inspector General: Washington, D.C. 20201: DEC 15 2003: Mr. David A. Powner: Director, Information Technology Management Issues: United States General Accounting Office: Washington, D.C. 20548: Dear Mr. Powner: Enclosed are the Department's comments on your draft report entitled, "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further: Improved." The comments represent the tentative position of the: Department and are subject to reevaluation when the final version of: this report is received. The Department appreciates the opportunity to comment on this draft: report before its publication. Sincerely, Signed by: Dara Corrigan: Acting Principal Deputy Inspector General: Enclosure: The Office of Inspector General (OIG) is transmitting the Department's response to this draft report in our capacity as the Department's designated focal point and coordinator for General Accounting Office reports. OIG has not conducted an independent assessment of these comments and therefore expresses no opinion on them. Comments of the Department of Health and Human Services on the General Accounting Office's Draft Report, "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (GAO-04-49): The Department of Health and Human Services (Department) appreciates the opportunity to comment on this draft report. The Department concurs with the findings and recommendations contained in the report. [End of section] Appendix IX: Comments from the Department of Housing and Urban Development: DEPARTMENT OF HOUSING AND URBAN DEVELOPMENT: WASHINGTON, DC 20410-3000: ASSISTANT SECRETARY FOR ADMINISTRATION/CHIEF INFORMATION OFFICER: DEC 15 2003: Ms. Linda Lambert: Assistant Director, Information Technology Management Issues: U.S. General Accounting Office: Washington, DC 20548: Dear Ms. Lambert: Thank you for the opportunity to comment on the GAO's draft report on "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can be Further Improved (GAO-04-49)." The report contains six recommendations to improve HUD's information technology (IT) strategic management/ performance measurement processes and five recommendations to improve the investment management processes. We are in agreement with all the recommendations in the report. We are pleased that of the 29 applicable practices GAO examined at HUD, 17 are in place and 8 are partially implemented. In the coming months, we will focus on implementing the four practices that are not in place and strengthening those processes where further improvements can be made. The information in the report will help form a baseline against which we will measure our continuing improvement efforts. Should you or your staff have any questions or require additional information, please contact Mary P. Barry, Acting Director, Office of Management and Planning, at (202) 708-1027, extension 123. Sincerely, Signed by: Vickers B. Meadows: Assistant Secretary for Administration/Chief Information Officer: cc: David Powner: [End of section] Appendix X: Comments from the Department of the Interior: United States Department of the Interior: OFFICE OF THE ASSISTANT SECRETARY: POLICY, MANAGEMENT AND BUDGET: Washington, DC 20240: David A. Powner: Director, Information Technology Management Issues: United States General Accounting Office: 441 G Street, NW: Washington, DC 20548: Dear Mr. Powner: The Department of the Interior reviewed the Draft GAO Report entitled, "Information Technology Management: "Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (GAO-04-49). This report follows closely upon the completion of the GAO Report on information technology investment management (ITIM) in Interior entitled "Information Technology: Departmental Leadership Crucial to Success of Investment Reforms at Interior" (GAO-03-1028), in which GAO commended Interior for the progress made to date in ITIM and provided recommendations for further improvement. In response to this report, Interior committed to develop a comprehensive plan to implement departmentwide improvements to the ITIM process based on the Stage 2 and Stage 3 critical processes of GAO's ITIM framework, and strengthen departmental oversight of bureau investment management processes. The draft report (GAO-04-49), which covers several agencies, acknowledges that Interior's progress in improving information technology investment management (ITIM) has been evident and is ongoing. The report also provides very useful comparisons to ITIM progress in other agencies. In order to capitalize on the successes at other agencies, it would perhaps be more helpful to provide more examples for emulation in areas where agencies are successful. Interior acknowledges the assistance provided by the GAO auditors during the review process in highlighting a few examples of successes in areas where Interior needs greater improvement. Further examples of successes would provide the information needed to benchmark, as recommended in the draft report. Interior is committed to continue to move forward aggressively to execute key practices for ITIM considering competing priorities for this and other initiatives. The Department of the Interior agrees the recommendations in this report would further improve ITIM at Interior, and have plans in place to incorporate many of the suggestions noted. However, some of the recommendations go beyond what Interior could reasonably accomplish within current and projected budgets for information technology. Interior remains committed to address the recommendations of the final report GAO-03-1028. For additional information, please contact Mr. W. Hord Tipton, at 202 208 6194. Sincerely, Signed by: Lynn Scarlett: Acting Assistant Secretary for Policy, Management and Budget: [End of section] Appendix XI: Comments from the Department of Justice: U.S. Department of Justice: Washington, D.C. 20530: DEC 2 2003: David A. Powner, Director: Information Technology Management Issues: U.S. General Accounting Office: 441 G St., N. W. Washington, D.C. 20548: Dear Mr. Powner: I would like to thank you for affording the Department of Justice (Department) the opportunity to respond to the General Accounting Office (GAO) report entitled "Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved.": The Attorney General appointed me Chief Information Officer for the Department in April 2002. In July 2002,1 re-issued the Information Technology (IT) Strategic Plan conveying a fundamental reorientation of the role of IT in the Department of Justice. In May 2003, I received congressional approval to reorganize my office and realign staff to operate more effectively to carry out the management goals identified in the IT Strategic Plan. During the past 12 months, I have initiated the following specific initiatives in these areas in order to achieve measurable progress in IT management within the Department: * Initiated an outside assessment of IT investment management with recommendations to improve the Department's investment management processes; * Completed a pilot project to test and assess new processes to manage IT investments in the Select Phase resulting in a portfolio-centered analysis of candidate investments and recommendations for consolidated business cases as part of the Department's Fiscal Year 2005 budget formulation process; * Implemented an ongoing pilot project to test, assess, and phase in new processes to manage IT investments in the Control Phase, including the use of an IT Investment Board, which will enable departmental senior executives to actively set direction and monitor high priority, high risk and department wide IT investments; * Initiated a web-based "Project Dashboard" to collect current data on cost, schedule, and performance for major projects across the Department; * Consolidated business cases for enterprise architecture and IT infrastructure - two areas critical for achieving my strategic objectives; * Developed of a project oversight process in collaboration with the FBI to oversee major IT investments. I am currently institutionalizing and documenting those key management processes listed above to strengthen strategic planning, performance measurement, and IT investment across the Department. The recommendations of your report will assist me in further defining those management practices. Overall, I concur with the findings and recommendations in your report; however, I want to modify one recommendation to ensure it is an actionable item that reflects the future direction of our oversight reporting requirements. That recommendation relates to Practice 1.6 - "The agency's performance plan required by GPRA [Government Performance and Results Act] includes the resources and time periods required to implement the information security program plan required by FISMA [Federal Information Systems Management Act].": The Department's self-assessment and the GAO review results are in agreement on the Partial rating for this assessment. As discussed in the Exit Conference, the Department lays out the full FISMA program, including major milestones and required resources, in the annual Security Report and the Plan of Actions and Milestones Report submitted to the Office of Management and Budget (OMB), GAO, and Congress but includes only selected performance measures in the GPRA Plan. However, the OMB will replace the annual GPRA performance plan with performance budgeting in FY 2005. In light of this change, the following modification in language is requested: Include the resources and time periods required to implement the information security program plan required by FISMA in reporting document(s) as directed by OMB guidance. Thank you for the opportunity to comment on this report. If you need additional information, please do not hesitate to contact me. Sincerely, Signed by: Vance E. Hitch: Chief Information Officer: The following are GAO's comments on the Department of Justice's letter dated December 2, 2003. GAO Comments: 1. GAO has ongoing work looking at OMB's initiative. However, the Federal Information Security Management Act of 2002 requires agencies to include in the performance plans required by the Government Performance and Results Act the resources and time periods to implement its information security program. Accordingly, we did not change the recommendation. [End of section] Appendix XII: Comments from the Department of Labor: U.S. Department of Labor: Office of the Assistant Secretary for Administration and Management: Washington, D.C. 20210: DEC - 2 2003: Mr. David Powner: Director, Information Technology Management Issues U.S. General Accounting Office: 441 G. Street, NW Washington, D.C. 29548: Dear Mr. Powner: The enclosed comments are in reference to GAO's draft report entitled Information Technology Management: Government-wide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved (GAO 04-49). As you know, the objectives of this study were to determine the extent to which major government agencies have in place practices associated with key legislative and other requirements for (1) IT strategic planning/performance measurement and (2) IT investment management. The Department concurs with GAO's finding and is proposing two changes, found on the enclosed document. If you have any questions about these comments or this matter in general, please contact Thomas Wiesner, Deputy CIO, at wiesner.thomas@dol.gov or (202) 693-4200. Sincerely, Signed by: Patrick Pizzella: Assistant Secretary for Administration and Management, Chief Information Officer: Enclosure: DOL Response to GAO-04-49 IT Management Practices Recommendations to Departments and Agencies: 1) p. 25. "For example, the Department of Labor's... does not address all required goals.": Recommended change: After this sentence, insert new sentence reading: "However, this single IT goal is supported by a number of IT indicators, which are specific, measurable and tied to Labor's mission - and address the Clinger-Cohen and PRA requirements. The CIO is assigned responsibility for achieving the goal, with component agencies responsible for carrying out the indicators appropriate to their mission.": 2) p. 71, second bullet under Labor's recommendations, which begins "develop a documented process...": Recommended change: remove this bullet entirely for the following reasons: As required by the Government Performance Results Act (GPRA), the Department of Labor (DOL) prepares an annual performance plan covering each program activity set forth in the DOL annual budget, including information technology. This plan is built upon strategic goals established by the Department and are directly related to the Department's mission. In addition, performance indicators are established to measure outcomes. These indicators are tracked and reported quarterly through the GPRA process. DOL's IT strategic goal, "to provide better and more secure service to citizens, businesses, government and Labor employees to improve mission performance" contain a number of IT indicators, which support this goal. The indicators are specific, measurable goals that support the strategic goal and are reported in GPRA on an annual basis. In accordance with the Clinger-Cohen Act and DOL Secretary's Order 3- 2003, the Office of the Chief Information Officer is responsible for establishing, monitoring and evaluating Departmental IT goals to improve efficiencies, contain costs, streamline business processes, provide better access to Government information and services, and promote a secure environment. In addition, consistent with their statutory responsibilities and other applicable Secretary's Orders and guidelines, all DOL Agency Heads are assigned responsibility to implement Department-wide IT initiatives approved by the Department's Management Review Board (MRB). The following are GAO's comments on the Department of Labor's letter dated December 2, 2003. GAO Comments: 1. Because Labor did not disagree with our characterization of its IT goal, no changes were made to our report. 2. We agree with Labor's characterization of its IT strategic goal and order 3-2003. Nevertheless, the recommendation, and related practice 1.7, refers to the documentation of the process used to develop IT goals and measures and the responsibility for achieving them. Labor neither provided documentation of such a process nor took issue with our assessment of practice 1.7, in which we stated that the agency did not have this practice in place. Moreover, Labor's self-assessment referenced a draft performance measurement guidebook and quarterly review process in support of this practice. However, these mechanisms relate to performance measures associated with IT projects, not Labor's enterprisewide IT goal. Finally, as we noted in our report, unlike other agencies in our review, Labor does not have goals in its IRM plan. Accordingly, we did not change this recommendation. [End of section] Appendix XIII: Comments from the National Aeronautics and Space Administration: National Aeronautics and Space Administration: Office of the Administrator: Washington, DC 20546-0001: December 8, 2003: Mr. David A. Powner, Director: Information Technology Management Issues: United States General Accounting Office: Washington, DC 20548: Dear Mr. Powner: Enclosed is the National Aeronautics and Space Administration (NASA) response to the General Accounting Office (GAO) Draft Report, "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (GAO-04-49). The Agency concurs with your recommendations for corrective action. Enclosed are NASA's detailed comments on each individual recommendation. My point of contact for information technology management is Nancy Kaplan in the Office of the Chief Information Officer (CIO). She may be contacted by e-mail at Nancy.Kaplan@nasa.gov or by telephone at (202) 358-1372. Cordially, Signed by: Frederick D. Gregory: Deputy Administrator: Enclosure: NASA Response to Draft GAO Report: "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (GAO-04-49): GAO Recommendations: To improve the Agency's Information Technology (IT) strategic management/performance measurement processes, we recommend that the Administrator of the National Aeronautics and Space Administration: 1. Document the Agency's IT strategic management processes; NASA Response to GAO Recommendation 1: Concur. NASA has produced an Information Resources Management (IRM) Strategic Plan and is currently documenting the strategic planning process which will provide the framework for ensuring that the Agency's IT management strategies as defined in the IRM Strategic Plan remain clearly linked to the Agency's vision, mission, and strategic goals as defined in the NASA Strategic Plan. This activity is scheduled to be completed by June 30, 2004. 2. Document how the Agency's IT management operations and decisions are integrated with human resources management processes; NASA Response to GAO Recommendation 2: Concur. The human resources management processes are critical to our success in meeting the NASA mission, and the NASA CIO has been working closely with the NASA Office of Human Resources to tightly integrate NASA's IT human capital needs with their efforts. NASA has developed a Strategic Human Capital Plan (SHCP) that establishes a systematic, Agencywide approach to human capital management, aligned with the Agency's vision and mission. The SHCP is built around five pillars--strategic alignment, strategic competencies, learning, performance culture, and leadership. A companion document to the SHCP, the Strategic Human Capital Implementation Plan (SHCP), has also been developed which contains detailed action plans for achieving the goals defined in the SHCP. Neither of these plans reference specific disciplines (e.g., IT), but are intended to pertain to all disciplines that are critical to developing and maintaining NASA's strategic competencies. The current NASA IRM Strategic Plan does not address human resources management since the SHCP and the SHCIP address this from a multidiscipline perspective, but does reference the NASA Strategic Plan, which identifies the criticality of IT as a strategic competency to NASA through the definition of Implementing Strategy 2, "Demonstrate leadership in the use of information technology." However, the IRM Strategic Plan is currently in the process of being revised and will include language that clarifies the linkage between the various strategic plans and the integration of IT management with human resources management. The revised plan is scheduled to be completed by September 30, 2004. 3. Include in the Agency's annual performance plan the resources and time periods required to implement the information security program plan required by FISMA; NASA Response to GAO Recommendation 3: Partially Concur. NASA has been including the resources and time periods required to implement the information security changes necessary to correct the identified weakness contained in the OMB FISMA report. OMB has instructed NASA to exclude this information in any submit to Congress due to budget data being embargoed at this time. NASA will summarize this information in our annual performance plan. 4. Develop a documented process to develop IT goals in support of Agency needs, measure progress against these goals, and assign roles and responsibilities for achieving these goals; develop measures related to the effectiveness of controls to prevent software piracy; NASA Response to GAO Recommendation 4: Concur. The strategic planning process referenced in (1) above is the process through which IT goals and performance measures are developed. NASA is currently in the process of revising NASA Policy Directive (NPD) 2800.1, Managing Information Technology, which identifies the roles and responsibilities for achieving the Agency's IT goals and objectives. This revision is scheduled for the first quarter of calendar year 2004. NASA will benchmark other agencies as well as private industry to identify best practices with respect to developing performance measures related to the effectiveness of controls to prevent software piracy. The benchmarking activity will be completed by September 30, 2004. 5. Track actual-versus-expected performance for the Agency's Enterprisewide IT performance measures in its IRM plan; and develop a mechanism for benchmarking the Agency's IT management processes, when appropriate. NASA Response to GAO Recommendation 5: Partially Concur. NASA has established performance measures for major IT investments that are identified in the Agency's Exhibit 300 submissions to OMB and have included performance measures in our IRM Strategic Plan. We acknowledge the need to expand our benchmarking efforts. 6. Revise the Agency's IT investment management policy and guidance to describe the relationship of this process to the Agency's Enterprise architecture; NASA Response to GAO Recommendation 6: Partially Concur. The NASA IT investment management policy is contained in NPD 2800.1, Managing Information Technology. The NASA IT Capital Planning and Investment Control (CPIC)Process document describes the process NASA will use for ensuring that all IT capital investments align with the Agency's mission, Enterprise architecture, and business needs. NASA will ensure that the planned revision of NPD 2800.1, referenced in (4) above, clarifies the relationship between these two documents and the Agency's Enterprise architecture. This revision is scheduled for the first quarter of calendar year 2004. 7. Include interfaces in its inventory of the Agency's major information systems, implement a standard, documented procedure to maintain this inventory, and develop a mechanism to use the inventory as part of managerial decisionmaking; NASA Response to GAO Recommendation 7: Concur. NASA has started the process of replacing our existing asset management system, and the NASA CIO has provided IT asset-tracking requirements to the project responsible for implementing this new system. In addition, the CIO has established a database that captures all NASA systems and has begun to integrate the use of these data into our management processes to facilitate decisionmaking. We will be integrating our existing database with the new asset management system once that system is deployed. 8. Within the Agency's IT investment selection process, implement a mechanism to identify possible conflicting, overlapping, strategically unlinked or redundant proposals, implement a scoring model, and develop a prioritized list of investments; NASA Response to GAO Recommendation 8: Concur. The NASA CPIC Process document referenced in (6) above describes the selection phase of the process NASA will use. The selection phase is divided into two stages- -Concept Screening and Business Case Development and Screening. The Concept Screening stage permits the evaluation of a number of candidate investments with a minimum amount of information. If the investment is deemed viable after this initial screening, then it will move into the business case development and screening stage where it will undergo the development of a full business case and project plan. Investments that pass the initial screening will be rated and ranked against other proposed investments and in the context of ongoing projects. As described in the CPIC process, a portfolio analysis will be conducted for rating and ranking each investment in the context of a total investment portfolio. NASA will have completed this process for the FY 2006 budget submission by September 30, 2004. 9. Document the role, responsibility, and authority of its IT investment management boards, including work processes, alignment and coordination of decisionmaking among its various boards and processes for controlling and evaluating IT investments, such as those in practices 2.15, 2.16, 2.17, and 2.18. NASA Response to GAO Recommendation 9: Concur. NASA will ensure that the role, responsibility, and authority of its IT investment management board(s) in the referenced practices (2.15, 2.16, 2.17, and 2.18) is clarified in the revision of NPD 2800.1, Managing Information Technology, referenced in (4) and (6) above. This revision is scheduled for the first quarter of calendar year 2004. The following are GAO's comments on the National Aeronautics and Space Administration's (NASA) letter dated December 8, 2003. GAO Comments: 1. Our practice dealt with enterprise-level measures, not project- specific measures. In addition, although we agree that NASA's IRM plan included performance measures, the agency generally does not track actual-versus-expected performance for these enterprisewide measures. [End of section] Appendix XIV: Comments from the Nuclear Regulatory Commission: UNITED STATES NUCLEAR REGULATORY COMMISSION: WASHINGTON, D.C. 20555-0001: December 5, 2003: Mr. David A. Powner, Director: Information Technology Management Issues: United States General Accounting Office: 441 G Street, NW: Washington, D.C. 20548: Dear Mr. Powner: I would like to thank you for the opportunity to review and submit comments on the draft report, "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can be Further Improved," (GAO-04-49). The U.S. Nuclear Regulatory Commission appreciates the time and effort that you and your staff have taken to review this important topic. Overall, this report provides useful information on Federal agencies' use of 12 IT strategic planning/performance measurement practices and 18 IT investment management practices. We agree that these practices are important for ensuring effective use of government IT investments, and we support the effort to encourage best practices across Federal agencies. We have no specific comments on the report. Should you have any questions, please contact either Mr. William Dean at 301-415-1703 or Ms. Melinda Malloy at 301-415-1785 of my staff. Sincerely, Signed by: William D. Travers: Executive Director for Operations: cc: Linda J. Lambert, GAO: [End of section] Appendix XV: Comments from the Social Security Administration: SOCIAL SECURITY: The Commissioner: December 3, 2003: Mr. David A. Powner: Director, Information Technology Management Issues: U.S. General Accounting Office, Washington, D.C. 20548: Room 4075: Dear Mr. Powner: Thank you for the opportunity to review the draft report, "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (GAO- 04-49). Our comments are enclosed. If you have any questions, please have your staff contact Mark Welch at (410) 965-0374. Sincerely, Signed by: Jo Anne B. Barnhart: Enclosure: COMMENTS OF THE SOCIAL SECURITY ADMINISTRATION ON THE GENERAL ACCOUNTING OFFICE'S (GAO) DRAFT REPORT "INFORMATION TECHNOLOGY MANAGEMENT: GOVERNMENTWIDE STRATEGIC PLANNING PERFORMANCE MEASUREMENT, AND INVESTMENT MANAGEMENT CAN BE FURTHER IMPROVED" (GAO-04-49): Thank you for the opportunity to review and provide comments on this GAO draft report. The Social Security Administration (SSA) has taken many positive actions in recent years, and continues efforts, to fully comply with federal guidance relating to information technology (IT) management and improve overall performance in this area. We are pleased that this GAO report acknowledges the numerous IT management practices fully and partially in place at SSA, and that our performance in this area compares quite favorably with that of other federal agencies. Recommendation I: Include in its annual performance plan (APP) the resources and time periods required to implement the information security program plans required by the Federal Information Security Management Act of 2002 (FISMA). Comment: We agree in part. We are concerned about including the details of our information security program, including information about the resources and time periods required for implementation in a public document, because such information might aid and abet those seeking to compromise the Agency's information security. However, we recognize that Section 3544(d) of the FISMA calls for agencies to include as part of their APP, which are public documents, the time periods and resources that are necessary to implement their information security programs. Therefore, we will include this information in future APPS in a manner that will not risk compromise of the Agency's information security. Recommendation 2: Develop measures related to the performance of the Agency's IT programs and the effectiveness of controls to prevent software piracy. Comment: We agree and we believe the Agency already has effective measures in place to ensure effective SSA IT program performance. SSA's Strategic Information Resources Management Plan and associated IT programs are driven by the Agency's strategic goals and objectives which are defined in the Agency Strategic Plan and APP. IT programs provide the automation support required for the projects that help SSA to achieve those goals and objectives. At the tactical level, the Control Phase of the Agency's Capital Planning and Investment Control Process provides the appropriate oversight process for IT initiatives, including measuring compliance with the cost, schedule and performance goals established for these IT initiatives. SSA also has availability, stability and performance measures for many components of its enterprise IT architecture. Regarding software piracy, SSA currently tests the effectiveness of controls to prevent software piracy. Since the performance measures in the APP are at a high level and are focused on the four Agency goals, it seems inappropriate to include a goal focused at the tactical level of preventing software piracy. This issue is more appropriately addressed in the oversight provided in the Control Phase of the Agency's Capital Planning and Investment Control Process. Recommendation 3: Develop a mechanism for benchmarking the agency's IT management processes, when appropriate. Comment: We agree that the use of benchmarking of IT management processes is useful in certain situations. However, we do not agree with any recommendation that requires the Agency to develop a mechanism (that is, an explicit strategy or policy) for benchmarking its IT management processes. We are in compliance with the Clinger/Cohen Act, Section 5123(4) requirement that "the head of an executive agency shall, where comparable processes and organizations in the public or private sectors exist, quantitatively benchmark agency process performance against such processes in terms of cost, speed, productivity, and quality of outputs and outcomes." SSA performs benchmarking, when appropriate, and has provided GAO a list of Agency benchmarking activities carried out in response to the Clinger/Cohen Act requirement. Recommendation 4: Develop work processes and procedures for the agency's investment management board. Comment: We agree and SSA's investment management board (the Information Technology Advisory Board - ITAB) already follows established work processes and procedures. The role of the ITAB is evolving as we explore ways to refine and improve our Capital Planning and Investment Control Process to ensure continued compliance with legislation and Office of Management and Budget (OMB) guidance. While high-level documentation concerning the ITAB's operations is in place, we are working to more fully document ITAB's work processes and procedures. Recommendation 5: Establish a policy requiring modularized IT investments. Comment: We agree and SSA has already established and is following a modularized investment policy. SSA's current review of business cases includes an assessment of major projects against the "Raines Rules" criteria that include "modularized" system implementations. The Raines Rules criteria are incorporated into guidance in OMB Circulars No. A-11 and A-130. The provisions of these circulars are incorporated by reference in SSA's Capital Planning and Investment Control Process. SSA software development initiatives follow a modular, release-oriented strategy and major IT technology deployments are performed through a phased- implementation approach. Each IT investment is assessed to determine whether it is worth doing before approval is given to proceed. Any problems with the results from the implementations of earlier "modules" of a system are considered in the approval process for continued investments. Recommendation 6: Document the role, responsibility, and authority of its IT investment management board for the oversight of IT investments, such as what is outlined in practices 2.15, 2.16, and 2.18. Comment: We agree. The role, responsibility and authority of SSA's investment management board (the ITAB) for the oversight of IT investments is documented at a high level. As noted above, the role of the ITAB is evolving as we explore ways to refine and improve our Capital Planning and Investment Control Process to ensure continued compliance with legislation and OMB guidance. We are working to ensure that the ITAB's role, responsibility and authority for the oversight of IT investments are more explicitly documented. Recommendation 7: Require that corrective actions be tracked, and reported to the investment management board for under performing projects. Comment: We agree. SSA tracks the progress of all IT projects, including both projects that are proceeding as planned, and those that are not. Any surfacing problems are: separately tracked and monitored within the Risk Identification and Mitigation System and: Problem and Issues Reporting System. These repositories collect and track risk/problem information and the corrective actions taken, which are reported upon at various points along a project's lifecycle. The ITAB has access to this information. Therefore, we will more explicitly document how corrective actions for under-performing projects are tracked and reported to the ITAB. The following are GAO's comments on the Social Security Administration's (SSA) letter dated December 3, 2003. GAO Comments: 1. We agree that SSA needs to consider the level of detail that is appropriate to include in its performance plans so as not to compromise security. 2. We requested documentation to support SSA's assertion that it has performance measures associated with the performance of IT programs (e.g., the percentage of IT projects that are meeting cost, schedule, and performance goals), but none were provided. Accordingly, we did not modify our report. 3. We agree that it is not appropriate to include measures related to the effectiveness of controls to prevent software piracy in agency performance plans. Neither our practice nor our recommendation specifies the document or process that should be used to address software piracy. 4. As we noted in this report, SSA performs benchmarking in an ad hoc manner. We believe that taking a more systematic approach is necessary to ensure that benchmarking is performed at suitable times using an appropriate methodology. Without a systematic approach, it is not possible to validate that the agency performs benchmarking "when appropriate." Accordingly, we did not modify our report. 5. References to OMB's Circular A-11 in agency policy documentation alone do not ensure that these practices are met. In particular, we believe that agency policies related to modularized IT investments should be explicit and that it is neither prudent nor practical to rely on users of SSA's documentation of its capital planning and investment control process to review a secondary source. [End of section] Appendix XVI: Comments from the Department of State: United States Department of State: Assistant Secretary and Chief Financial Officer: Washington, D. C 20520: DEC 9 2003: Dear Ms. Westin: We appreciate the opportunity to review your draft report, "INFORMATION TECHNOLOGY MANAGEMENT: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved," GAO- 04-49, GAO Job Code 310448. The enclosed Department of State comments are provided for incorporation with this letter as an appendix to the final report. If you have any questions concerning this response, please contact Karen Mummaw, Computer Specialist, Bureau of Information Resource Management at (202) 663-1415. Sincerely, Signed by: Christopher B. Burnham: Enclosure: As stated. cc: GAO/IT - David Powner; State/OIG - Luther Atkins; State/IRM - Bruce Morrison; State/H - Paul Kelly: Ms. Susan S. Westin, Managing Director, International Affairs and Trade, U.S. General Accounting Office. Department of State Comments on GAO Draft Report: Information Technology Management, Government-wide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved (GAO-04-49, GAO Job Code 310448): Thank you for the opportunity to review and comment on your draft report "Information Technology Management, Government-wide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved". In general, the findings of the report are consistent with the discussions between GAO staff and our Bureau of Information Resource Management staff in September 2003. However, we have made progress in several of the areas since those discussions took place. So that GAO might have the opportunity to reflect the most current information in its final report, we want to bring the following information to your attention: * Practice Area 2.13, pg. 49 - Agency has established a structured selection process. GAO scored this item as partial for the Department and commented that the Department's documentation of its selection process is in draft form. On September 25, 2003, the Department published the final version of the Department's Capital Planning and Investment Control (CPIC) Program Guide on the IT Planning Division's web page on the Department's Intranet. This guide contains the framework for the Department's information technology CPIC process. Section 4 of the guide contains detailed procedures for the Select phase and addresses process, prerequisites and data requirements, scoring criteria, and outputs. The Department suggests that GAO consider changing the score to a "yes" for this practice area and amend the comment section accordingly. (See pp. 11-14 of the final CPIC guide that is attached.): * Practice Area 2.16, pg. 51 - Agency has established an oversight mechanism of funded investments. GAO scored this item as partial for the Department and commented that State had draft procedures for control phase reviews in place, but they are not fully implemented. Section 5 of the Department's final CPIC Guide, published on the Department's Intranet, contains detailed procedures for the Control phase of the process. The Department has begun the process of implementing these procedures for controlling IT investments, which address the initiative review process, portfolio review process, prerequisites and data requirements, scoring criteria, and outputs. (See pp. 16-21 of the CPIC Guide attached.) The Department suggests that GAO amend the comments section for this practice area to state that the Department recently issued final procedures for control phase reviews, which are not yet fully implemented. Practice Area 2.17, pg. 51 - Corrective actions for projects are agreed upon, documented, and tracked by the agency wide investment management board. GAO scored this item as partial for the Department and commented that State had draft procedures for control phase reviews in place, but they are not fully implemented. See our comment for Practice Area 2.16. The Department suggests that GAO amend the comments section for this practice area to state that the Department recently issued final procedures for control phase reviews, which are not yet fully implemented. * Practice Area 2.18, pg. 52 - Agency-wide investment management board required that post-implementation reviews be conducted. GAO scored this item as partial for the Department and commented that State has a policy related to this practice, but did not provide evidence that it is completely implemented. Section 6.2 of the Department's final CPIC Guide provides for post- implementation reviews of IT investments and sets forth the prerequisites and data requirements for such reviews. (See pp. 24-25.) The Department suggests that GAO amend the comments section for this practice area to state that the Department recently issued final procedures for conducting post implementation reviews, which are not yet fully implemented. The following are GAO's comments on the Department of State's letter dated December 9, 2003. GAO Comments: 1. We based our evaluation on the agency's draft Capital Planning and Investment Control Program Guide that was provided during our review. However, based on State's newly finalized Capital Planning and Investment Control Program Guide, we changed this evaluation in our report. 2. We based our evaluation on the agency's draft Capital Planning and Investment Control Program Guide that was provided at the time of our review. Based on the final version of the Capital Planning and Investment Control Program Guide provided by State in its response, we modified the language in our report, as appropriate. 3. See comment 2. 4. See comment 2. [End of section] Appendix XVII: Comments from the U.S. Agency for International Development: USAID: DEC 9 2003: U.S. AGENCY FOR INTERNATIONAL DEVELOPMENT: Mr. David A. Powner: Director: Information Technology Management Issues: U.S. General Accounting Office: 441 G Street, N.W. Washington, DC 20548: Dear Mr. Powner: I am pleased to provide the U.S. Agency for International Development's (USAID's) formal response on the draft GAO report entitled "Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved" (November 2003.): We have reviewed the draft recommendations to USAID and appreciate the thorough review which has been performed by your team. We do take exception to your assessment of USAID's compliance at the "No" level with the following practices: l. Practice 2.11 (page 48): "The agency requires that the proposed IT investment, at a minimum: * support work processes that it has simplified or redesigned to reduce costs and improve effectiveness, and: * make maximum use of commercial-off-the-shelf (COTS) software.": 2. Practice 2.14 (page 49): "Agency policy calls for investments to be modularized (e.g., managed and procured in well-defined useful segments or modules that are short in duration and small in score) to the maximum extent achievable.": We believe that USAID implements both practices at a \\partial" level because we require that USAID IT investment requests be submitted in accordance with OMB Circular A-11 which contains such policy requirements. For example, USAID's Automated Directives System (ADS) Chapter 577 (Capital Planning and Investment Control), Table l, Investment Category Documentation and Review Requirements, states that OMB Exhibit 300 must be submitted to the CIO requesting approval of Level II and Level III investments, which are defined on that table. Authors of Level I Investment Decision Requests (USAID 3001) must follow instructions provided in OMB Exhibit 300 in OMB Circular A-11. OMB Circular A-11 states that, "Agencies must develop, implement and use a capital programming process to develop their capital asset portfolio, and must: ... * Simplify or otherwise redesign work processes to reduce costs, improve effectiveness, and make maximum use of commercial services and off-the-shelf technology; ... * Structure major acquisitions into useful segments with a narrow scope and brief duration... ..": (ADS 577 is available on the USAID website at: http://www.usaid.gov/ policy/ads/500/577.pdf. Table 1 is available at: http://www.usaid.gov/policy/ads/500/577maa.doc. The Action Memorandum is available at: http://www.usaid.gov/policy/ads/500/577mag.pdf and OMB Circular A-11, Part 7; page p. 300-8 is available at: http://www.cio.gov/documents/ s300.pdf.): In addition, USAID evaluates and scores IT investment proposals, in part, according to the following criteria: - whether or not the project includes the use of commercial off-the-shelf (COTS) solutions and minimizes the use of customer-designed components and - whether the Agency uses phased successive chunks. (See ADS 577, Table 3, page 3, item 2.2, which is available on the USAID website at: http:// www.usaid.gov/policy/ads/ 500/577mab.pdf.): Thank you for the opportunity to respond to the GAO draft report and for the courtesies extended by your staff in the conduct of this review. If you have any questions, you may contact me or contact Mark Kneidinger, Deputy CIO for Policy and Deputy Assistant Administrator for Management at (202)712-1200. Sincerely, Signed by: John Marshall: Assistant Administrator: Bureau for Management: The following are GAO's comments on the U.S. Agency for International Development's (USAID) letter dated December 9, 2003. GAO Comments: 1. References to OMB's Circular A-11 in agency policy documentation alone do not ensure that these practices are met. In particular, we believe that agency policies related to practices 2.11 and 2.14 should be explicit and that it is neither prudent nor practical to rely on users of USAID's directives to review a secondary source. Regarding USAID's comments that it uses the criteria in practices 2.11 and 2.14 as part of its evaluation and scoring of investments, we agree that the agency does ask some questions on the use of commercial-off-the-shelf software and whether the agency uses "successive chunks" within its proposed IT investment scoring model. However, addressing these criteria as part of a scoring model does not address our practice because scoring projects on the basis of the questions asked does not necessarily preclude projects from continuing if they do not fully meet the criteria. Additionally, the questions asked as part of the scoring model do not fully meet the requirements of the practices. Accordingly, we did not modify our report. [End of section] Appendix XVIII: Comments from the Department of Veterans Affairs: THE SECRETARY OF VETERANS AFFAIRS: WASHINGTON: December 5, 2003: Mr. David A. Powner: Director: Information Technology Management Issues: U.S. General Accounting Office: 441 G Street, NW: Washington, DC 20548: Dear Mr. Powner: The Department of Veterans Affairs (VA) has reviewed your draft report, Information Technology Management: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved (GAO 04-49). VA concurs with all seven of GAO's recommendations. I am pleased to advise that VA has already implemented five recommendations and plans implementing the remaining two recommendations by the end of April 2004. VA's detailed comments specific to the report's recommendations are outlined in the enclosure. Thank you for the opportunity to comment on your draft report. Sincerely yours, Signed by: Anthony J. Principi: Enclosure: Enclosure: Department of Veterans Affairs Comments To GAO Draft Report, INFORMATION TECHNOLOGY MANAGEMENT: Governmentwide Strategic Planning, Performance Measurement, and Investment Management Can Be Further Improved (GAO-04-49): To improve the department's IT strategic management/performance measurement processes, GAO recommends that the Secretary of Veterans Affairs: * Include in the department's annual performance plan the resources required to implement the information security program plan required by FISMA. Concur: The Department currently collects information on the costs to implement the provision of Federal Information Security Management Act (FISMA) of 2002 through several inter-related efforts and reports this information to the Office of Management and Budget. This information will be included in future Government Performance and Results Act submissions. * Develop a documented process to measure progress against the department's IT goals and assign goals and responsibilities for achieving these goals. Concur: VA's Acting Chief Information Officer (CIO) is revising the Department's IT Strategic Plan. The revision will include a section that outlines the performance measures that are associated with specific IT goals and objectives and which organizations are responsible for carrying them out. Progress against these performance measures will be assessed as part of the monthly performance reviews that all projects within the IT Portfolio must undergo. The process for measuring this performance will be documented in guidance being prepared to address IT Portfolio and Project Management and will be developed by April 2004. The IT Strategic Plan update will be developed by February 2004. VA anticipates implementing this recommendation by the end of April 2004. * Develop measures related to the effectiveness of controls to prevent software piracy. Concur: In fiscal year 2004, VA will field its Department-wide Security Configuration and Management Program (SCAMP). SCAMP will establish and deploy an enterprise-level, network configuration management framework capability to centrally manage all desktops, servers, and communications and security devices in the VA environment. This initiative will greatly enhance VA's existing capabilities to preclude the installation of any unauthorized (including pirated) software on VA networks and desktops, and should provide data to measure the effectiveness of the controls. * Track actual-versus-expected performance for the department's enterprisewide IT performance measures in its IRM plan. Concur: Monthly performance reports are provided to VA's Strategic Management Council for all OMB Exhibit 300-level projects and programs actual-versus-expected performance metrics including: * Acquisition requirements: * Funding: * Staffing: * Schedule performance: * Budget performance: * Quality performance: The Department's Strategic Management Council is chaired by the Deputy Secretary and is comprised of senior officers from the Department's administrations and staff offices. This recommendation has been implemented. To improve the department's IT investment management process, GAO recommends that the Secretary of Veterans Affairs: Document the alignment and coordination of responsibilities of the department's various IT investment management boards for decision making related to IT investments, including cross-cutting investment(s). Concur: Within VA, the sole board responsible for the overall governance of IT decision making, as it is related to the IT Portfolio of VA investments, is the Enterprise Information Board (EIB). The EIB is chaired by the VA Chief Information Officer and includes within its membership the senior IT leadership of the Department, as well as fiscal officer representation. The EIB uses the Integrated Management Process as its tool for managing IT investments. The Integrated Management Process ensures appropriate planning, technical, and fiscal review at key decision points ("milestone") within a project's life cycle. Projects must fully complete all requirements for a given milestone before they are permitted unqualified approval to proceed to the next one. There are five milestones in the life cycle of a project: (0) Project Concept Development; (1) Project Planning; (2) Project Pilot/Prototyping; (3) Project Development/Roll-Out; and (4) Project Operation (i.e., "Post-Implementation"). VA constituent organizations (administrations and staff offices) are free to develop appropriate decision making mechanisms to vet investments and investment progress decisions prior to their review by the EIB. Nevertheless, those boards are still subject to the overall governance of the EIB. Detailed guidance on the Integrated Management Process, IT Portfolio Management, and IT Project Management is being prepared and will be completed by April 2004. VA anticipates implementing this recommendation by the end of April 2004. Within the agency's IT investment selection process, implement a mechanism to identify possible conflicting, overlapping, strategically unlinked or redundant proposals and prioritize its IT investment(s). Concur: VA has already implemented an IT Capital Planning and Investment Control (CPIC) process that fully implements the requirement of the Office of Management and Budget (OMB), as defined within OMB Circular A-11, Section 300. A key component of the CPIC process is the annual review of the entire IT Portfolio through the analyses of Exhibit 300 - Capital Asset Plan and Business Case - for each of the Department's major IT investments. VA has fully aligned its review process with that of OMB to the extent that internal evaluation of Exhibit 300s (prior to their dispatch to OMB) uses the same scoring template and analytical paradigm as used by OMB. All Exhibit 300 documents are reviewed by a small group of technical subject matter experts. Through such means, it becomes possible to identify those investments that may be in conflict with others, potentially offer duplicative or overlapping efforts, or not advance the mission, goals, and objectives of the Department (i.e., "strategically unlinked"). Based on this review, the EIB is offered recommendations on the IT Portfolio. The EIB, through its capability to decide whether to recommend continuance, modification, or termination of projects to the VA CIO, can then resolve conflicting, overlapping, or non-aligned investment proposals. This recommendation has been implemented. * Develop a process to use independent verification and validation reviews, when appropriate. Concur: As part of VA's Integrated Management Process, the VA CIO, or the EIB, can request "in process" reviews of investments during their life cycle, particularly if such an investment appears "off track" (that is, cost and/or schedule is no longer within planned values plus 10 percent, or the project is not delivering the performance that was expected). In addition, after implementation, projects are subject to post-implementation reviews. This latter set of reviews determines whether a project, now fully implemented, provided the Department with what was initially expected. These reviews are conducted by someone in other than the office responsible for the investment and often by outside consultants. This recommendation has been implemented. The following are GAO's comments on the Department of Veterans Affairs' (VA) letter dated December 5, 2003. GAO Comments: 1. VA's response indicates that the department will address this recommendation in the future and, therefore, we did not remove this recommendation. 2. See comment 1. 3. See comment 1. 4. VA's monthly performance reports track project-specific measures, not enterprisewide IT performance measures. VA's draft IRM plan states that it will establish metrics to measure performance for IT strategic initiatives. However, progress toward doing so was not addressed by VA in its comments. Therefore, we do not believe this recommendation has been fully addressed. 5. See comment 1. 6. Although VA describes a process followed for reviewing investment proposals, it did not provide evidence to support that this practice was actually followed. In addition, VA did not address the element of our recommendation related to prioritizing its IT investments. Therefore, we did not remove this recommendation. 7. On the basis of the additional information provided, we agree that the recommendation has been implemented and modified our report accordingly. [End of section] Appendix XIX: GAO Contacts and Staff Acknowledgments: GAO Contacts: Linda J. Lambert, (202) 512-9556 Mark D. Shaw, (202) 512-6251: Staff Acknowledgments: Joseph P. Cruz, Lester P. Diamond, Laurence P. Gill, David B. Hinchman, Robert G. Kershaw, David F. Plocher, Susan S. Sato, and Patricia D. Slocum made key contributions to this report. (310448): FOOTNOTES [1] Office of Management and Budget, Budget of the U.S. Government, Fiscal Year 2004, Report on IT Spending for the Federal Government for Fiscal Years 2002, 2003, and 2004. We did not verify these data. [2] U.S. General Accounting Office, High-Risk Series: An Update, GAO- 03-119 (Washington, D.C. January 2003) and Major Management Challenges and Program Risks: A Governmentwide Perspective, GAO-03-95 (Washington, D.C. January 2003). [3] The Paperwork Reduction Act of 1995 revised the information resources management responsibilities established under the Paperwork Reduction Act of 1980, as amended in 1986. [4] Office of Management and Budget, Circular A-11, Preparation, Submission, and Execution of the Budget (July 27, 2002) and Circular A- 130, Management of Federal Information Resources (Nov. 28, 2000). [5] U.S. General Accounting Office, Information Technology Investment Management: A Framework for Assessing and Improving Process Maturity, GAO/AIMD-10.1.23, Exposure Draft (Washington, D.C. May 2000). [6] This section of the United States Code requires 24 departments and agencies to establish chief financial officers. We did not include the Federal Emergency Management Agency in our review, even though it is one of the 24 departments and agencies, because this agency has been transferred to the Department of Homeland Security. [7] Six percent were not applicable. [8] One percent were not applicable. Percentages do not add up to 100 percent due to rounding. [9] DOD submitted a single letter that included comments from the Departments of the Air Force, Army, and Navy. [10] This responsibility is in addition to OMB's role in assisting the President in reviewing agency budget submissions and compiling the President's budget, as discussed in 31 U.S.C. chapter 11. [11] Government Information Security Reform, Title X, Subtitle G, Floyd D. Spence National Defense Authorization Act for Fiscal Year 2001, P.L. 106-398, Oct. 30, 2000. [12] GAO-03-95. [13] U.S. General Accounting Office, Electronic Government: Selection and Implementation of the Office of Management and Budget's 24 Initiatives, GAO-03-229 (Washington, D.C. Nov. 22, 2002). [14] U.S. General Accounting Office, Information Management: Challenges in Managing and Preserving Electronic Records, GAO-02-586 (Washington, D.C. June 17, 2002). [15] U.S. General Accounting Office, Electronic Records: Management and Preservation Pose Challenges, GAO-03-936T (Washington, D.C. July 8, 2003). [16] U.S. General Accounting Office, Privacy Act: OMB Leadership Needed to Improve Agency Compliance, GAO-03-304 (Washington, D.C. June 30, 2003). [17] U.S. General Accounting Office, Information Security: Opportunities for Improved OMB Oversight of Agency Practices, GAO/AIMD- 96-110 (Washington, D.C. Sept. 24, 1996). [18] U.S. General Accounting Office, Computer Security: Progress Made, but Critical Federal Operations and Assets Remain at Risk, GAO-03-303T (Washington, D.C. Nov. 19, 2002). [19] U.S. General Accounting Office, Information Security: Continued Efforts Needed to Fully Implement Statutory Requirements, GAO-03-852T (Washington, D.C. June 24, 2003). [20] U.S. General Accounting Office, High-Risk Series: Protecting Information Systems Supporting the Federal Government and the Nation's Critical Infrastructures, GAO-03-121 (Washington, D.C. January 2003). [21] U.S. General Accounting Office, Information Technology: Leadership Remains Key to Agencies Making Progress on Enterprise Architecture Efforts, GAO-04-40 (Washington, D.C. Nov. 17, 2003) and Information Technology: Enterprise Architecture Use across the Federal Government Can Be Improved, GAO-02-6 (Washington, D.C. Feb. 19, 2002). [22] For example, see U.S. General Accounting Office, Information Technology: Inconsistent Software Acquisition Processes at the Defense Logistics Agency Increase Project Risks, GAO-02-9 (Washington, D.C. Jan. 10, 2002); and HUD Information Systems: Immature Software Acquisition Capability Increases Project Risks, GAO-01-962 (Washington, D.C. Sept. 14, 2001). [23] U.S. General Accounting Office, Information Technology: Leading Commercial Practices for Outsourcing of Services, GAO-02-214 (Washington, D.C. Nov. 30, 2001). [24] U.S. General Accounting Office, Information Technology Investment Management: A Framework for Assessing and Improving Process Maturity, Exposure Draft, GAO/AIMD-10.1.23 (Washington, D.C. May 2000). [25] For example, U.S. General Accounting Office, Information Technology: Departmental Leadership Crucial to Success of Investment Reforms at Interior, GAO-03-1028 (Washington, D.C. Sept. 12, 2003); Bureau of Land Management: Plan Needed to Sustain Progress in Establishing IT Investment Management Capabilities, GAO-03-1025 (Washington, D.C. Sept. 12, 2003); United States Postal Service: Opportunities to Strengthen IT Investment Management Capabilities, GAO- 03-3 (Washington, D.C. Oct. 15, 2002); Information Technology: DLA Needs to Strengthen Its Investment Management Capability, GAO-02-314 (Washington, D.C. Mar. 15, 2002); and Information Technology: INS Needs to Strengthen Its Investment Management Capability, GAO-01-146 (Washington, D.C. Dec. 29, 2000). [26] Office of Management and Budget, Circular A-11, Preparation, Submission, and Execution of the Budget (July 2002) and Circular A-130, Management of Federal Information Resources (Nov. 30, 2000). [27] U.S. General Accounting Office, Information Technology Investment Management: A Framework for Assessing and Improving Process Maturity, GAO/AIMD-10.1.23, Exposure Draft (Washington, D.C. May 2000). [28] The Departments of Agriculture, the Air Force, the Army, Commerce, Defense, Education, Energy, Health and Human Services, Housing and Urban Development, the Interior, Justice, Labor, the Navy, State, Transportation, the Treasury, and Veterans Affairs; and the Environmental Protection Agency, General Services Administration, National Aeronautics and Space Administration, National Science Foundation, Nuclear Regulatory Commission, Office of Personnel Management, Small Business Administration, Social Security Administration, and U.S. Agency for International Development. [29] Each year, OMB's Office of Information and Regulatory Affairs publishes an Information Collection Budget by gathering data from executive branch agencies on the total number of burden hours it approved for collection of information at the end of the fiscal year and agency estimates of the burden for the coming fiscal year. [30] In fulfilling its responsibilities under this act, OMB requires agencies to report to OMB on their plans for providing the public with the option of submitting, maintaining, and disclosing required information electronically, instead of on paper. [31] U.S. General Accounting Office, Information Resources Management: Comprehensive Strategic Plan Needed to Address Mounting Challenges, GAO-02-292 (Washington, D.C. Feb. 22, 2002). [32] U.S. General Accounting Office, Paperwork Reduction Act: Record Increase in Agencies' Burden Estimates, GAO-03-619T (Washington, D.C. April 11, 2003). [33] U.S. General Accounting Office, Executive Guide: Measuring Performance and Demonstrating Results of Information Technology Investments, GAO/AIMD-98-89 (Washington, D.C. March 1998). [34] GAO/AIMD-98-89. [35] GAO/AIMD-98-89. [36] VA included the time periods to implement its information security program in its performance plan. [37] GAO/AIMD-10.1.23 and U.S. General Accounting Office, Assessing Risks and Returns: A Guide for Evaluating Federal Agencies' IT Investment Decision-making, GAO/AIMD-10.1.13 (Washington, D.C. February 1997). [38] We have previously reported that certain DOD system acquisitions were not utilizing incremental management best practices or were just beginning to do so. For example, see U.S. General Accounting Office, DOD Systems Modernization: Continued Investment in Standard Procurement System Has Not Been Justified, GAO-01-682 (Washington, D.C. July 31, 2001). [39] GAO-03-1028. [40] With a scoring model, the assessment body typically attaches numerical scores and "relative value" weights to each of the individual selection criteria. Investments are then assessed relative to these scores and then against weights associated with each individual criterion. Finally, the weighted scores are summed to create a numerical value for each investment. [41] U.S. General Accounting Office, DOD Business Systems Modernization: Improvements to Enterprise Architecture Development and Implementation Efforts Needed, GAO-03-458 (Washington, D.C. Feb. 28, 2003). [42] U.S. General Accounting Office, DOD Business Systems Modernization: Important Progress Made to Develop Business Enterprise Architecture, but Much Work Remains, GAO-03-1018 (Washington, D.C. Sept. 19, 2003). [43] The technical review board, which serves as the department's first-tier investment review board, is chaired by the deputy CIO, and its members consist of IRM managers and administrative officers from each component. [44] DOD submitted a single letter that included comments from the Departments of the Air Force, Army, and Navy. [45] AIMD-10-1-23. [46] The CIO Council is the principal interagency forum for improving agency practices related to the design, acquisition, development, modernization, use, operation, sharing, and performance of federal government information resources. GAO's Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.