Results-Oriented Management

Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information Gao ID: GAO-09-676 August 17, 2009

Since 1997, periodic GAO surveys indicate that overall, federal managers have more performance information available but have not made any greater use of this information for decision making. Based on GAO's most recent survey in 2007, GAO was asked to (1) identify agencies with relatively low use of performance information and the factors that contribute to this condition; and (2) examine practices in an agency with indications of improvement in use of performance information. GAO analyzed results from its surveys of federal managers across 29 agencies, reviewed key agency documents related to using performance information--such as Performance and Accountability Reports--and interviewed agency and selected subunit managers about their management practices. GAO also compared management practices, at selected agencies with those GAO has identified as promoting the use of performance information for decision making.

According to GAO's 2007 survey of federal managers on their use of performance information for decision making, the Federal Emergency Management Agency (FEMA) and the Department of the Interior (Interior), ranked 27 and 28 out of 29 agencies. Several factors contributed to this relatively low use. At both FEMA and Interior, the demonstrated commitment of agency leaders to using performance information--a key management practice--was inconsistent. While some FEMA programs and regions encouraged use of performance information to plan for and respond to unpredictable events, others expressed uncertainty as to how they could use performance information in the face of uncontrollable external factors. FEMA managers were also hampered by weak alignment among agency, program, and individual goals, as well as limited analytic capacity to make use of performance information. At Interior and the National Park Service (NPS), managers reported a proliferation of measures, including some that, while meaningful for department-level accountability, were not relevant to their day-to-day management. Managers at NPS and the Bureau of Reclamation also said that poorly integrated performance and management information systemscontributed to an environment where the costs of performance reporting--in terms of time and resources--outweighed what they described as minimal benefits. While both FEMA and Interior have taken some promising steps to make their performance information both useful and used, these initiatives have thus far been limited. The experience of the Centers for Medicare & Medicaid Services (CMS) highlights the role that strengthened management practices can play. According to GAO's 2000 and 2007 survey results, the percentage of managers at CMS reporting use of performance information for various management decisions increased by nearly 21 percentage points--one of the largest improvements among agencies over that period. CMS officials attributed this change to a combination of key management practices they had employed, including, but not limited to: leadership commitment to using performance information; alignment of strategic and performance goals; improving the usefulness of performance information; and building the analytic capacity to collect and use performance information.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-09-676, Results-Oriented Management: Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information This is the accessible text file for GAO report number GAO-09-676 entitled 'Results-Oriented Management: Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information' which was released on September 24, 2009. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: United States Government Accountability Office: GAO: August 2009: Results-Oriented Management: Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information: GAO-09-676: GAO Highlights: Highlights of GAO-09-676, a report to congressional requesters. Why GAO Did This Study: Since 1997, periodic GAO surveys indicate that overall, federal managers have more performance information available but have not made any greater use of this information for decision making. Based on GAO‘s most recent survey in 2007, GAO was asked to (1) identify agencies with relatively low use of performance information and the factors that contribute to this condition; and (2) examine practices in an agency with indications of improvement in use of performance information. GAO analyzed results from its surveys of federal managers across 29 agencies, reviewed key agency documents related to using performance information”such as Performance and Accountability Reports”and interviewed agency and selected subunit managers about their management practices. GAO also compared management practices, at selected agencies with those GAO has identified as promoting the use of performance information for decision making. What GAO Found: According to GAO‘s 2007 survey of federal managers on their use of performance information for decision making, the Federal Emergency Management Agency (FEMA) and the Department of the Interior (Interior), ranked 27 and 28 out of 29 agencies. Several factors contributed to this relatively low use. At both FEMA and Interior, the demonstrated commitment of agency leaders to using performance information”a key management practice”was inconsistent. While some FEMA programs and regions encouraged use of performance information to plan for and respond to unpredictable events, others expressed uncertainty as to how they could use performance information in the face of uncontrollable external factors. FEMA managers were also hampered by weak alignment among agency, program, and individual goals, as well as limited analytic capacity to make use of performance information. At Interior and the National Park Service (NPS), managers reported a proliferation of measures, including some that, while meaningful for department-level accountability, were not relevant to their day-to-day management. Managers at NPS and the Bureau of Reclamation also said that poorly integrated performance and management information systems contributed to an environment where the costs of performance reporting”in terms of time and resources”outweighed what they described as minimal benefits. While both FEMA and Interior have taken some promising steps to make their performance information both useful and used, these initiatives have thus far been limited. Figure: Survey question: [Refer to PDF for image: horizontal bar graph] My agency's top leadership demonstrates a strong commitment to using performance information to guide decision making: Percentage responding to a ’great“ or ’very great“ extent: FEMA: 32%; Interior: 37%; Rest of Government: 50%. Source: GAO. [End of figure] The experience of the Centers for Medicare & Medicaid Services (CMS) highlights the role that strengthened management practices can play. According to GAO‘s 2000 and 2007 survey results, the percentage of managers at CMS reporting use of performance information for various management decisions increased by nearly 21 percentage points”one of the largest improvements among agencies over that period. CMS officials attributed this change to a combination of key management practices they had employed, including, but not limited to: leadership commitment to using performance information; alignment of strategic and performance goals; improving the usefulness of performance information; and building the analytic capacity to collect and use performance information. What GAO Recommends: GAO is making recommendations to the Departments of Homeland Security and the Interior for improvements to key management practices to promote greater use of performance information at FEMA, NPS, Reclamation, as well as at Interior. Interior agreed in principle and DHS generally concurred, but disagreed that FEMA should develop an interim performance management plan. GAO clarified this recommendation to address the concern. View [hyperlink, http://www.gao.gov/products/GAO-09-676] or key components. For more information, contact Bernice Steinhardt at (202) 512-6806 or steinhardtb@gao.gov. [End of section] Contents: Letter: Background: FEMA and Interior Were Hindered in Using Performance Information for Decision Making by Weak or Inconsistent Application of Key Management Practices: Officials at CMS Headquarters and Selected Programs Said Key Management Practices Had Promoted Use of Performance Information for Decision Making: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Objectives, Scope, and Methodology: Appendix II: Agency Rankings Based on Index of 2007 Survey Results: Appendix III: Timeline of Major Government Results-Oriented Management Reforms: Appendix IV: Comments from the Department of Homeland Security: Appendix V: Comments from the Department of the Interior: Appendix VI: GAO Contact and Staff Acknowledgments: Related GAO Products: Figures: Figure 1: Percentage of Federal Managers Who Reported Having Performance Measures: Figure 2: Percentage of Federal Managers Who Reported Using Information Obtained from Performance Measurement for Various Management Decision- Making Functions: Figure 3: Practices That Can Promote the Use of Performance Information for Decision Making: Figure 4: Percentage of Federal Managers Who Reported That Agency's Top Leadership Demonstrated a Strong Commitment to Using Performance Information to Guide Decision Making: Figure 5: Percentage of Federal Managers Who Identified Lack of Leadership Commitment to Using Performance Information as a Hindrance to Measuring Performance or Using Performance Information: Figure 6: Percentage of Federal Managers Who Reported They Were Held Accountable for Their Agency's Accomplishments of Strategic Goals: Figure 7: Percentage of Federal Managers Who Reported Agency Investment in Performance Data Capacity: Figure 8: Percentage of Federal Managers Who Reported Top Leadership Commitment to Using Performance Information to Guide Decision Making: Figure 9: Interior Managers Reported Being Held Accountable to a Similar Extent as Rest of Government: Figure 10: Interior Managers Reported Using Performance Information to Identify Problems, Take Corrective Actions, or Develop Strategy to a Lesser Extent than Rest of Government: Figure 11: Percentage of Federal Managers Who Reported That Difficulty in Determining Meaningful Measures Hinders Using Performance Information: Figure 12: Percentage of CMS Managers Who Reported Top Leadership Demonstrated Commitment to Achieving Results: Figure 13: CMS Region IV Communicated Performance Information with Stakeholders to Improve Quality of Care in Nursing Homes: Figure 14: Percentage of CMS Managers Who Reported That Agency Managers at Their Level Are Held Accountable for the Results of Their Programs: Figure 15: CMS Reported Alignment among Department and Agency Goals and Individual Performance Objectives: Figure 16: Percentage of CMS Managers Who Reported That Difficulty Determining Meaningful Measures Hinders Using Performance Information: Figure 17: A CMS Region IV Manager Described How Easier Access to Performance Data Contributed to Improved Nursing-Home Survey Frequency in Alabama: Figure 18: Percentage of CMS Managers Who Reported That Training Was Provided to Help Accomplish Key Management Tasks: Figure 19: Average Change in Percentage of Federal Managers Reporting Use of Performance Information to a Great or Very Great Extent, 2000- 2007: Figure 20: Agency Ranking Based on 2007 Survey Results on Use of Performance Information: Abbreviations: CMS: Centers for Medicare & Medicaid Services: CQISCO: Consortium for Quality Improvement and Survey & Certification Operations: DHS: Department of Homeland Security: FEMA: Federal Emergency Management Agency: FIRM: Flood Insurance Rate Maps: GPRA: Government Performance and Results Act: Interior: Department of the Interior: NPS: National Park Service: OCSQ: Office of Clinical Standards and Quality: OMB: Office of Management and Budget: PAR: Performance and Accountability Report: PART: Program Assessment Rating Tool: PMA: President's Management Agenda: PMAP: Performance Management Appraisal Program: PMDS: Performance Management Data System: PMIS: Project Management Information System: PPS: Performance Plan System: Recovery Act: American Recovery and Reinvestment Act: SCHIP: State Children's Health Insurance Program: SES: Senior Executive Service: [End of section] United States Government Accountability Office: Washington, DC 20548: August 17, 2009: The Honorable Thomas R. Carper: Chairman: The Honorable John McCain: Ranking Member: Subcommittee on Federal Financial Management, Government Information, Federal Services, and International Security: Committee on Homeland Security and Governmental Affairs: United States Senate: The Honorable Tom Coburn: United States Senate: How the federal government performs and the results it achieves have a significant effect on many of the most pressing issues of concern to the American public--whether it be the creation of jobs by providing timely and targeted aid for recovery programs, rigorous oversight of financial markets, effective responses to natural disasters, reduction in pollutants that contribute to climate change, or delivery of water to arid regions of the country. Given increasing public demands for a more effective, transparent, and accountable federal government, it is more important than ever that federal agencies establish meaningful goals for improving performance, monitor progress in achieving their goals, and use information about performance to make decisions that can improve results. For the purposes of this report, we define performance information to mean data collected to measure progress toward achieving an agency's established mission or program-related goals. Performance information can focus on various dimensions of performance such as outcomes, outputs, quality, timeliness, customer satisfaction, or efficiency. It can inform key management decisions such as setting program priorities, allocating resources, identifying program problems and taking corrective action to solve those problems; or it can help determine progress in meeting the goals of programs or operations. Performance information may be collected to address internal management needs or external reporting requirements such as the Government Performance and Results Act of 1993 (GPRA),[Footnote 1] or the Program Assessment Rating Tool (PART), used by the Office of Management and Budget (OMB) under the previous administration.[Footnote 2] Our periodic surveys on performance and management issues since 1997 [Footnote 3] have indicated that federal managers today have significantly more performance information available for the programs they manage than they did 10 years ago. However, on the whole, federal managers have shown little or no progress in increasing their use of performance information to manage for results. While some agencies have reported significant improvements, others remain unchanged.[Footnote 4] In an effort to increase the use of performance information by agency managers, you asked that we conduct reviews at selected agencies to better understand what may hinder their use of performance information in managerial decision making and to identify opportunities for improvement. Our objectives were to: (1) identify agencies with relatively low use of performance information and the factors that contribute to this condition; and (2) examine practices in an agency where there were indications of improvement in its use of performance information. To address both of our objectives, we reviewed our prior work on results-oriented management, including key practices that can promote greater use of performance information. We also reviewed prior reports and other relevant materials on GPRA and PART. To address our first objective, we first drew on our 2007 survey results to identify agencies where relatively fewer managers reported making extensive use of performance information. Based on this ranking and other considerations, we chose the Department of the Interior (Interior) and the Federal Emergency Management Agency (FEMA), which ranked 27th and 28th respectively out of 29 agencies (see appendix II, figure 20). We then conducted interviews with senior-level officials responsible for operations, budget, human capital, and performance- reporting functions at each agency to gain an understanding of the performance-based management policies and practices established at the top levels of their organizations. We also asked officials and managers to identify areas where they faced difficulties in using performance information for decision making. To obtain the perspective of bureau, program, and field managers on challenges they faced in using performance information at their level, we interviewed officials from selected component organizations that covered significant and diverse aspects of each agency's mission. At Interior, we selected the National Park Service (NPS) and Bureau of Reclamation (Reclamation) for review; and at FEMA, we selected the Disaster Assistance and Mitigation Directorates. To address our second objective of examining practices in an agency where improvement in the managers' use of performance information appears to have progressed, we selected the Centers for Medicare & Medicaid Services (CMS). Comparing survey results from 2000 and 2007, CMS managers' reported use of performance information across selected areas of key decision making increased by nearly 21 percentage points. However, because CMS scored significantly below the rest of government in responses to survey items on managerial use of performance information in 2000, their 2007 responses to these items reflect a significant turnaround (see appendix I, figure 19). At CMS, we interviewed top headquarters officials, officials and managers in Regions IV and IX, and in two lines of business--the Consortium for Quality Improvement and Survey & Certification Operations (CQISCO) and the Consortium for Financial Management and Fee for Service Operations. It should be noted that we did not systematically assess the quality of the performance information used in the examples we cite. In addition, although we describe how performance information was used to make decisions in our examples, we did not examine whether such use ultimately resulted in improved outcomes. At all three agencies, we interviewed selected officials and managers to gauge the extent to which key management practices--that we previously reported can promote use of performance information to manage for results--had been implemented. We also reviewed agency policies, procedures, and documentation related to results-oriented management such as their strategic plans, performance measures, and individual performance-management systems. See appendix I for a more detailed discussion of our scope and methodology. We performed our work in the Washington, D.C., metropolitan area; Boston, Massachusetts; San Francisco and Sacramento, California; and Atlanta, Georgia; from March 2007 to May 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Background: Over the past 16 years, a succession of legislative reforms and executive guidance have been aimed at improving the effectiveness of federal programs by transforming the departments and agencies that administer those programs to be more results-oriented and performance- based. A key element of these reforms is the Government Performance and Results Act of 1993 (GPRA), which among other things required executive agencies to establish results-oriented goals and performance measures and report on the progress achieved. More recently, OMB created the Program Assessment Rating Tool (PART), a diagnostic tool intended to provide a consistent approach for evaluating federal programs. (See appendix III for a timeline of results-oriented-management reforms.) As we reported in July 2008,[Footnote 5] we have seen a positive transformation in the capacity of the federal government to manage for results. This capacity includes an infrastructure of outcome-oriented strategic plans, performance measures, and accountability reporting that provides a solid foundation for improving the performance of federal programs. In particular, significantly more federal managers reported to a great or very great extent having the types of performance measures called for by GPRA and PART than they did 10 years ago (see figure 1).[Footnote 6] Figure 1: Percentage of Federal Managers Who Reported Having Performance Measures: [Refer to PDF for image: multiple horizontal bar graph] Survey question: Output measures[A]: 1997, percentage responding to a ’great“ or ’very great“ extent: 37.8; 2007, percentage responding to a ’great“ or ’very great“ extent: 54.2. Survey question: Efficiency measures[A]: 1997, percentage responding to a ’great“ or ’very great“ extent: 25.9; 2007, percentage responding to a ’great“ or ’very great“ extent: 44.1. Survey question: Customer Service measures[A]: 1997, percentage responding to a ’great“ or ’very great“ extent: 31.5; 2007, percentage responding to a ’great“ or ’very great“ extent: 41.6. Survey question: Quality measures[A]: 1997, percentage responding to a ’great“ or ’very great“ extent: 30.9; 2007, percentage responding to a ’great“ or ’very great“ extent: 40.2. Survey question: Outcome measures[A]: 1997, percentage responding to a ’great“ or ’very great“ extent: 31.8; 2007, percentage responding to a ’great“ or ’very great“ extent: 48.9. Source: GAO. Notes: Data are from GAO 1997 and 2007 surveys. [A] There is a statistically significant difference between 1997 and 2007 surveys. Hereafter, the differences in percentages reported are statistically significant unless otherwise indicated. [End of figure] However, the ultimate benefit of collecting performance information-- improved decision making and results--is only fully realized when this information is used to support management planning and decision-making functions. The results of our 2007 survey showed that despite having more performance measures available, the extent to which managers make use of this information to improve performance has remained relatively unchanged. As shown in figure 2, six of the eight categories of management activities we asked about in both 2000 and 2007 showed no statistically significant change over the past 10 years: Figure 2: Percentage of Federal Managers Who Reported Using Information Obtained from Performance Measurement for Various Management Decision- Making Functions: [Refer to PDF for image: multiple horizontal bar graph] Survey question: Setting program priorities[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 65.8; 2007, percentage responding to a ’great“ or ’very great“ extent: 58.1. Survey question: Allocating resources[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 62.5; 2007, percentage responding to a ’great“ or ’very great“ extent: 59. Survey question: Adopting new program approaches or changing work processes; 1997, percentage responding to a ’great“ or ’very great“ extent: 66.1; 2007, percentage responding to a ’great“ or ’very great“ extent: 53. Survey question: Coordinating program efforts with other internal or external organizations[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 56.8; 2007, percentage responding to a ’great“ or ’very great“ extent: 50.5. Survey question: Refining program performance measures[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 51.5; 2007, percentage responding to a ’great“ or ’very great“ extent: 46.3. Survey question: Setting new or revising existing performance goals[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 58.5; 2007, percentage responding to a ’great“ or ’very great“ extent: 52.1. Survey question: Setting individual job expectations for the government employees I manage or supervise[A]; 1997, percentage responding to a ’great“ or ’very great“ extent: 60.8; 2007, percentage responding to a ’great“ or ’very great“ extent: 61.9. Survey question: Rewarding government employees I manage or supervise; 1997, percentage responding to a ’great“ or ’very great“ extent: 52.6; 2007, percentage responding to a ’great“ or ’very great“ extent: 60.9. Survey question: Developing and managing contracts[B]; 1997, percentage responding to a ’great“ or ’very great“ extent: 0; 2007, percentage responding to a ’great“ or ’very great“ extent: 40.5. Source: GAO. Notes: Data are from GAO 1997 and 2007 surveys. [A] Differences in percentages between 1997 and 2007 were not statistically significant. [B] This question was not asked in 1997. [End of figure] As our survey results showed, despite legislative and administration efforts to focus federal management decisions on maximizing the results achieved with federal funds, changing the way federal managers make decisions is not simply a matter of making more program performance information available. Based on our work on management reform efforts as well as analysis of federal managers' responses to our surveys, we have identified several key management practices that can promote the use of performance information (see figure 3).[Footnote 7] Figure 3: Practices That Can Promote the Use of Performance Information for Decision Making: [Refer to PDF for image: illustration] Practices: Demonstrating management commitment; Aligning agencywide goals, objectives, and measures; Improving the usefulness of performance information; Developing capacity to use performance information; Communicating performance information frequently and effectively. Uses: Identify problems and take corrective action; Develop strategy and allocate resources; Recognize and reward performance; Identify and share effective approaches. Practices promote uses, which result in improved results. Source: GAO. [End of figure] Our prior report grouped these practices into five categories as described below. Demonstrating Management Commitment: The commitment of agency managers to results-oriented management is critical to increased use of performance information for policy and program decisions. Demonstrating the willingness and ability to make decisions and manage programs on the basis of results, and inspiring others to embrace such a model, are important indicators of management's commitment. Management can show this type of commitment by leading frequent, regular performance-review meetings to discuss progress made toward the achievement of results, and by involving staff from different organizational levels in performance-review meetings. These methods can assist agencies in identifying performance problems and in developing performance-improvement plans based on collected performance information. Aligning Agencywide Goals, Objectives, and Measures: Agencies can encourage greater use of performance information by aligning agencywide goals and objectives, and by aligning program performance measures at each operating level with those goals and objectives. GPRA requires that agencies use performance measurement to reinforce the connection between their long-term strategic goals and the day-to-day activities of their managers and staff. To meet the GPRA requirements, an agency should cascade its goals and objectives throughout the organization and should align performance measures to the objectives from the executive level down to the operational levels. Furthermore, a greater focus on results can be created by cascading organizational goals and objectives down to the individual performance level. This alignment increases the usefulness of the performance information collected to decision makers at each level, and reinforces the connection between strategic goals and the day-to-day activities of managers and staff. Improving the Usefulness of Performance Information to Better Meet Management's Decision-Making Needs: To ensure that performance information will be both useful and used in decision making throughout the organization, agencies need to consider users' differing policy and management information needs. To be useful, performance information must meet users' needs for completeness, accuracy, consistency, timeliness, validity, and ease of use. Other attributes that affect the usefulness of information include, but are not limited to, relevance, credibility, and accessibility. Measures should be selected specifically on the basis of their ability to inform the decisions made at each organizational level, and should be appropriate to the responsibilities and control at each level. In that regard, involving managers in the development of performance goals and measures is critical to increasing the relevance and therefore the usefulness of performance information to their day-to-day activities. Developing Agency Capacity: The practice of building analytical capacity to use performance information--both in terms of staff trained to do analysis and availability of research and evaluation resources--is critical to using performance information in a meaningful fashion. Such capacity can be enhanced by training to develop the competencies and skills of managers to plan strategically, develop robust measures of performance, and analyze what the performance data mean. Performance management literature also states that training is a key factor in improving employees' capabilities and enabling employee involvement in achieving performance improvements. Communicating Performance Information Frequently and Effectively: Improving the communication of performance information among staff and stakeholders can facilitate the use of performance information by agency managers. Improvements can be achieved through frequent and routine communication, and the use of effective communication tools, such as visual aids. Frequent, regular communication is key for managers to inform staff and other stakeholders of their commitment to achieve the agency's goals and to keep these goals in mind as they pursue their day-to-day activities. Frequently reporting performance information also allows managers to review the information in time to take action to make improvements. Program managers can also communicate performance information upward through the management hierarchy, and across operating units. Vehicles for such communication include poster displays, performance scorecards, intranet sites, e-mail, and distribution of monthly performance-review meeting minutes. Agencies Reviewed: As noted above, to better understand what may hinder use of performance information and to identify opportunities for improvement, we selected FEMA, Interior, and CMS for more extensive review: Federal Emergency Management Agency: Although an independent agency originally, FEMA has been a part of the Department of Homeland Security (DHS) since 2003. FEMA's primary mission is to reduce the loss of life and property and protect the nation from all hazards, including natural disasters, acts of terrorism, and other man-made disasters, by leading and supporting the nation in a risk-based, comprehensive emergency-management system of preparedness, protection, response, recovery, and mitigation. There are eight directorates within FEMA, each dedicated to formulating policy and administering programs from the headquarters office located in Washington, D.C., and from 10 regional offices located across the United States. For fiscal year 2009, Congress appropriated approximately $15.6 billion for FEMA and FEMA programs, including $610 million in the American Recovery and Reinvestment Act (Recovery Act). [Footnote 8] For this study, we reviewed the Mitigation Directorate, which manages a range of programs designed to reduce future losses to homes, businesses, schools, public buildings, and critical facilities from floods, earthquakes, tornadoes, and other natural disasters; and the Disaster Assistance Directorate. Within the Disaster Assistance Directorate, we focused on the Public Assistance program, which administers FEMA's grants for emergency work, such as debris removal, and permanent work such as repairing and replacing damaged buildings following major disasters to state and local governments, authorized tribal organizations, and specific types of nonprofit organizations. Department of the Interior: Interior oversees nine separate agencies and bureaus with a wide range of responsibilities including resource use and protection, providing recreation opportunities on public lands, and honoring the nation's obligations to American Indians and Alaskan Natives. For this study, we examined the National Park Service (NPS), which is responsible for preserving the natural and cultural resources and protecting the wildlife of the national parks so that they will remain unimpaired for the enjoyment of this and future generations, and the Bureau of Reclamation (Reclamation), which is responsible for managing, developing, and protecting water and related resources in an environmentally and economically sound manner. In the Omnibus Appropriations Act, 2009, Congress appropriated more than $10 billion for Interior, with NPS and Reclamation receiving approximately $2.56 billion and $1.11 billion respectively.[Footnote 9] In addition, the Recovery Act provided approximately $3 billion for Interior, including $750 million for NPS and $1 billion for Reclamation.[Footnote 10] Centers for Medicare & Medicaid Services: CMS, a component of the Department of Health and Human Services, is the largest purchaser of health care in the United States, serving about 95 million Medicare, Medicaid, and State Children's Health Insurance Program (SCHIP) beneficiaries. CMS' fiscal year 2009 budget of $776.3 billion supports the entitlement programs of Medicare, Medicaid, and SCHIP and includes $35.9 billion in the Recovery Act. CMS has approximately 4,570 employees located in its headquarters and in 10 regions throughout the country. In February 2007, CMS reorganized its regional management from a geography-based reporting structure to a consortia structure based on the agency's key lines of business: Medicare health plans, Medicare financial management, Medicare fee for service operations, Medicaid and children's health, survey and certification of health care providers, and quality improvement. For this study, we reviewed the Consortium for Quality Improvement and Survey & Certification Operations (CQISCO), which is responsible for, among other activities, oversight of state surveys and activities intended to monitor the quality of nursing homes and other types of health care facilities that participate in Medicare and Medicaid. [Footnote 11] Our prior work had identified more than 40 recommendations to CMS to improve oversight of nursing homes, which included recommendations intended to improve CMS' use of performance data.[Footnote 12] Also, both Medicare and Medicaid programs are on GAO's high-risk list due to their size and complexity, as well as their susceptibility to mismanagement and improper payments.[Footnote 13] FEMA and Interior Were Hindered in Using Performance Information for Decision Making by Weak or Inconsistent Application of Key Management Practices: According to our 2007 survey of federal managers, FEMA and Interior were two of the lowest users of performance information among the 29 federal agencies surveyed. Several factors contributed to this relatively low use. At both FEMA and Interior, the demonstrated commitment of agency leaders to using performance information--a key management practice--was inconsistent. FEMA managers were also hampered by weakly aligned goals and limited analytic capacity to make use of performance information. At Interior and NPS, we observed, and officials and managers reported, a proliferation of measures, including some that, while meaningful for department-level accountability, were not relevant to their day-to-day management. Field managers we interviewed at NPS also said that poorly integrated performance and management information systems contributed to an environment where the costs of performance reporting--in terms of time and resources-- outweighed what they described as minimal benefits to their decision making. While both FEMA and Interior have taken some promising steps to make their performance information both useful and used, these initiatives have thus far been limited. FEMA Leadership Inconsistently Demonstrated Commitment to Using Performance Information in Certain Directorates, Regions: As we have previously reported, demonstrating the willingness and ability to make decisions and manage programs on the basis of results and inspiring others to embrace such a model are important indicators of leadership's commitment to using performance information.[Footnote 14] Our 2007 survey results indicated that, compared to the rest of government, a smaller percentage of FEMA managers agreed their top leadership demonstrated a strong commitment to using performance information to guide decision making (see figure 4). Figure 4: Percentage of Federal Managers Who Reported That Agency's Top Leadership Demonstrated a Strong Commitment to Using Performance Information to Guide Decision Making: [Refer to PDF for image: horizontal bar graph] Survey question: My agency‘s top leadership demonstrates a strong commitment to using performance information to guide decision making: Percentage responding to a ’great“ or ’very great“ extent: FEMA: 32%; Rest of Government: 50%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] At the same time, a significantly greater percentage of FEMA managers-- 27 percentage points more than managers across the rest of government-- agreed that lack of leadership commitment is a hindrance to their use of performance information (see figure 5). Figure 5: Percentage of Federal Managers Who Identified Lack of Leadership Commitment to Using Performance Information as a Hindrance to Measuring Performance or Using Performance Information: [Refer to PDF for image: horizontal bar graph] Survey question: Lack of ongoing top executive commitment or support for using performance information to make program/funding decisions. Percentage responding to a ’great“ or ’very great“ extent: FEMA: 52%; Rest of Government: 25%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] Our interviews with officials at FEMA were consistent with these survey results, indicating that management commitment was demonstrated inconsistently across the program directorates and regions we reviewed. Leaders and managers we spoke to throughout the management hierarchy were clearly committed to carrying out FEMA's mission. The level of commitment to using performance information for decision making, however, appeared to vary among those we interviewed. Further, several FEMA headquarters officials said that top leadership commitment to using performance information, demonstrated by one of the agency's former leaders, had not permeated the organizational culture and that it was unclear whether recently instituted practices would be sustained under the new administration. A former Deputy Administrator, who served as FEMA's chief operating officer before he left his position as part of the change in administration, said that he believed strongly in using performance information to identify areas for improvement. He said that when he arrived at FEMA in 2006, the agency culture was response driven and there was little recognition that performance information could be used to make improvements. As a first step, he said he had focused on improving FEMA's business practices in areas such as hiring, financial management, and information systems. He also introduced performance information briefings on disaster assistance-related areas of concern- -such as housing for Hurricane Ike victims--in an effort to improve performance. For example, he said that by reviewing data on various aspects of the post-Hurricane Ike replacement housing situation--such as individual housing needs, the inventory of mobile homes, and the rate of mobile home installation--he was able to identify and fix a bottleneck in providing an adequate supply of temporary housing. Another FEMA official with responsibilities for performance reporting said that such disaster assistance metrics were important first steps in using performance information to improve disaster response. However, he noted that these metrics were specific to that disaster and that more work is needed to develop performance measures that can be applied to all disaster response situations. More recently, the former Deputy Administrator said he had begun to turn his attention to FEMA's performance at the regional level. For example, he said he had begun to work with the regions to require strategic plans and quarterly performance reporting to his office. He acknowledged, however, that some regions are better than others at using performance information and that these quarterly reporting efforts were still immature. Further, our interviews with other officials indicated that only one reporting cycle had been completed before the former Deputy Administrator had left. Our interviews with top officials and regional program managers in the Disaster Assistance and Mitigation Directorates indicated that leadership commitment to using performance information varied among directorates and regions. In the Disaster Assistance Directorate, one headquarters official told us that he does not need performance targets to help him determine whether his directorate is accomplishing its mission. He says he relies primarily on verbal communications with the leadership and from FEMA's regions, joint field offices, and members of Congress to identify issues to be addressed and areas that are running well. Although he said he does use data to monitor how well the directorate is responding to post-disaster inquiries for assistance, such as the call center statistics from FEMA's National Processing Service Center, his description indicated that these data mostly reflect workload and activity levels, rather than performance against goals. Another headquarters official within the Disaster Assistance Directorate's Public Assistance program said he does not receive formal performance reports from regional program managers, nor are any performance reports required of him by his supervisors. He noted that goals for the Public Assistance program are established for the field at the regional level and that the directorate's role is primarily one of developing policies and guidance. Although regional staff completed reports on various activities related to his programs, he said he was uncertain how he would use these regional reports since performance monitoring was not his focus. Instead, he said that he spoke to the regions on an ad hoc basis as performance problems arose. Officials responsible for Disaster Assistance Directorate programs in two of three regions we reviewed similarly described the ad hoc nature of performance reporting to headquarters. One said that, although he has begun to issue quarterly reports on various data, such as funding obligation rates, to his Regional Administrator that are shared with headquarters, he had not received any comments or feedback on what is reported. Another regional Disaster Assistance Directorate official told us that although regional program supervisors and staff meet quarterly to discuss timeliness related to funding obligations and project worksheet completion,[Footnote 15] among other items, they do not communicate performance information to the directorate head or headquarters on a regular basis. Officials from the third region we interviewed said that, although they anticipated that new strategic planning initiatives might change how they report on performance in the future, they did not review performance information with their regional staff, nor did they communicate it to headquarters. Officials we interviewed within the Disaster Assistance Directorate expressed reluctance toward holding their staff accountable for meeting performance goals due to external factors, such as the unpredictability of disasters beyond their control. Further, some expressed uncertainty as to how they could use performance information in the face of uncontrollable external factors. In contrast, according to several officials from the Mitigation Directorate, they had begun to use performance information to more effectively manage under unpredictable circumstances. These officials said that the former Mitigation Administrator's commitment to performance and accountability helped change the directorate culture to one that encouraged use of performance information to plan for and respond to factors outside of their control. For example, storms and other natural events can disrupt the Mitigation Directorate's production work related to flood-plain maps modernization. To plan for possible disruptions, Mitigation Directorate officials said they review performance information on progress toward map modernization goals on a monthly basis, in connection with weather forecasts. This review helps them to determine in advance if they are at risk of missing performance targets and to identify corrective actions or contingency plans in order to get back on track toward achieving their goals. They also described their own commitment to using performance information as a means to demonstrate the value of their programs, make improvements, and achieve results (see example 1). [Text box: Example 1: Mitigation Directorate Leadership Used Performance Information to Engage Partners in Improving Map Modernization Outcomes: The Mitigation Directorate works with multiple stakeholders, including state and local governments, the insurance community, and private contractors to ensure that flood-prone communities have the most current and reliable flood data available, and that those communities are in compliance with regulations referencing current Flood Insurance Rate Maps (FIRM), which are used to regulate land development through flood-plain management and for flood insurance purposes. Communities that fail to adopt the new maps by the FIRM effective date are suspended from the National Flood Insurance Program, which can negatively affect local real-estate transactions and limit the community‘s eligibility for disaster assistance. In an effort to improve community compliance, the Mitigation Directorate set an annual performance target of 93 percent of communities adopting new maps by the FIRM effective date. They closely monitored performance by incorporating the map adoption rate into state grant agreements, map modernization contracts, and FEMA‘s regional performance scorecards. According to Mitigation Directorate officials, they frequently reviewed map modernization performance information with their external stakeholders and FEMA‘s regional management, which sent a clear signal that they were paying attention to outcomes. According to these headquarters officials, they were able to meet or exceed their performance target of 93 percent, in part as a result of their frequent communication and review of performance information. End of text box] In addition, Mitigation Directorate officials from both the national office and from two of the three regions we reviewed noted frequent communication and review of program performance information. For example, each of the Mitigation Directorate's three divisions developed score cards including performance measures that are reviewed quarterly. Mitigation Directorate officials also told us they involved staff throughout the directorate and their external stakeholders including insurance companies, the lending community, and state and local officials in their efforts to establish preliminary performance goals and measures. One regional manager we spoke to said that directorate and regional leadership fostered collaboration by encouraging regional staff to develop and share ideas for metrics and targets through weekly conference calls devoted to discussing performance information. Mitigation Directorate officials said that developing measures and holding staff and contractors accountable for their performance was not an easy transformation. They said that one key to this culture change was for the leadership to strike an appropriate balance between holding managers accountable for agency goals and building trust among managers and staff that performance information would be used as an improvement tool, rather than as a punitive mechanism. Finally, Mitigation Directorate officials said that once managers and staff began to see that measuring performance can actually help them to improve results, they became more supportive of their leadership's efforts to use performance information in their decision making. FEMA Has Not Consistently Aligned Agency, Program, and Individual Performance Goals: As we reported previously,[Footnote 16] agencies can promote use of performance information by aligning agencywide goals and objectives, and by aligning program performance measures at each operating level with those goals and objectives. FEMA's current strategic plan includes high-level strategic goals, such as, "deliver easily accessible and coordinated assistance for all programs." FEMA officials said they had recently completed an addendum to the strategic plan that can be updated to reflect evolving circumstances. These officials acknowledged, however, that these goals and measures are high level and that establishing performance goals at the regional or division level would help FEMA to cascade organizational goals down to individual staff. According to our interviews with other top officials, FEMA had started to develop regional performance measures and align them with different operating levels of the agency. In mid-2008, FEMA's leadership directed all regions to begin developing regional strategic plans; however, several officials described challenges in ensuring that these plans included meaningful performance measures that aligned with those at the directorate and agency level. The regions' early efforts produced performance measures that, from the program directorates' perspective, were often not aligned with their program goals. Subsequently, FEMA performance and operations officials worked with the regions and national program directorate leaders to refine the measures and to provide training, which included information on aligning activities and outputs with strategic goals. As of late 2008, FEMA had developed an initial set of 38 regional performance measures that are intended to link to the agency's strategic goals. It was unclear, however, when the regions will begin collecting data on these metrics to establish baseline performance levels and targets for improvement. Top officials acknowledged that efforts to align regional goals with the agency's strategic objectives is a work in progress and will take some time to complete. We have reported that a greater focus on results can be created by cascading organizational goals and objectives down to the individual performance level, helping individuals to see the connection between their daily activities and organizational goals and providing a basis for holding individuals accountable for their results.[Footnote 17] However, our 2007 survey results indicated 44 percent of FEMA managers reported being held accountable for agency strategic goals, compared to 60 percent of their counterparts in the rest of government (see figure 6). Figure 6: Percentage of Federal Managers Who Reported They Were Held Accountable for Their Agency's Accomplishments of Strategic Goals: [Refer to PDF for image: horizontal bar graph] Survey question: Agency managers/supervisors at my level are held accountable for agency accomplishment of its strategic goals. Percentage responding to a ’great“ or ’very great“ extent: FEMA: 44%; Rest of Government: 60%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] FEMA lacked a performance-management system that can cascade organizational goals to the individual performance level--that is, create a "line of sight" linking individual goals and organizational success. In September 2008, FEMA human capital management officials told us that aligning agency goals with non-Senior Executive Service (SES) performance objectives was being accomplished through the implementation of DHS' ePerformance Management System.[Footnote 18] At that time, according to officials, approximately 30 percent of permanent full-time employees had been converted to the new system. However when Congress disallowed DHS from further implementing the human-resources-management system then in place, in October 2008, [Footnote 19] DHS rescinded its ePerformance Management and related human-resources systems. For fiscal year 2009, FEMA chose to revert to its previous appraisal system, which it had established in the mid- 1990s. Although human-capital officials told us that this appraisal system encourages supervisors and employees to develop work plans, our review of policies and appraisal documents indicated no requirement to align individual performance objectives with agency or program performance goals. A human-capital official at FEMA noted that, in an effort to forestall multiple, time-consuming changes to FEMA's performance-management system, the agency did not want to invest further resources in this area until the agency is able to ascertain how DHS is proceeding with department-level performance-management policies and systems. This official also explained that changes to FEMA's current performance management policies or guidance--which covers both bargaining-unit and non-bargaining-unit employees--would be subject to negotiation with its employee labor unions. Interviews with management officials from three regions further indicated that the practice of cascading organizational goals to the individual performance level was applied inconsistently across regions. The top official from one region said that the region does not include GPRA goals or other agency-level goals in individual performance agreements and noted that FEMA is not very mature in individual performance management. Although the top official from another region said that in the absence of direction from headquarters, he had worked to link individual performance objectives to FEMA's strategic goals, the performance agreement example provided to us included expected outcomes that were not easily measurable, such as, "— build trust and confidence with the state and local partners." An official from a third region said that while the now-defunct ePerformance Management System was useful in helping the region to establish such linkages, FEMA's current performance management system is a "sham" because it provides no tool to measure performance against goals. Limited goal alignment in the area of individual performance management may hinder managers' ability to understand how their roles and responsibilities affect broader results. For example, one official told us that without such alignment, it was difficult to show staff how their efforts supported FEMA's mission. It was also difficult to hold managers accountable for results. FEMA Officials Said Inadequate Information Systems, Analytic Skills Hindered Use of Performance Information: The practice of building analytic capacity is critical to using performance information in a meaningful fashion. Our review of FEMA's analytic capacity to use performance information--in terms of both management information systems and trained employees--revealed some weaknesses. According to our 2007 survey, the percentage of FEMA managers reporting that their agency is investing in resources to improve the agency's capacity to use performance information is lower than the rest of government (see figure 7). Figure 7: Percentage of Federal Managers Who Reported Agency Investment in Performance Data Capacity: [Refer to PDF for image: horizontal bar graph] Survey question: My agency is investing in resources to improve the agency‘s capacity to use performance information. Percentage responding to a ’great“ or ’very great“ extent: FEMA: 19%; Rest of Government: 33%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] These survey results were consistent with the perspective we heard from key headquarters officials who told us that poorly integrated systems made it difficult for FEMA managers to use performance information. According to one official, in order to gather performance information, it was necessary to write programs to generate specific reports for each of the systems and then manually integrate the information, making it difficult to produce repeatable and verifiable reports. For example, in order to pull performance information together at the program- directorate level, this official told us he had to ask the few technical staff capable of working with the systems to devote a significant amount of time and effort to producing the information. FEMA officials told us they were pursuing a new budget system, RM Online, which includes a component intended to make high-level program and program performance information readily available to senior managers. However, the agency was still evaluating the system and it was unclear when it might be implemented. In addition to weaknesses in information-systems capacity, according to several officials we interviewed, there were a limited number of staff with the analytic capacity necessary to work with performance metrics. The former Deputy Administrator said that when he joined FEMA in 2006, managers and staff did not use or understand performance data. A high- ranking directorate official told us that he was trying to increase use of performance information to improve workflow and other operations in his area. However, he said he lacked staff with the skills to analyze information for decision-making purposes. Another official with responsibilities for performance measurement said that the lack of analytically-skilled staff throughout the agency posed a challenge to using performance information. According to this same official, in order to improve the agency's capacity to use performance information, FEMA has begun to provide training on performance measurement to directorate and regional managers. Our review of the training materials indicated that they addressed specific areas that we have identified as critical to using performance information, including strategic planning, developing robust performance measures, and analyzing what the performance data mean. FEMA has also developed strategic planning guidance that outlines an approach for developing performance measures and evaluating performance, among other goals. However, so far, the training has been provided only to representatives from each region and directorate managers in headquarters, and a key official acknowledged that not all directorates have been equally effective at pushing the training out to their managers and staff in the regions. Interior, NPS Officials and Managers Reported Uneven Leadership Commitment to Using Performance Information for Decision Making: We have previously reported that to drive continuous improvement throughout an agency and inspire employees to accomplish challenging goals, it is critical that leadership demonstrates its commitment to results-oriented management.[Footnote 20] We have also reported that top leadership can demonstrate such commitment by clearly communicating how they use performance information for decision making. On survey items related to managers' perceptions of their leadership's commitment to using performance information, Interior's 2007 results were lower than those in the rest of government (see figure 8). Figure 8: Percentage of Federal Managers Who Reported Top Leadership Commitment to Using Performance Information to Guide Decision Making: [Refer to PDF for image: horizontal bar graph] Survey question: My agency‘s top leadership demonstrates a strong commitment to using performance information to guide decision making. Percentage responding to a ’great“ or ’very great“ extent: Interior: 37%; Rest of Government: 50%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] Our interviews with top leadership and managers provided further insight into these survey results. At all levels, we observed that leaders and managers conveyed a strong commitment to accomplishing the agency's mission. However, their commitment to using performance information for decision making was less evident. For example, the former Deputy Secretary of the Interior said that, although she reviewed performance information at the end of the year in connection with preparing the department's annual performance report, she was not involved in regularly monitoring the performance information reported under GPRA and PART.[Footnote 21] Another Interior official characterized the department's strategic plan as more of a vehicle for communicating high-level goals and accomplishments than as a tool for management decision making. This view that top leadership did not use performance information to make decisions was supported by several NPS managers who referred to the performance reporting process as "feeding the beast," because they receive little or no communication from either Interior or NPS headquarters in response to the information they are required to report, leading them to assume that no one with authority reviews or acts on this information. As we have previously reported, leaders can demonstrate their commitment to using performance information for various management functions in a number of ways, such as: 1. holding individuals accountable for results by evaluating their performance against goals; 2. identifying problems in existing programs, to try to identify the causes of problems and to develop corrective actions; 3. developing strategies, planning and budgeting, identifying priorities, and making resource allocation decisions to affect programs in the future; and: 4. identifying more effective approaches to program implementation and sharing those approaches more widely across the agency.[Footnote 22] At Interior our 2007 survey results on questions related to accountability indicated that Interior's managers were similar to those in the rest of government. For example, 75 percent of Interior managers--similar to those at other agencies--reported being held accountable for results (see figure 9). Figure 9: Interior Managers Reported Being Held Accountable to a Similar Extent as Rest of Government: [Refer to PDF for image: horizontal bar graph] Survey question: Agency managers/supervisors at my level are held accountable for agency accomplishment of its strategic goals[A]. Percentage responding to a ’great“ or ’very great“ extent: Interior: 56%; Rest of Government: 60%. Survey question: Agency managers/supervisors at my level are held accountable for the result(s) of the program(s)/operation(s)/project(s) they are responsible for[A]. Percentage responding to a ’great“ or ’very great“ extent: Interior: 75%; Rest of Government: 72%. Source: GAO. Notes: Data are from GAO 2007 survey. [A] The differences in responses between Interior and the rest of government on these two items are not statistically significant. [End of figure] Several NPS managers we interviewed corroborated these survey results, citing how performance information was used to hold them individually accountable for achieving certain performance goals. However, this focus on individual accountability did not appear to extend to using performance information for other management functions that leading organizations employ, such as identifying problems, taking corrective actions, or developing strategy (see figure 10). Figure 10: Interior Managers Reported Using Performance Information to Identify Problems, Take Corrective Actions, or Develop Strategy to a Lesser Extent than Rest of Government: [Refer to PDF for image: horizontal bar graph] Survey question: Developing program strategy. Percentage responding to a ’great“ or ’very great“ extent: Interior: 38%; Rest of Government: 51%. Survey question: Identifying program problems to be addressed. Percentage responding to a ’great“ or ’very great“ extent: Interior: 40%; Rest of Government: 61%. Survey question: Taking corrective action to solve program problems. Percentage responding to a ’great“ or ’very great“ extent: Interior: 41%; Rest of Government: 60%. Source: GAO. Notes: Data are from GAO 2007 survey. [End of figure] Our interviews with top leaders and managers at NPS may help to explain these survey results. A senior headquarters official at NPS responsible for park operations said that he is not involved in regularly monitoring the performance of the park system in achieving GPRA and PART-related performance goals and does not use this information to manage park operations. Rather, he regularly communicated with his staff about other management issues, such as the rate at which funds are obligated.[Footnote 23] However, obligation rates, while helpful in assessing the pace at which projects are progressing, do not provide information about the results achieved with these funds and therefore may not be useful in driving performance improvements. He also said that although GPRA and PART-related goals were not useful to him in making operational and program management decisions, NPS had included these goals in managers' performance agreements in order to comply with Interior's individual performance management policies.[Footnote 24] However, this official--and other NPS senior officials and managers we interviewed--was concerned that GPRA and PART information was sometimes used to evaluate individual performance without appropriate context or recognition of what is outside managers' control. Some noted, for example, that storms and flooding can have an effect on maintenance and repair performance targets, and air and water quality goals are dependent on environmental factors outside of the parks. We have previously reported[Footnote 25] that successful organizations typically create ambitious performance goals aimed at achieving significant improvements in performance, rather than marginal improvements of just a few percentage points. However, several NPS officials and managers told us that GPRA and PART targets were not always set at ambitious levels. An NPS headquarters official explained that targets set at the park and regional levels are aggregated at the agency level and subject to evaluation in the context of past performance and anticipated funding. If necessary, headquarters will adjust performance targets to appropriate levels. He also noted that parks and regions are advised at the beginning of every planning cycle that targets should be based on what can be realistically accomplished during the course of the fiscal year. However, some park managers we spoke to said that they were directed by their superiors to set targets at levels they can safely meet because they and their superiors could otherwise be penalized. Without sufficiently ambitious goals, managers may not have incentives to use performance information to identify opportunities for significant improvement. Although our interviews indicated some concerns that NPS leadership had focused more on using performance information for accountability than to help make improvements, some regional and park managers did note examples of specific program areas, such as cultural resources and facilities management, where they had begun to see their NPS leadership take a broader approach to using performance information. For example, one regional director said that the cultural-resources program leadership based in headquarters had effectively communicated to the regions how they used information from the PART review to inform their funding decisions. He went on to say that even though some of the decisions they made about prioritizing certain projects over others were painful, staff in the field understood how the performance information they submitted contributed to the program's funding decisions. He noted that his park superintendents appreciated when program leaders communicated that the information they provided was useful and demonstrated how it informed their decision-making processes. Measures That Lacked Credibility to Bureau Level Managers, in Combination with a Proliferation of Measures, Detracted from Usefulness of Performance Information: We reported previously that performance information must be useful and meet differing policy and management information needs in order to encourage its use for decision making.[Footnote 26] Our 2007 survey results show that, compared to the rest of government, a significantly higher percentage of Interior managers reported that difficulty determining meaningful measures is a barrier to using performance information (see figure 11). Figure 11: Percentage of Federal Managers Who Reported That Difficulty in Determining Meaningful Measures Hinders Using Performance Information: [Refer to PDF for image: horizontal bar graph] Survey question: Difficulty determining meaningful measures hinders measuring performance or using the performance information. Percentage responding to a ’great“ or ’very great“ extent: Interior: 58%; Rest of Government: 39%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] Our interviews at NPS and Reclamation may help to explain these survey results, with managers describing several types of problems that detract from the usefulness of the performance information they are required to report. Some GPRA, PART Measures Lacked Credibility for Decision Making at the Bureau Level: We reported previously that, to be useful, performance information should be consistent, timely, valid, relevant, and credible, among other attributes.[Footnote 27] At both NPS and Reclamation, managers we interviewed said they do not consider certain GPRA and PART measures to be meaningful indicators of performance. For example, Reclamation managers report on a performance measure of the amount of water they deliver, but according to a number of managers within the bureau, this measure is not meaningful because it reflects unpredictable changes in weather conditions that could affect their customers' needs, rather than how well they are performing (see example 2). [Text box: Example 2: Reclamation Managers Described How a Performance Measure That Lacked Relevance Was Not Useful for Decision Making: Reclamation delivers approximately 10 trillion gallons of water to more than 31 million customers including farmers, municipalities, and irrigation districts each year. According to Reclamation officials, it is critical that they frequently monitor performance information related to water delivery in order to mitigate the effect of water shortages on their customers. Reclamation is responsible for identifying potential water shortages as early as possible and restricting delivery when supply is running low. To identify potential shortages, Reclamation managers need information that can tell them how well they are managing their supply of water. Instead, Reclamation managers report against a target for the amount of water delivered, which is not adjusted for changes in critical factors such as weather conditions or customer needs. For example, if there is a lot of rainfall, Reclamation customers do not need the bureau to deliver water. By not delivering water, Reclamation is meeting customer needs and being a responsible steward of a natural resource, but it is also failing to meet its performance goal. Under these circumstances, Reclamation officials pointed out, if a manager were to make a decision based on a performance goal related to the amount of water delivered, he would have a perverse incentive to deliver water that is not needed. End of text box] Some managers suggested it would make more sense to have a measure that takes into account how well they manage water supply in relation to customer needs, rather than amount delivered. Headquarters officials at Interior and Reclamation said they view the measure as an important statistic related to one of Reclamation's primary missions, to deliver water, and that they do not expect managers to meet the water-delivery targets if circumstances change. One official also noted that performance information on this measure is accompanied by narrative that explains whatever factors may have affected the amount of water delivered. However, Interior included annual goals for this measure for fiscal years 2009 and 2010, and reported on the results of this goal for the past four years in its fiscal year 2008 Performance Accountability Report. Several Reclamation managers we interviewed said that this measure lacked credibility and was not used to make decisions. Despite the concerns expressed by regional and frontline managers, an Interior official noted that Reclamation had not proposed an alternative measure, which may be due to the difficulty of defining a quantifiable, verifiable measure. At both NPS and Reclamation, managers we interviewed described performance information that lacked credibility because the measures do not accurately define comparable elements or do not take into account different standards across bureaus or units. For example, several NPS managers noted that one of the measures they report, "percent of historic structures in good condition," does not differentiate between a large, culturally significant structure such as the Washington Monument and a smaller, less significant structure such as a group of headstones. Consequently, a manager could achieve a higher percentage by concentrating on improving the conditions of numerous less significant properties. Similarly, measures related to trail maintenance do not take into account the baseline conditions or the purpose of the trails, which can vary greatly and affect the level of resources needed for maintenance. Some park managers expressed concern that certain performance measures falsely imply a consistency among units being compared, which could lead to inaccurate conclusions if decision makers do not take the proper context into consideration when reviewing information. Some managers also said that the long-term nature of some of the GPRA goals--many of which they acknowledged as critical--made it difficult to use the related performance information for daily decision making. For example, one headquarters official explained that the goal of restoring damaged park lands to a desired condition is an important aspect of NPS' mission, but progress may be slow and the result may not be achieved for many years. As we have reported,[Footnote 28] involving managers in the development of performance goals and measures is critical to increasing the credibility and therefore the usefulness of performance information to their day-to-day activities. A couple of NPS managers acknowledged that in the early stages of GPRA implementation, they were more involved, and felt more committed to the process of developing goals and measures than they do currently, but that is no longer the case. Several Reclamation managers also said that they believed they had little influence in defining or revising GPRA and PART measures and that as a result, they do not take ownership of the performance information. According to a senior Interior headquarters official, however, the department has continued to provide a forum for bureaus to propose modifications or alternatives to existing measures. For example, every 3 years, Interior requests proposals for changes or revisions, which are then reviewed by bureau and department officials, OMB, and the general public. However, several managers we interviewed at NPS and Reclamation said that when they have proposed changes to measures that they find misleading or irrelevant, they did not always see evidence that their suggestions were considered or acted upon. An Interior headquarters official explained that suggestions received from the bureau level may not always be appropriate for adoption at the department level, where it is important to have measures that can provide Interior decision makers with a view over its bureaus' aggregate performance against mission goals. In these cases, the department forwards the suggestions to the bureau policy makers responsible for performance measurement and reporting for their consideration. This official pointed out that Interior's aggregated performance measures do not preclude the bureaus from reviewing disaggregated performance information; or other performance information that is more meaningful or relevant to their particular management decision-making needs. Proliferation of Performance Measures at Interior and NPS Detracted from Usefulness: We have previously reported that to be useful and meaningful to managers and staff across an agency, performance measures should be limited at each organizational level to the vital few that provide critical insight into the agency's core mission and operations. [Footnote 29] Setting such a limit also helps to ensure, among other things, that the costs involved in data collection and analyzing the data do not become prohibitive. However, in the 7 years since the inception of the former administration's PART assessment initiative, Interior has expanded its performance reporting to include 440 PART program measures, in addition to its approximately 200 strategic performance measures that satisfy GPRA reporting requirements on its nine bureaus. A senior headquarters official at Interior said that the annual Performance and Accountability Report[Footnote 30] contains so much data that it is difficult for senior leaders and managers to focus on priorities and easily identify performance gaps among the different program areas. At NPS, managers are required to report on over 120 performance measures related to GPRA and PART, covering a broad range of programs and projects including concessions, facilities, volunteer hours, visitor satisfaction, recreation opportunities, safety elements, and the condition of natural, historical, and cultural resources. At NPS, a senior headquarters official echoed the concern stated by an Interior official, noting that the volume and scope of performance information that managers are required to collect and report make it difficult for them to determine which aspects of performance are important to department and service leadership. Moreover, regional and park managers stated that the resources needed to collect and report on the large volume of GPRA and PART measures are extensive. Managers staffed to smaller parks with fewer employees were particularly concerned about the workload associated with collecting and reporting on all of these measures, with one noting that it was at the direct expense of park operations and maintenance activities. Several managers told us that new measures and reporting requirements were frequently introduced with new programs and initiatives, but that they were unaware of any efforts at NPS to review or retire performance measures that may no longer represent management priorities. According to officials at Interior and NPS, they were aware of issues related to the usefulness of performance information as currently collected and reported. The former Deputy Secretary of the Interior said that department-level measures were not necessarily useful for making day-to-day management decisions. She attributed this to Interior's effort to create measures with reach, which she defined as a few simple measures that can provide a view of how all of their bureaus are performing against broad outcome goals. She contrasted the department-level measures with rich measures that are intended to provide detailed information that makes sense on the ground and can be used to manage programs. Although this official said that, by definition, there should only be a few high-level reach measures, she acknowledged that there had been a proliferation of measures, which made it difficult to manage. Interior and NPS Have Undertaken Efforts to Improve the Usefulness of Performance Information: To address these concerns, some efforts had been initiated at both organizations to improve the usefulness of performance information without adding to the existing data-collection and reporting process. A senior headquarters official at Interior described how he worked with department and bureau managers to select a subset of 26 key performance indicators from the department's more than 200 strategic performance measures to help internal and external audiences focus on the department's critical mission areas. These 26 measures are aggregated across bureaus with shared performance goals and link their performance with cost information. For example, Interior reported that in 2007, the department spent approximately $114.4 million to control about 635,000 acres--or 1.7 percent--of total acres with invasive plant species. These data capture expenditures and progress made by the four bureaus- -NPS, Reclamation, Bureau of Land Management, and Fish and Wildlife Service--that collectively contribute to the goal. According to this official, these summary indicators are intended to help senior officials navigate the large volume of performance information so they can more easily identify areas that are achieving results and focus on areas that need improvement. By linking budget information to these key indicators, this official said he was able to improve the usefulness of the performance information as a decision-making tool for Interior, providing top leaders with a snapshot of the cost of meeting various goals. At NPS we observed several initiatives that were intended to improve the usefulness of performance information. The NPS Scorecard, for example, was developed by the NPS Comptroller--with input from headquarters, the regions, and the parks--as a diagnostic tool to evaluate performance and efficiency across the organization. The current version of the scorecard includes 34 financial, organizational, and strategic performance indicators derived from existing data sources and scores each park relative to its counterparts. Although still under development, some managers described the NPS scorecard as a promising instrument because it provided them with an analysis of their performance and efficiency in comparison to other parks, among other reasons. They also appreciated that the scorecard pulled information from existing NPS databases and did not require additional investment in data collection or reporting. However, some managers expressed concerns that the scorecard compared park performance to an average of all parks rather than an established benchmark. On an annual basis, to meet NPS-wide goals, specific performance targets are established for individual parks. Therefore, it is possible that a park that met or exceeded these agreed upon performance targets, but was still below the NPS-wide average, would receive an unfavorable rating on the scorecard. Another approach that was being adopted in some of the NPS regions we reviewed--the Core Operations Analysis--is a park-level funding and staffing planning process that is intended to improve park efficiency and ensure that a park's resource-allocation decisions are linked to its core mission goals. Regional-level managers who engaged in the Core Operations Analysis said it was useful in establishing goals based on the park's priorities, monitoring progress toward achieving those goals, and holding park superintendents accountable for meeting established goals, in contrast to the GPRA and PART goals, which several managers perceived as relating to department-or program-level planning efforts and long-term goals. One region had taken the additional step of establishing linkages between their analysis and GPRA goals with assistance from a senior Interior official. In addition, two park superintendents we interviewed said they were able to make GPRA and PART measures useful for their day-to-day management decision making by linking them to shorter-term, operational goals. For example, one had created an annual work plan that linked GPRA goals to intermediate tasks necessary to achieve those goals so that staff could see how their efforts contributed to aggregate or long- term outcomes. Specifically, the GPRA target that "87 percent of visitors understand and appreciate the site's significance" was linked to two intermediate goals: "research and preserve information on stories and resources" and "conduct 7 guided tours per day, 362 days per year." Another said that he incorporated targets related to these GPRA goals in all of his managers' performance plans in order to make the performance information more meaningful on an annual basis. Although these initiatives appeared to be positive steps toward ensuring that performance measures were useful for decision making at different operating levels--as well as to help focus leaders and managers on performance priorities--they did not appear to reduce the volume of performance information that NPS and Reclamation were required to report. As a result, efforts to ensure that performance information was useful at the bureau level may have been hindered. For example, the two park managers who had linked shorter-term or disaggregated targets to GPRA goals told us that they were atypical in taking this approach. They explained that many of their colleagues were not able to develop bridge measures or interim goals due to lack of training, time, or staff resources relative to their reporting workload; or because they were responsible for reporting on too many measures as a result of the scope of their park's operations. This perspective was borne out by our interviews with other park superintendents who said they expended significant resources to report on many measures that lacked relevance to their management decision making. One park manager told us that he did not have the staff available to collect and use information over and above what was already required. Labor-Intensive and Poorly Integrated Data Systems Increase Burden of Performance Reporting at NPS and Reclamation: According to NPS managers we interviewed at all levels, lack of integration among NPS' multiple data systems--and little flexibility to modify its primary performance information system to accommodate evolving programs and requirements--contributed to a time-consuming, labor-intensive performance-reporting process. At NPS, managers were required to use the Performance Management Data System (PMDS) to collect GPRA performance information. However, regional and park managers we spoke to said that the system was not easy to use for data entry or for obtaining information needed to manage programs or operations. One headquarters official confirmed that the system had not been designed as a management tool, but was designed simply to aggregate performance from the parks. Several managers described the system as slow and difficult to use and noted this was especially problematic for NPS' remote parks that do not have high-speed system access.[Footnote 31] NPS managers we interviewed also said that PMDS cannot be used to integrate GPRA data with other information that would be useful for decision making such as park projects status and funding information, facilities condition information, or other park program performance information that is not covered by current GPRA measures. Several park managers described how another system, the Project Management Information System (PMIS), captures the park's project proposals and is useful in tracking park-level projects' status and funding. Although PMIS has a field where park managers can enter expected performance outcomes related to GPRA goals, because there are no automated linkages between the two systems, doing so would require additional data entry. Further, a NPS headquarters official said that PMDS does not have the flexibility to generate new reports to satisfy evolving reporting requirements, which can be problematic when NPS is expected to demonstrate results associated with special funding initiatives such as the Centennial Commitment--a multiyear initiative that provided NPS with $100 million in fiscal year 2008 to augment existing park funding. According to this official, PMDS was not designed to differentiate Centennial dollars--or any other special appropriations--from their annual appropriation; rather the system was set up to collect performance information, regardless of funding source, from individual park units. According to this same official, PMDS does not attribute specific performance results to discrete fund sources because gains in performance are often the result of multiple fund sources over multiple years. To satisfy Centennial reporting requirements, managers had to collect and report additional performance information, using other systems, to show what they did with Centennial dollars and what types of results they accomplished. One park manager noted that although the lack of integration and flexibility was an inconvenience for the largest parks, they generally had enough staff to absorb the workload. However, some managers responsible for smaller, geographically isolated parks that do not employ many staff said that these issues presented a challenge to their ability to enter data and generate required reports. A top official at Interior also acknowledged that this was especially difficult for smaller parks where a few staff members have to serve in many roles and may not have enough time to collect all of the required data. Some Reclamation managers we spoke to also said that a lack of integrated systems made it difficult for them to collect and report on performance information. According to a regional manager responsible for GPRA reporting, there is no one centralized database to which a Reclamation executive can go to find out how the bureau is doing on all of Reclamation's required performance goals. The lack of linkage among the different Reclamation systems required managers to enter the same data multiple times, which some managers said is a burden. Officials at CMS Headquarters and Selected Programs Said Key Management Practices Had Promoted Use of Performance Information for Decision Making: In 2000, significantly fewer managers at CMS--then known as the Health Care Financing Administration--reported using performance information for various management decisions, as compared to their counterparts in the rest of government. Between our 2000 and 2007 surveys, however, CMS showed one of the largest average increases in the percentage of managers who reported using performance information for certain decisions. This increase placed CMS in about the middle of our agency rankings, which were based on an index of 2007 survey results designed to reflect each agency's managers' reported use of performance information (see figure 20 in appendix II). Selected officials we interviewed attributed this change to the combined effect of key management practices they employed, including leadership commitment to using performance information; alignment of strategic and performance goals; improving the usefulness of performance information; and building the analytic capacity to collect and use performance information. They also cited the legislative mandate for expanded performance reporting included in the Medicare Prescription Drug, Improvement, and Modernization Act of 2003 as another key change factor.[Footnote 32] According to these managers and officials, their increased use of performance information helped them to identify problems and solutions to improve results. CMS Managers Said Highly Visible Leadership Commitment, Frequent Communication Fostered Use of Performance Information: Our 2007 survey results indicated that CMS managers are at about the same level as the rest of government in reporting leadership commitment to results-oriented management, but compared to 2000, significantly more CMS managers agreed that their leadership is committed to using performance information (see figure 12). Figure 12: Percentage of CMS Managers Who Reported Top Leadership Demonstrated Commitment to Achieving Results: [Refer to PDF for image: horizontal bar graph] Survey question: My agency's top leadership demonstrates a strong commitment to achieving results. Percentage responding to a ’great“ or ’very great“ extent: 2000: 46%; 2007: 69%. Source: GAO. Note: Data are from GAO 2000 and 2007 surveys. [End of figure] Nearly all of the CMS officials we interviewed credited the commitment of one or more agency leaders--such as the CMS Administrator, Chief Operating Officer, or a Consortium head--for their increased use of performance information to achieve results. Some of them further described how leadership demonstrated their commitment. For example, a budget official told us that at the first staff meeting of the year he distributes the Chief Financial Officer's performance plan and priorities to lay out for staff their performance goals for the year. He also described how each group director is given a poster-sized chart outlining his/her performance goals and program-specific GPRA goals for the year, which hang in their offices. As another example, an official we interviewed in Region IX described how top management discusses performance goals and the accomplishment of goals in staff meetings. She also noted that they feature information on performance milestones in their organizational newsletter. CMS Survey & Certification Division managers of Region IV of the Consortium for Quality Improvement and Survey & Certification Operations (CQISCO) attributed improvements in performance against GPRA goals--such as reducing the incidence of pressure ulcers among nursing- home patients--to leadership commitment to using performance information and frequent communication of performance information with stakeholders, among other factors. According to CMS Region IV managers we interviewed, they are several steps removed from nursing-home health- care delivery, which in the past had been seen as a limiting factor in their ability to affect outcomes among nursing-home patients. One manager cited the regional leadership's commitment to getting external stakeholders to the table--even those outside of CMS' realm of regulatory oversight--as a critical factor to improving outcomes. She further described frequent, effective communication of performance information among stakeholders as a means to getting them to work together. See figure 13 for more information on these efforts. Figure 13: CMS Region IV Communicated Performance Information with Stakeholders to Improve Quality of Care in Nursing Homes: [Refer to PDF for image: illustration and accompanying text] Performance goal: Reduce incidence of pressure ulcers by 11%. Collaboration on goals and shared performance information by External stakeholders and Internal stakeholders. Performance goal achieved: 2,441 fewer long-stay residents with pressure ulcers[A]. When Region IV was charged with reducing the prevalence of pressure ulcers among nursing home residents by more than 11 percent[A]”an outcome the region could not directly control”a Region IV official said ’we knew we had to try something different.“ According to the Associate Regional Administrator, this different approach included collaboration with the internal and external stakeholders that could contribute to improved performance against their regional pressure ulcer reduction goal. They engaged hospital and nursing home personnel, patient advocates, emergency medical technicians, quality improvement organizations, state survey agencies, and others. They shared performance information about the problem and collaborated on possible causes and solutions. According to CMS regional officials, collaboration on goals and shared performance information were among key factors in bringing about improvements in the prevalence of pressure ulcers among nursing home residents in each of the eight states in their region. In Region IV, between fiscal years 2006 and 2008, this improvement translated into nearly 2,500 fewer long-stay nursing home residents with pressure ulcers. Source: GAO analysis of CMS data. [A] In the 2-year period during fiscal years 2006-2008, Region IV reduced the percentage of long-stay nursing-home residents with pressure ulcers from 9.3 percent to 8.3 percent, which represented an 11 percent decrease, or 2,441 fewer cases. [End of figure] According to several headquarters officials we interviewed, as a result of leadership support that emphasized communication internally and externally and other key drivers, CMS has undergone a culture change that places a greater emphasis on using performance data to achieve results. CMS Managers Reported Alignment among Agency, Program, and Individual Performance Goals: Our survey results indicated that between 2000 and 2007, a significantly greater percentage of CMS managers reported that they were held accountable for program results (see figure 14). Figure 14: Percentage of CMS Managers Who Reported That Agency Managers at Their Level Are Held Accountable for the Results of Their Programs: [Refer to PDF for image: horizontal bar graph] Survey question: Agency managers/supervisors at my level are held accountable for the results of the program(s) they are responsible for. Percentage responding to a ’great“ or ’very great“ extent: 2000: 42%; 2007: 77%. Source: GAO. Note: Data are from GAO 2000 and 2007 surveys. [End of figure] Top CMS headquarters officials said that a new performance-management system that required linkages between organizational, program, and individual goals had made individual accountability for program results more explicit. In 2006, CMS began to implement the Department of Health and Human Services' four-tiered system, Performance Management Appraisal Program (PMAP), for non-SES employees at the agency.[Footnote 33] Under PMAP, employees are held accountable for both administrative and program performance. Top CMS officials described how agency goals and objectives were embedded in the Administrator's performance agreement and cascaded down through the management hierarchy, so that each level of management understood their accountability for achieving the agency's broad goals. For example, broad goals for preventive health care cascaded from the Department of Health and Human Services to a Health Insurance Specialist in CMS' Office of Clinical Standards and Quality (OCSQ), who was responsible for communications to raise awareness among beneficiaries (see figure 15). Figure 15: CMS Reported Alignment among Department and Agency Goals and Individual Performance Objectives: [Refer to PDF for image: illustration] Strategic goal/performance goal: Department of Health and Human Services: Improve the safety, quality, affordability, and accessibility of health care including behavioral health care and long-term care; Strategic objective/performance objective: * Promote and encourage preventive health care including mental health, lifelong healthy behaviors, and recovery. Strategic goal/performance goal: CMS: Improve early detection of breast cancer; Strategic objective/performance objective: * Improve early detection of breast cancer among Medicare beneficiaries age 65 years and older by increasing the percentage of women who receive a mammogram. Strategic goal/performance goal: Director, Quality Improvement Group, OCSQ: Increase in utilization of Medicare preventive benefits (defined as influenza vaccination, screening mammography, and initial preventive services examination); Strategic objective/performance objective: * Achieve 2 percent increase in mammography for the targeted participating providers by end date of this work period; * Increase provider awareness of the importance of preventive services by educating 80 percent of targeted physician groups for the core prevention theme related to screening mammography by end date of this period. Strategic goal/performance goal: Supervisor, Health Insurance Specialist, OCSQ: Promote and encourage preventive health care, including mental health, lifelong health behaviors and recovery specifically through collaborative partnerships; Strategic objective/performance objective: * Fosters a team atmosphere whereby effective communications and collaboration are promoted both within and outside of the division; * Ensures effective staff communications and coordination in addressing media and congressional requests; recommends one proactive communication effort per quarter to benefit beneficiary awareness or stakeholders, or both, through outreach. Source: GAO analysis of CMS information. [End of figure] CMS Officials Said Selecting a Vital Few Performance Measures with Relevance to Managers Encouraged Use of Performance Information: Our survey results show that between 2000 and 2007, there was a significant decline in the percentage of CMS managers who reported that difficulty developing meaningful measures was a hindrance to using performance information (see figure 16). Figure 16: Percentage of CMS Managers Who Reported That Difficulty Determining Meaningful Measures Hinders Using Performance Information: Survey question: Difficulty determining meaningful measures hinders measuring performance or using the performance information. Percentage responding to a ’great“ or ’very great“ extent: 2000: 65%; 2007: 48%. Source: GAO. Note: Data are from GAO 2000 and 2007 surveys. [End of figure] Our interviews with CMS officials provided insight into steps they had taken to ensure that performance information was useful to managers. They said they selected measures for GPRA reporting purposes that were useful for decision making and limited the number of measures to the 31 that represented the agency's priorities. According to an official responsible for strategic planning and performance reporting, because quality of care is a top priority for CMS, many of their measures--such as incidence of pressure ulcers among nursing-home residents--reflect this focus. This official noted that it would be unmanageable to measure and report on every aspect of their programs and processes. They ultimately settled on a set of performance goals that helped managers and staff identify performance gaps and where there are opportunities to improve performance to close the gaps. Improved Data Systems and Training Opportunities Enhanced CMS' Capacity to Identify Problems and Devise Solutions: At CMS, our prior work identified the need for the agency to develop better management information systems, among other actions, to improve oversight of nursing-home quality and safety.[Footnote 34] More recently, we acknowledged that CMS had pursued important upgrades in the system used to track the results of state survey activities and has increased its analyses of data to improve oversight.[Footnote 35] This is consistent with CMS finance and administrative officials' statements that the agency had invested in systems infrastructure to manage using performance information. A CQISCO manager responsible for Survey & Certification performance reporting in Region IV noted that easier access to performance information through improved data systems had enabled them to use performance information more effectively to identify problems and develop solutions. CQISCO managers and staff can now use more than 350 standard performance reports in exercising their oversight of state agencies responsible for surveying nursing-home quality. For example, a CQISCO Region IV official told us that they use many of these performance reports at quarterly meetings with state survey officials and that the performance information is helpful in identifying aberrant trends and illustrating these trends to the states. They also compare performance to prior years to determine whether positive outcomes have occurred. CQISCO Region IV officials provided an example of how a report on the frequency of state nursing- home quality surveys, which includes information that had not been easily accessible several years ago, helped them to improve outcomes in a particular state (see figure 17). Figure 17: A CMS Region IV Manager Described How Easier Access to Performance Data Contributed to Improved Nursing-Home Survey Frequency in Alabama: [Refer to PDF for image: horizontal bar graph and accompanying text] Every nursing home receiving Medicare or Medicaid payment must be inspected by the state survey agency against federal quality-of-care and fire safety standards not less than once every 15 months.[A] Performance is measured in terms of the percentage of surveys a state conducts within this required interval. In CMS‘s Southeast Region, a Survey & Certification Division official said that investments in data systems made it easier to obtain the performance information that helped them improve outcomes related to Alabama‘s nursing-home oversight activities. In the middle of fiscal year 2007, the regional Survey & Certification team was reviewing a performance report on the frequency of Alabama‘s nursing home surveys and found that over 90 percent had not met standards for timeliness. When states fail to conduct timely surveys, there is a risk that quality of care issues, such as preventing avoidable pressure ulcers, weight loss, or accidents, will go undetected in between surveys. In response, the team put together a ’Request for Action Plan“ to Alabama‘s State Survey Agency that included a range of data reports on the state‘s performance. ’When the state agency sees the problems presented in black and white, it really encourages them to take action,“ the manager said. The team continued to provide the information to the state and, over time, fewer and fewer surveys missed the timeliness standard. Partway into fiscal year 2009, CMS‘s reports on Alabama‘s nursing-home survey frequency showed 100 percent compliance with the maximum survey interval standard (see fig. below). Several years ago, the official noted, the information on state survey frequency was not as easy to access. ’These days, it doesn‘t take a technical expert to run these reports or interpret the information”they‘ve been designed so that managers and staff can use them.“ According to the official, improved availability of performance information has contributed to improved results. Percentage of Alabama nursing-home quality surveys that did not meet standard for timeliness, FY2007–FY2009: Fiscal year: 2007; Percentage: 76.6. Fiscal year: 2008; Percentage: 39.4. Fiscal year: 2009[B]; Percentage: 0. [A] In addition, the statewide average interval for these surveys must not exceed 12 months. CMS generally interprets these requirements to permit a statewide average interval of 12.9 months and a maximum interval of 15.9 months for each home. 42 U.S.C. § 1395i-3(g). [B] As of February 2009. Source: GAO analysis of CMS data. [End of figure] According to our prior work, inadequate staff expertise may have hindered CMS from using performance information in monitoring state performance of nursing-home oversight.[Footnote 36] In 2003, officials in three regions said lack of staff expertise, among other issues, prevented them from using reports that were available to aid them in overseeing state survey activities. Our survey results, along with recent interviews with several CMS officials, indicate that the agency has taken steps to develop its staff's capacity to use performance information. Between 2000 and 2007, there was a significant positive increase on all six survey questions related to managers' access to training over the past 3 years on use of performance information for various activities (see figure 18). Figure 18: Percentage of CMS Managers Who Reported That Training Was Provided to Help Accomplish Key Management Tasks: Survey question: Link the performance of program(s)/operation(s)/ project(s)to the achievement of agency strategic goals. Percentage responding to a ’great“ or ’very great“ extent: 2000: 22%; 2007: 54%. Survey question: Use program performance information to make decisions; Percentage responding to a ’great“ or ’very great“ extent: 2000: 17%; 2007: 34%. Survey question: Assess the quality of performance data; Percentage responding to a ’great“ or ’very great“ extent: 2000: 9%; 2007: 24%. Survey question: Develop program performance measures; Percentage responding to a ’great“ or ’very great“ extent: 2000: 15%; 2007: 49%. Survey question: Set program performance goals; Percentage responding to a ’great“ or ’very great“ extent: 2000: 29%; 2007: 58%. Survey question: Conduct strategic planning; Percentage responding to a ’great“ or ’very great“ extent: 2000: 38%; 2007: 57%. Source: GAO. Note: Data are from GAO 2000 and 2007 surveys. [End of figure] Key officials in CMS' headquarters told us that the agency had provided training on a range of topics related to performance measurement, such as "The Government Performance Logic Model," "Aligning Project Management with Organizational Strategy," and "Strategic Planning and Performance Measurement." According to one CQISCO Region IV official, increasing her staff's skills in conducting analyses of performance information and presenting findings was a gradual process that required training, coaching, and guidance. She said that there are one to two formal training opportunities for staff every year and that the Consortium holds quarterly meetings that address topics such as how to analyze data to identify problems. Another key approach was to hold staff individually accountable for using performance information in reports they present to senior management. Additionally, in 2006, Region IV's Survey & Certification Division began to include a performance element in all of its managers' performance agreements related to performance information use. Conclusions: It has been more than 16 years since Congress passed GPRA in an effort to ensure that federal agencies have the infrastructure and tools they need to improve results. Across the federal government, agencies have developed and implemented strategic plans and are routinely generating performance information to report progress toward their strategic goals. Our survey of government managers, however, showed that GPRA's legislative requirements and other performance improvement initiatives are not sufficient to ensure that managers will actually use performance information to manage for results. At FEMA and Interior, inconsistencies or weaknesses in key management practices at these agencies--such as demonstrating leadership commitment, aligning goals, ensuring usefulness of performance measures, and building analytic capacity--appeared to hinder their use of performance information. At FEMA, despite their strong commitment to achieving mission results, some of the officials we interviewed did not demonstrate the same level of commitment to using performance information in decision making, especially in the face of unpredictable circumstances such as natural disasters. Strengthening the commitment of FEMA's leaders alone, however, would not be enough to ensure that managers throughout the agency are well-positioned to use performance information to manage for results. The agency also had gaps in performance information, such as immature performance measurement at the regional level, which made it difficult to establish a line of sight over strategic, program, regional, and individual performance goals. These gaps, coupled with a lack of a performance-management system that required goal alignment, made it challenging for managers to hold individuals accountable for achieving results. Furthermore, FEMA faces other hurdles, such as a lack of trained staff and inadequate information systems, to ensure that performance information can be easily collected, communicated, and analyzed. However, despite the presence of these challenges, there are some emerging efforts that FEMA's leadership could build on, such as consistent and timely regional reporting against performance goals. At Interior and NPS we observed a management culture where performance information was primarily used for after-the-fact accountability and reporting purposes, but not as a tool for improving results. Leaders at Interior and NPS were not effectively communicating how, if at all, they used performance information to identify performance gaps and develop strategies to better achieve results. Instead, the greater emphasis on using performance information to hold individuals accountable for achieving goals may be contributing to the perception that it is being used to punish poor performers rather than to improve overall performance. Leaders who do not effectively strike a balance among such uses of performance information run the risk of creating perverse incentives where managers are afraid to fail rather than inspired to succeed. Under GPRA and PART, Interior collectively tracks nearly 650 performance measures, which made it difficult for leadership and management at all levels to focus on critical priorities. Even where there was management commitment to using performance information proactively, some bureau-level managers at NPS and Reclamation said GPRA and PART measures were not useful for decision making, either because there were too many or they were not credible. A labor- intensive, cumbersome performance-information system further hindered NPS and Reclamation managers' efforts to use performance information to inform their decision making. Interior's more recent focus on key performance indicators and NPS' efforts to develop more useful performance information for park-level decision making could be the foundation for further improvements. Our review of selected areas at CMS indicated a possible roadmap for agencies seeking to increase their use of performance information to improve results. There, it was clear that agency leaders were committed to using performance information and they made it a priority to build the analytic capacity to do so. Managers credited the use of the very same management practices that were weak or inconsistent at our other case agencies with helping improve their ability to manage for results. Although these managers noted that they could not control ultimate quality-of-care outcomes--such as the incidence of pressure ulcers among nursing-home patients--they said that communicating performance information to states, nursing homes, and other stakeholders helped them to work collaboratively to improve results. Managing for results will become even more critical under the Recovery Act: both Interior and FEMA have seen significant increases in their funding; in the case of Reclamation, funding has more than doubled. With the economic health of the nation at stake, federal managers of these and other agencies will be expected to allocate these resources to achieve critical results. Recommendations for Executive Action: The Secretary of DHS should direct the Administrator of FEMA to take the following three actions: 1. direct agency leadership to demonstrate its commitment to using performance information for decision making by reviewing performance results with subordinate managers on a regular and recurring basis and communicating decisions based on performance information to show that performance information is reviewed and acted upon; 2. augment FEMA's analytic capacity to collect and analyze performance information by: a. continuing to build upon recent efforts to provide training to directorate and regional managers to enhance their use of performance information, which includes topics such as strategic planning, developing robust performance measures, and analyzing what the performance data mean; and: b. reviewing performance information systems to address users' needs for integrated, timely, and relevant performance information; 3. improve linkages among agency, program, and individual performance by: a. continuing to engage program and regional managers in efforts to develop, and where appropriate refine, intermediate, measurable performance targets that cascade from agency strategic goals; and: b. in the absence of a DHS-wide performance-management system, developing interim guidance for FEMA's current performance-appraisal system, covering supervisors and managers, on how to align individual performance objectives with program and agency goals. Such guidance could include information on how work plans can be used to align individual and agency performance goals and objectives, examples of alignment from subunits within FEMA that are already implementing this practice, or other approaches to promoting such alignment. The Secretary of the Interior should take the following two actions: 1. direct departmental leadership and the Director of NPS to demonstrate their commitment to using performance information for decision making by reviewing performance results with subordinate managers on a regular and recurring basis and communicating decisions based on performance information to show that performance information is reviewed and acted upon; 2. direct departmental leadership, the Director of NPS, and the Commissioner of Reclamation in conjunction with OMB to review the usefulness of their performance measures and refine or discontinue performance measures that are not useful for decision making. The review should also consider options for reducing the burden of collecting and reporting performance information. This review should involve managers at all levels to take into account their differing needs for performance information. Agency Comments and Our Evaluation: We provided a draft of this report to the Secretary of DHS, the Secretary of the Interior, and the Secretary of the Department of Health and Human Services for comment. We received written comments from DHS and Interior, which are reprinted in appendices IV and V. The Department of Health and Human Services provided only technical comments, which were incorporated as appropriate. DHS concurred with two of our three recommendations and partially concurred with the third. DHS also noted efforts underway that may ultimately address some of these recommendations, such as FEMA's recently established Performance Improvement Council, which is working in conjunction with other FEMA teams to address performance issues across the agency. With regard to our third recommendation concerning improved linkages among agency, program, and individual performance goals, DHS agreed that FEMA should continue to develop or refine intermediate performance targets that cascade from agency strategic goals. However, DHS did not concur that, in the absence of a DHS-wide performance-management system, FEMA should develop an interim plan to allow managers to hold staff accountable for the accomplishment of agency strategic goals. DHS commented that an interim plan is not advisable at this time because the current FEMA Employee Performance System covers both non-bargaining and bargaining unit employees, and an interim performance management system would require FEMA to support two systems until it could be bargained. Since DHS appeared to interpret our recommendation to mean that FEMA should develop and maintain one or more interim systems, which was not our intention, we contacted officials from both agencies for clarification. Upon further discussion with FEMA human capital officials, we revised the language of our recommendation to state that the agency should develop interim guidance for its current performance appraisal system for managers and supervisors. FEMA officials agreed that developing interim guidance for its managers and supervisors would not require them to develop dual systems or to negotiate with its employee labor unions. They further acknowledged that such guidance could help lay the groundwork for implementing DHS' future performance management system and indicated their willingness to address the recommendation as clarified. Interior agreed in principle with the recommendations in the report and noted that it is in the process of revising its strategic plan in a manner that will be responsive to our recommendations. The department also provided additional comments on some of the findings contained in the report. First, Interior noted that Bureau of Reclamation, rather than Interior, sets goals related to water delivery. In response, we revised our report to indicate that Interior included--rather than set-- a goal for this measure for fiscal years 2009 and 2010, and reported on the results of this goal for the past four years in its fiscal year 2008 Performance Accountability Report (PAR). Interior also commented that the specific water delivery measure we discussed as lacking credibility is provided only for reference at the end of Interior's PAR and is not emphasized. While it is true that this measure is shown at the end of the report, it is nonetheless labeled as a performance measure and Reclamation must collect and report the performance data, even though both Interior and Reclamation officials agreed that these data were not used in decision making. Second, Interior expressed concern related to our finding that at both NPS and Reclamation, managers we interviewed described certain performance measures as lacking credibility. Interior stated that they would like to continue to emphasize outcome measures and suggested that our report associated long-term measures with a lack of credibility. In addition, Interior commented that our report implies a need for an increase in specific, narrower-focused measures to improve credibility and use among managers. We disagree that we equated long-term measures with low credibility among managers. As we reported, managers we interviewed said that credibility issues were a result of measures that do not accurately define comparable elements or do not take into account different standards across bureaus or units. Furthermore, we agree that Interior should continue to emphasize outcome measures, and do not believe that our report encourages adoption of additional, narrow measures as a solution to the credibility issues we noted. On the contrary, one of our findings was the large volume of performance measures may be hindering managers' efforts to relate short-term or disaggregated performance targets to longer-term outcome goals. As we stated in the report, limiting measures to the vital few at each organizational level can ensure there is useful, meaningful performance data available to managers and staff across an agency. Interior also provided additional perspective on our recommendation that it work with NPS, Reclamation, and OMB to review the usefulness of their performance measures. As agreed with your offices, unless you publicly announce the contents of this report earlier, we plan no further distribution until 30 days from the date of this letter. We will then send copies of this report to the Secretary of DHS, the Secretary of the Interior, and the Secretary of the Department of Health and Human Services, and other congressional committees interested in results-oriented government and management issues at DHS, Interior, and CMS. In addition, the report will be available at no charge on GAO's Web site at [hyperlink, http://www.gao.gov]. If you or your staff have any questions about this report, please contact me at (202) 512-6806 or at steinhardtb@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. Individuals who made key contributions to this report are listed in appendix VI. Signed by: Bernice Steinhardt: Director, Strategic Issues: [End of section] Appendix I: Objectives, Scope, and Methodology: To develop a better understanding of what practices may inhibit or promote the use of performance information in managerial decision making our objectives were to: (1) identify agencies with relatively low use of performance information and the factors that contribute to this condition; and (2) examine practices in an agency where there were indications of improvement in its use of performance information. We conducted a review of three agencies to identify what barriers and challenges hinder managers' use of performance information and, in addition for objective two, what practices appear to have contributed to agency efforts to improve managerial use of performance information. To address both of our objectives, we began by examining the results of our four surveys of performance and management issues,[Footnote 37] reviewing our prior reports and other relevant materials on results- oriented management, the Government Performance and Results Act of 1993 (GPRA), and the Office of Management and Budget's (OMB) Performance Assessment Ratings Tool (PART). We also compared management practices at selected agencies with those we have previously identified as enhancing the use of performance information for decision making. [Footnote 38] To identify agencies for our review, we used our survey results from 2000 and 2007 since both were designed to provide analysis of data at the agency and department level as well as governmentwide. For purposes of selecting agencies for review under objective one, we used our 2007 survey data to calculate an agency average "core uses" index score to identify agencies with a lower percentage of managers reporting performance information use relative to their counterparts at other agencies. See appendix II for detailed information on how we developed the core uses index score and a table showing each agency's ranking and index score. In deciding which agencies to select, we focused on those agencies with the lowest ranking. For that subset of agencies, we reviewed additional performance-related material such as the Office of Personnel Management's Federal Human Capital Survey, and OMB's PART assessments to aid us in the selection process. Since we expected that our review could potentially produce recommendations concerning agency-specific performance management issues we observed in the course of our review, we also conducted an environmental scan of recent and ongoing GAO research at our potential case-study agencies to avoid duplicating similar work.[Footnote 39] Based on these considerations, we selected the Department of the Interior (Interior) and Federal Emergency Management Agency (FEMA). In the initial phase of our review of these two agencies, we conducted interviews with senior-level officials responsible for operations, budget, human resources, and performance-reporting functions at each agency to gain an understanding of their performance-based management policies and practices established at the top levels of the organization. To help guide us in determining where to further focus our review, we also asked officials and managers to identify particular areas that could provide illustrative examples of the challenges and difficulties they faced in using performance information for decision making. Based on these initial interviews and other information, we then identified programs and operational areas for more in-depth review. At Interior, we selected the National Park Service (NPS) and Bureau of Reclamation (Reclamation) for review. At both NPS and Reclamation, we interviewed senior-level officials responsible for operations, budget, policy, human resources, and performance-management functions. At NPS, we interviewed management and program officials from four of seven regions: Pacific West, National Capital, Southeast, and Intermountain Regions. We interviewed park superintendents and program managers from two parks within each of the four regions for a total of eight parks. Of the four regions and eight parks, we conducted site visits at three of the regional management offices, and four national parks; the other interviews were conducted over the phone. In our sample, we included a mix of small, medium, and large parks, which we defined by the size of their operating budget: (1) small parks had an operating budget of less than $1 million; (2) medium parks had an operating budget of $1 million to $10 million; and: (3) large parks had an operating budget of more than $10 million. We also selected parks on the basis of their diversity across resource types that included national historic parks, national battlefields, national recreation areas, and national seashores. We interviewed senior and line managers from three of five of their major national program areas, including Natural Resources, Stewardship, and Science programs; Cultural Resources programs; and Park Planning, Facilities, and Lands programs. At Reclamation, of the five regions, we visited the Mid Pacific Region and spoke to officials from two other regions: the Upper Colorado, and Great Plains Regions. Within these three regions, we interviewed the regional director, GPRA Coordinator, and one area or project manager. These three regions are geographically diverse and include a range of operations and projects including new construction, dams, and power- generating plants. At FEMA, we followed a similar procedure of meeting with senior-level officials and soliciting their suggestions for areas or operations that were demonstrative of the challenges and difficulties of using performance information in managerial decision making. At FEMA, we interviewed top officials and program managers from two of their eight Directorates along with regional officials and line managers responsible for general operations as well as Mitigation and Disaster Assistance Directorate program delivery in 3 of FEMA's 10 regions: Region I in Boston, Region IV in Atlanta, and Region IX in San Francisco. To address our second objective of identifying an agency where improvement in the managers' use of performance information appears to have progressed, we used results from a set of selected items that we asked on both the 2000 and 2007 survey. For each agency, we calculated the difference between the percent of agency managers reporting their use of performance information to a great or very great extent in 2000 and 2007 across a set of nine items that addressed the use of performance information in areas of managerial decision making such as setting program priorities, allocating resources, and setting or revising work goals. We then calculated the average of these nine differences as an overall descriptive indicator of change in each agency's managers' reporting on their use of performance information. [Footnote 40] See figure 19 for the results of this analysis. Figure 19: Average Change in Percentage of Federal Managers Reporting Use of Performance Information to a Great or Very Great Extent, 2000- 2007: [Refer to PDF for image: vertical bar graph] Agency: NSF; Average change 2000 to 2007: 29.1%. Agency: NRC; Average change 2000 to 2007: 21.3%. Agency: CMS; Average change 2000 to 2007: 20.9%. Agency: FAA; Average change 2000 to 2007: 19.9%. Agency: EPA; Average change 2000 to 2007: 18.0%. Agency: Education; Average change 2000 to 2007: 15.9%. Agency: Energy; Average change 2000 to 2007: 15.8%. Agency: IRS; Average change 2000 to 2007: 14.6%. Agency: SSA; Average change 2000 to 2007: 13.6%. Agency: FEMA; Average change 2000 to 2007: 11.9%. Agency: NASA; Average change 2000 to 2007: 11.3%. Agency: DOT; Average change 2000 to 2007: 11.1%. Agency: AID; Average change 2000 to 2007: 11.1%. Agency: State; Average change 2000 to 2007: 10.1%. Agency: Labor; Average change 2000 to 2007: 9.15%. Agency: Forest Service; Average change 2000 to 2007: 5.6%. Agency: VA; Average change 2000 to 2007: 5.1%. Agency: DOJ; Average change 2000 to 2007: 3.9%. Agency: Treasury; Average change 2000 to 2007: 3.8%. Agency: GSA; Average change 2000 to 2007: 2.7%. Agency: USDA; Average change 2000 to 2007: 2.3%. Agency: HUD; Average change 2000 to 2007: 2.3%. Agency: Interior; Average change 2000 to 2007: 1.9%. Agency: Commerce; Average change 2000 to 2007: 1.4%. Agency: DOD; Average change 2000 to 2007: -0.7%. Agency: OPM; Average change 2000 to 2007: -1.8%. Agency: HHS; Average change 2000 to 2007: -3.2%. Agency: SBA; Average change 2000 to 2007: -9.6%. Source: GAO. Note: Data are from GAO 2007 survey. [End of figure] Similar to our selection procedures for objective one, we focused on those agencies showing the greatest positive shift in the percent of managers endorsing their use of performance information in their decision making. We then reviewed additional performance-related material for that subset of agencies showing the greatest change. In selecting an agency for review, we considered various factors such as agency size, mission, and workforce mix with a view that an agency facing substantive barriers and challenges to the use of performance information would yield good illustrative case examples of change. We selected the Centers for Medicare & Medicaid Services (CMS). In our 2000 survey, on five items asking managers about the extent to which they had five different types of performance measures, for example, outcome, output, customer service, CMS[Footnote 41] had the lowest ranking of all agencies on four of the five items.[Footnote 42] While CMS has shown notable change in its managers' reported use of performance information, its relative standing in 2007 as reflected in appendix II, figure 20, shows it to be in about the middle of the distribution of the 29 agencies we ranked. At CMS, we interviewed senior-level officials to gain insight into how they were able to improve their managers' use of performance information as indicated by our survey results. We asked agency officials to help us identify regional-and line-manager interview subjects who could articulate their experiences with such a change. Based on this input, we interviewed the top officials from two of CMS' four Consortia: the Consortium for Quality Improvement and Survey & Certification Operations and the Consortium for Financial Management and Fee for Service Operations. We also interviewed regional officials and managers responsible for Survey & Certification in Region IV and regional officials responsible for administration and outreach in Region IX. At our interviews, we asked officials to identify the barriers and challenges they faced in the use of performance information and strategies for overcoming them. At all the agencies where we conducted our work, our interviews and examination of agency documentation such as strategic plans, GPRA measures, and performance agreements incorporated a review of the extent to which these organizations implemented practices that our work has shown can promote the use of performance information for management decision making. We also examined our 2007 survey results for these agencies to explore what differences in agency responses could be useful in identifying conditions or perceptions that were relevant to or elaborated on our observations. Throughout the body of the report, the differences in percentages are significant unless noted otherwise. Although we reviewed available performance information in the examples we cite and describe how performance information was used to make decisions, we did not attempt to independently assess the reliability of the information or verify that the use resulted in improved outcomes, since they were not within the scope of our review objectives. We performed our work in the Washington, D.C., metropolitan area; Boston, Massachusetts; San Francisco and Sacramento, California; Atlanta, Georgia; from March 2007 to May 2009 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix II: Agency Rankings Based on Index of 2007 Survey Results: As part of our analyses of the 2007 survey data, we identified a set of nine items from the questionnaire that inquired about uses of performance information that we identified in a previous GAO report. [Footnote 43] Using those items, we developed an index that reflected the extent to which managers' perceived their own use of performance information for various managerial functions and decisions as well as that of other managers in the agency. To obtain an index score of reported use of performance information, we computed an average score for each respondent across the nine items we identified. We then averaged the respondent scores from each agency to produce an overall index score for each agency. By using this average index score, which yields values in the same range as the 5-point extent scale used on each item, we were able to qualitatively characterize index score values using the same response categories used for the items constituting the index.[Footnote 44] Figure 20 shows the relative ranking on the index score for each agency in the 2007 survey. Figure 20: Agency Ranking Based on 2007 Survey Results on Use of Performance Information: [Refer to PDF for image: list] Rank: 1; Agency/Component: National Aeronautics and Space Administration. Rank: 2; Agency/Component: Nuclear Regulatory Commission. Rank: 3. Agency/Component: Department of Veterans Affairs. Rank: 4; Agency/Component: Social Security Administration. Rank: 5; Agency/Component: National Science Foundation. Rank: 6; Agency/Component: General Services Administration. Rank: 7; Agency/Component: Department of Energy. Rank: 8; Agency/Component: Department of Housing And Urban Development. Rank: 9; Agency/Component: Department of Education. Rank: 10; Agency/Component: Department of the Treasury (excluding Internal Revenue Service). Rank: 11; Agency/Component: Internal Revenue Service. Rank: 12; Agency/Component: Environmental Protection Agency. Rank: 13; Agency/Component: Small Business Administration. Rank: 14; Agency/Component: Centers for Medicare & Medicaid Services. Rank: 15; Agency/Component: Department of Commerce. Rank: 16; Agency/Component: Office of Personnel Management. Rank: 17; Agency/Component: Federal Aviation Administration. Rank: 18; Agency/Component: Agency for International Development. Rank: 19; Agency/Component: Department of Labor. Rank: 20; Agency/Component: Department of Agriculture (excluding Forest Service). Rank: 21; Agency/Component: Department of Homeland Security (excluding Federal Emergency Management Agency). Rank: 22; Agency/Component: Department of Defense. Rank: 23; Agency/Component: Department of State. Rank: 24; Agency/Component: Department of Transportation (excluding Federal Aviation Administration). Rank: 25; Agency/Component: Department of Justice. Rank: 26; Agency/Component: Department of Health and Human Services (excluding Centers for Medicare & Medicaid Services). Rank: 27; Agency/Component: Department of the Interior. Rank: 28; Agency/Component: Federal Emergency Management Agency. Rank: 29; Forest Service. Source: GAO. [End of figure] [End of section] Appendix III: Timeline of Major Government Results-Oriented Management Reforms: 1993: Congress enacted the Government Performance and Results Act of 1993 (GPRA) (Public Law 103-62) to address several broad purposes, including improving federal program effectiveness, accountability, and service delivery; and enhancing congressional decision making by providing more objective information on program performance. GPRA requires executive agencies to complete strategic plans in which they define their missions, establish results-oriented goals, and identify the strategies that will be needed to achieve those goals. GPRA also requires executive agencies to prepare annual performance plans that articulate goals for the upcoming fiscal year that are aligned with their long-term strategic goals. Finally, GPRA requires executive agencies to measure performance toward the achievement of the goals in the annual performance plan and report annually on their progress in program performance reports. 1993: The Clinton administration launched the National Performance Review, later renamed the National Partnership for Reinventing Government which was intended to transform the federal government to be more results-oriented, performance-based, and customer-focused by among other things requiring consideration of employee and customer views and increasing the use of technology, especially the Internet, for delivery of services and information to the public. 1994-1996: Congress passed the Government Management Reform Act of 1994 (Public Law 103-356), Paperwork Reduction Act of 1995 (Public Law 104- 13), and Clinger-Cohen Act of 1996 (Public Law 104-106, div. D, E), which together provide a framework for developing and integrating information about agencies' missions and strategic priorities, the results-oriented performance goals that flow from those priorities, performance data to show the level of achievement of those goals, and the relationship of reliable and audited financial information and information technology investments to the achievement of those goals. 2001: With the President's Management Agenda, the Bush administration attempted to resolve long-standing federal management weaknesses by establishing five governmentwide management priorities including performance-budget integration and improved financial reporting. 2002: The Office of Management and Budget (OMB) created the Program Assessment Rating Tool (PART), a diagnostic tool that is intended to provide a consistent approach for evaluating federal programs as part of the executive budget formulation process. Through PART, OMB sought to create better ties between program performance and the allocation of resources. 2003: OMB issued changes to Circular A-11 requiring agencies to submit "performance budgets" in lieu of annual performance plans for their fiscal year 2005 budget submission to OMB and Congress. [End of section] Appendix IV: Comments from the Department of Homeland Security: U.S. Department of Homeland Security: Washington, DC 20528: July 7, 2009: Ms. Bernice Steinhardt: Director Strategic Issues: United States Government Accountability Office: 441 G Street, NW: Washington, DC 20548: Dear Ms. Steinhardt: Thank you for the opportunity to review and comment on the Government Accountability Office's (GAO's) Draft Report GAO-09-676 entitled Results Oriented Management: Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information. FEMA has reviewed the Draft Report and concurs with Recommendations 1, 2 a., 2 b., and 3 a. and non-concurs with Recommendation 3 b. The Secretary of DHS should direct the Administrator of FEMA to: Recommendation 1: direct agency leadership to demonstrate their commitment to using performance information for decision making by reviewing performance results with subordinate managers on a regular and recurring basis and communicating decisions based on performance information to show that performance information is reviewed and acted upon. Response: Concur. FEMA recently formed a Performance Improvement Council (PIC) to address program performance issues across the agency. This particular working group is made up of individuals within each program, office, and directorate that coordinate all aspects of program performance management within their spheres of influence. The PIC works in conjunction with FEMA's Investment Working Group (IWG) who, in turn, makes recommendations to FEMA's Investment Review Board, which comprises senior leadership, on relevant courses of action based on input from the PIC. The IWG coordinates many of the agency's cross- programmatic resource allocation decisions at various stages of the budget process. Recommendation 2: augment FEMA's analytic capacity to collect and analyze performance information by: a. continuing to build upon recent efforts to provide training to directorate and regional managers to enhance their use of performance information, which included topics such as strategic planning, developing robust performance measures, and analyzing what the performance data mean; and; b. reviewing performance information systems to address users' needs for integrated, timely, and relevant performance information. Response: Concur. The Office of Policy & Program Analysis, Program Analysis and Evaluation Division, chairs the FEMA PIC. The PIC is a recently developed cross-functional and cross-organizational group created to address program performance related issues across the Agency in order to improve programmatic outcomes. Most recently, the PIC provided performance measurement training to all FEMA Programs, Offices and Directorates to enhance that particular competence among these FEMA entities. Also, to address paragraph b. above--one of the elements of this particular working group's emphasis will be ensuring that information systems requirements to capture relevant data and information are identified and pursued. Recommendation 3: improve linkages among agency, program, and individual performance by: a. continuing to engage program and regional managers in effort to develop, and where appropriate, refine intermediate, measurable performance targets that cascade from agency strategic goals, and; Response: Concur. Recommendation: b. secondly, in the absence of a department-wide performance management system, developing an interim plan that allows managers to hold staff accountable for the accomplishment of agency strategic goals by providing them with a tool they can use to measure their staffs' performance against goals. Response: Non-concur. An interim plan is not advisable at this time, since the current FEMA Employee Performance System covers both non- bargaining and bargaining unit employees. An interim performance management system would require FEMA to support two systems until it could be bargained. It should be noted that during FY 2009, DHS Chief Human Capital Officer expects both the new Performance Management System Directive and Instructions to be sent for vetting to the components and the Request for Proposal for a new software system to be submitted for bidding. Thank you again for the opportunity to comment on this Draft Report and we look forward to working with you on future Homeland Security issues. Sincerely, Signed by: Jerald E. Levine: Director: Departmental GAO/OIG Liaison Office: [End of section] Appendix V: Comments from the Department of the Interior: Note: Page numbers in the draft report may differ from those in this report. United States Department of the Interior: Office Of The Secretary: Washington, DC 20240: June 29, 2009: Bernice Steinhardt: Director, Strategic Issues: U.S. Government Accountability Office: 441 G Street, N.W. Washington, D.C 20548: Dear Ms. Steinhardt: Thank you for providing the Department of the Interior the opportunity to review and comment on the draft Government Accountability Office Report entitled, "Results-Oriented Management: Strengthening Key Practices at FEMA and Interior Could Promote Greater Use of Performance Information," (GAO-09-676). We agree with the GAO recommendations in principle, so that the Department of the Interior will be working to: * Demonstrate commitment throughout all levels of management and leadership in promoting and practicing the consideration of performance as a factor in decision making. * Review and revise its performance measures to better align them with agency activities and ensure the appropriate balance of output/activity based performance measures, that are recognizable to program managers, and related to longterm/outcome performance measures. The Department is currently in the process of revising its strategic plan including a detailed review of the measures in a manner that will be responsive to the recommendations. The enclosure provides additional comments on the report regarding: * A clarification that the bureaus set their own performance targets; * Concerns about equating increasingly specialized measures with improving credibility, i.e., the value of long-term measures; * The need to better balance the application of program measures including over 400 PART measures; and; * Concerns about how measures are applied through the use of performance information. The time and effort put forth by the GAO team in conducting this analysis and preparing these observations and recommendations is much appreciated. We hope to pursue these recommendations with recognizable results in time for the next GAO report on the use of performance information. If you have any questions, or need additional information, please contact Richard Beck, Director, Office of Planning and Performance Management, at (202) 208-1818. Sincerely, Signed by: Pamela K. Haze: Deputy Assistant Secretary for Budget and Business Management: Enclosure: General Comments: Page 31, The role of bureaus in setting their own targets: * In pursuing effective performance management, it is important at the Department of the Interior that programs have as much ownership as possible of their performance measure targets and reporting. To improve the relevance of performance assessment to the programs being measured, programmatic performance targets are set by the bureaus and not by the Department. Therefore, the particular reference that "...Interior set annual goals for this measure..." should be corrected to recognize that the water delivery annual goals being discussed (on page 31) are set by the Bureau of Reclamation. * It is also important to note that the water delivery measure emphasized in the GAO report is not highlighted but rather provided for reference purposes in the back of the Performance and Accountability Report's (PAR) Part II tables. In the PAR, more emphasis is actually placed on the Representative Performance Measure about Reclamation's facility reliability, for which program managers have more direct influence, in the up front Management Discussion and Analysis section. Concerns about equating increasingly specialized measures with improving credibility: * We would like to continue to emphasize outcome measures. We are concerned with the depiction of long-term measures as "lacking credibility" and an impediment to day-to-day performance measurement. Also, while we agree with the disadvantages of too many measures, the report appears to suggest that an increased set of more specialized, narrower focused measures, e.g. for different sizes of historic structures, could improve their credibility, and use, among managers. The need to better balance the application of program measures: * As part of determining the most effective cadre of measures, it is important that the Department and its bureaus have the final decision on which measures to use. The present existence of over 400 PART measures needs to be reviewed for their relativity to effective program management. Concerns about how measures are applied on the use of performance information: * We also suggest that along with considering the construction of the measures, it may be beneficial to focus on managers' concerns about how the performance information would be used. We suggest that managers be given an opportunity to provide context along with the reported performance. We have been including more perspective within our performance reports, especially in our PAR Management Discussion and Analysis and Part II tables to present explanations of trends in performance achievement over time, in the context of what has been achieved, the significance relative to future achievement, and consideration of future management adjustments. While we believe that we have demonstrated some improvement in this type of performance trend analysis in our most recent PAR and Citizen's Report, placing performance in a perspective beyond just whether or not a single year's targets were met or not, will continue to be a focus of our performance assessment efforts. * Along with the recommendations provided by GAO, we recognize that senior leadership and management could help better demonstrate to managers that reporting the context surrounding the level of performance achieved is as important as the value of the target that has been met, not met, or exceeded. [End of section] Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: Bernice Steinhardt at (202) 512-6806 or steinhardtb@gao.gov: Staff Acknowledgments: Elizabeth Curda and Laura Miller Craig managed this assignment. Jessica Nierenberg, Thomas M. Beall, Nicholas Benne, and Kate Hudson Walker made key contributions to all aspects of the report. David Bixler, Peter DelToro, Daniel Dunn, Elizabeth Erdmann, Ellen Grady, William O. Jenkins, Jr., Kathleen M. King, Barbara Lancaster, Walter Ochinko, Melanie Papasian, Mark Ramage, Jerry Sandau, William Trancucci, and John Vocino also provided assistance. In addition, A.J. Stephens provided legal support and Donna Miller developed the report's graphics. [End of section] Related GAO Products: Medicare and Medicaid Participating Facilities: CMS Needs to Reexamine Its Approach for Funding State Oversight of Health Care Facilities. GAO- 09-64. Washington, D.C.: February 13, 2009. High-Risk Series: An Update. GAO-09-271. Washington, D.C.: January 2009. Disaster Recovery: FEMA's Public Assistance Grant Program Experienced Challenges with Gulf Coast Rebuilding. GAO-09-129. Washington, D.C.: December 18, 2008. Government Performance: 2007 Federal Managers Survey on Performance and Management Issues, an E-supplement to GAO-08-1026T. GAO-08-1036SP. Washington, D.C.: July 24, 2008. Government Performance: Lessons Learned for the Next Administration on Using Performance Information to Improve Results. GAO-08-1026T. Washington, D.C.: July 24, 2008. Nursing Home Reform: Continued Attention Is Needed to Improve Quality of Care in Small but Significant Share of Homes. GAO-07-794T. Washington, D.C.: May 2, 2007. President's Management Agenda: Review of OMB's Improved Financial Performance Scorecard Process. GAO-07-95. Washington, D.C.: November 16, 2006. National Park Service: Major Operations Funding Trends and How Selected Park Units Responded to Those Trends for Fiscal Years 2001 through 2005. GAO-06-431. Washington, D.C.: March 31, 2006. Nursing Homes: Despite Increased Oversight, Challenges Remain in Ensuring High-Quality Care and Resident Safety. GAO-06-117. Washington, D.C.: December 28, 2005. Program Evaluation: OMB's PART Reviews Increased Agencies' Attention to Improving Evidence of Program Results. GAO-06-67. Washington, D.C.: October 28, 2005. Performance Budgeting: PART Focuses Attention on Program Performance, but More Can Be Done to Engage Congress. GAO-06-28. Washington, D.C.: October 28, 2005. Managing for Results: Enhancing Agency Use of Performance Information for Management Decision Making. GAO-05-927. Washington, D.C.: September 9, 2005. Human Capital: Senior Executive Performance Management Can Be Significantly Strengthened to Achieve Results. GAO-04-614. Washington, D.C.: May 26, 2004. Results-Oriented Government: GPRA Has Established a Solid Foundation for Achieving Greater Results. GAO-04-38. Washington, D.C.: March 10, 2004. Highlights of a GAO Forum: High-Performing Organizations: Metrics, Means, and Mechanisms for Achieving High Performance in the 21st Century Public Management Environment. GAO-04-343SP. Washington, D.C.: February 13, 2004. Performance Budgeting: Observations on the Use of OMB's Program Assessment Rating Tool for the Fiscal Year 2004 Budget. GAO-04-174. Washington, D.C.: January 30, 2004. Results-Oriented Government: Using GPRA to Address 21st Century Challenges. GAO-03-1166T. Washington, D.C.: September 18, 2003. Nursing Home Quality: Prevalence of Serious Problems, While Declining, Reinforces Importance of Enhanced Oversight. GAO-03-561. Washington, D.C.: July 15, 2003. Results-Oriented Cultures: Implementation Steps to Assist Mergers and Organizational Transformations. GAO-03-669. Washington, D.C.: July 2, 2003. Program Evaluation: An Evaluation Culture and Collaborative Partnerships Help Build Agency Capacity. GAO-03-454. Washington, D.C.: May 2, 2003. Results-Oriented Cultures: Creating a Clear Linkage between Individual Performance and Organizational Success. GAO-03-488. Washington, D.C.: March 14, 2003. Results-Oriented Cultures: Insights for U.S. Agencies from Other Countries' Performance Management Initiatives. GAO-02-862. Washington, D.C.: August 2, 2002. Results-Oriented Budget Practices in Federal Agencies. GAO-01-1084SP. Washington, D.C.: August 2001. Managing for Results: Federal Managers' Views on Key Management Issues Vary Widely Across Agencies. GAO-01-592. Washington, D.C.: May 25, 2001. Managing for Results: Emerging Benefits From Selected Agencies' Use of Performance Agreements. GAO-01-115. Washington, D.C.: October 30, 2000. Managing for Results: Federal Managers' Views Show Need for Ensuring Top Leadership Skills. GAO-01-127. Washington, D.C.: October 20, 2000. Managing for Results: Challenges Agencies Face in Producing Credible Performance Information. GAO/GGD-00-52. Washington, D.C.: February 4, 2000. Nursing Homes: Additional Steps Needed to Strengthen Enforcement of Federal Quality Standards. GAO/HEHS-99-46. Washington, D.C.: March 18, 1999. The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven. GAO/GGD-97-109. Washington, D.C.: June 2, 1997. Executive Guide: Effectively Implementing the Government Performance and Results Act. GAO/GGD-96-118. Washington, D.C.: June 1, 1996. Government Reform: Goal-Setting and Performance. GAO/AIMD/GGD-95-130R. Washington, D.C.: March 27, 1995. [End of section] Footnotes: [1] Pub. L. No. 103-62, 107 Stat. 285 (Aug. 3, 1993). Congress enacted GPRA to address several broad purposes, including improving federal program effectiveness, accountability, and service delivery; and enhancing congressional decision making by providing more objective information on program performance. See appendix III for more information on GPRA and other federal management reforms. [2] OMB created PART, a diagnostic tool that was intended to provide a consistent approach for evaluating federal programs as part of the executive budget formulation process. Through PART, OMB sought to create better ties between program performance and the allocation of resources. Although PART was discontinued as of the change in administration in 2009, it is likely that OMB will continue some form of agency assessment that will require performance information. See appendix III for more information on PART and other federal management reforms. [3] Our surveys were completed in 1997, 2000, 2003, and 2007 and were designed to obtain the observations and perceptions of respondents on various aspects of results-oriented management topics such as the presence and use of performance measures, hindrances to measuring performance and using performance information, and agency climate. Most of the items on our surveys asked respondents to rate the strength of their perception on a 5-point extent scale ranging from "to no extent" at the low end of the scale to "to a very great extent" at the high end. For more information on our survey methodology and selected survey results see appendix I. [4] GAO, Government Performance: Lessons Learned for the Next Administration on Using Performance Information to Achieve Results, GAO- 08-1026T (Washington, D.C.: July 24, 3008). In addition to our testimony, our survey results are also available: GAO, Government Performance: 2007 Federal Managers Survey on Performance and Management Issues, an E-supplement to [hyperlink, http://www.gao.gov/products/GAO-08-1026T], [hyperlink, http://www.gao.gov/products/GAO-08-1036SP] (Washington, D.C.: July 24, 2008). [5] [hyperlink, http://www.gao.gov/products/GAO-08-1026T]. [6] [hyperlink, http://www.gao.gov/products/GAO-08-1026T]. Hereafter, when describing our survey results, we are reporting the percentage of federal managers who selected the "great" or "very great extent" response to survey items. [7] GAO, Managing for Results: Enhancing Agency Use of Performance Information for Management Decision Making, [hyperlink, http://www.gao.gov/products/GAO-05-927] (Washington, D.C.: Sept. 9, 2005). [8] Disaster Relief and Recovery Supplemental Appropriations Act, 2008, Pub. L. No. 110-329, div. B, 122 Stat. 3574, 3592 (Sept. 30, 2008); Department of Homeland Security Appropriations Act, 2009, Pub. L. No. 110-329, div. D, 122 Stat. 3574, 3670-676 (Sept. 30, 2009); Pub. L. No. 111-5, 123 Stat. 115 (Feb.17, 2009). The Recovery Act provided stimulus funding for preserving and creating jobs and promoting economic recovery and for investment in transportation, environmental protection, and other infrastructure. The Congressional Budget Office (CBO) estimates that the Recovery Act's combined spending and tax provisions will cost $787 billion, of which over $580 billion will be in additional spending. [9] Energy and Water Development and Related Agencies Appropriations Act, 2009, Pub. L. No. 111-8, div. C, 123 Stat. 524, 609 (Mar. 11, 2009); Department of the Interior, Environment, and Related Agencies Appropriations Act, 2009, Pub. L. No. 111-8, div. E, 123 Stat. 701-725. Interior was funded by continuing resolutions for the 2009 fiscal year up until the Omnibus Appropriations Act was enacted. See Pub. L. No. 111-6, 123 Stat. 522 (Mar. 6, 2009); Consolidated Security, Disaster Assistance, and Continuing Appropriations Act, 2009, Pub. L. No. 110- 329, div. A, Continuing Appropriations Resolution, 2009, 122 Stat. 3574, 3575, 3593 (Sept. 30, 2008). [10] Pub. L. No. 111-5, 123 Stat. 115, 137, 166-168 (Feb. 17, 2009). [11] CMS contracts with states to assess the quality of care provided by Medicare and Medicaid participating facilities. [12] GAO, Nursing Home Reform: Continued Attention Is Needed to Improve Quality of Care in Small but Significant Share of Homes, [hyperlink, http://www.gao.gov/products/GAO-07-794T] (Washington, D.C.: May 2, 2007). For a full bibliography of our prior work on nursing homes, see GAO, Medicare and Medicaid Participating Facilities: CMS Needs to Reexamine Its Approach for Funding State Oversight of Health Care Facilities, [hyperlink, http://www.gao.gov/products/GAO-09-64] (Washington, D.C.: Feb. 13, 2009). [13] GAO, High-Risk Series: An Update, [hyperlink, http://www.gao.gov/products/GAO-09-271] (Washington, D.C.: January 2009). [14] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [15] A critical step in the Public Assistance program process is the completion of a project worksheet, which documents eligible work and estimated cost. [16] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [17] GAO, Results-Oriented Cultures: Creating a Clear Linkage between Individual Performance and Organizational Success, [hyperlink, http://www.gao.gov/products/GAO-03-488] (Washington, D.C.: Mar. 14, 2003). [18] FEMA's SES managers were covered by DHS' human-capital-management system, which required linkages between agency goals and individual performance objectives. [19] Pub. L. No. 110-329, div. D, § 522. [20] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [21] In discussing hindrances to using performance information, managers we interviewed at Interior generally referred to performance information developed to meet GPRA and PART reporting requirements. Both GPRA and PART were intended to enhance decision making by requiring agencies to develop results-oriented performance goals linked to agency missions and report on the results achieved. See appendix III for more information on GPRA and PART. [22] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [23] Funds are obligated when a definite commitment is made that creates a legal liability of the government for the payment of goods and services ordered or received. An agency may incur an obligation, for example, when it places an order, signs a contract, awards a grant, or purchases a service. [24] More recently, according to another senior headquarters official, Interior had discontinued the requirement to include GRPA goals in SES individual performance plans. However, NPS guidance concerning its SES members' fiscal year 2008 performance plans indicated that GPRA performance information would be taken into consideration in individual performance evaluations. [25] GAO, Government Reform: Goal-Setting and Performance, [hyperlink, http://www.gao.gov/products/AIMD/GGD-95-130R] (Washington, D.C.: Mar. 27, 1995). [26] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [27] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [28] [hyperlink, http://www.gao.gov/products/GAO-05-927]. [29] GAO, Executive Guide: Effectively Implementing the Government Performance and Results Act, [hyperlink, http://www.gao.gov/products/GGD-96-118] (Washington, D.C.: June 1, 1996). [30] Current OMB guidance calls for agencies to combine the annual performance report required by GPRA with their financial statement and accountability report into a Performance and Accountability Report (PAR). [31] A headquarters official indicated that in 2007, changes were made to PMDS to improve the data-entry process. However, none of the managers we interviewed in 2008 through early 2009 commented on these changes. [32] Pub. L. No. 108-173, 117 Stat. 2066 (Dec. 8, 2003). [33] CMS' SES are covered separately by the agency's automated program, the Performance Plan System (PPS). The PPS supports the guidelines and requirements as outlined in the Department of Health and Human Services' Senior Executive and Organizational Performance Management System. [34] GAO, Nursing Homes: Additional Steps Needed to Strengthen Enforcement of Federal Quality Standards, [hyperlink, http://www.gao.gov/products/GAO/HEHS-99-46] (Washington, D.C.: Mar. 18, 1999). [35] GAO, Nursing Homes: Despite Increased Oversight, Challenges Remain in Ensuring High-Quality Care and Resident Safety, [hyperlink, http://www.gao.gov/products/GAO-06-117] (Washington, D.C.: Dec. 28, 2005). [36] GAO, Nursing Home Quality: Prevalence of Serious Problems, While Declining, Reinforces Importance of Enhanced Oversight, [hyperlink, http://www.gao.gov/products/GAO-03-561] (Washington, D.C.: July 15, 2003). [37] Our surveys were conducted in 1997, 2000, 2003, and 2007. For information on the design and administration of each of the four surveys, see GAO, The Government Performance and Results Act: 1997 Governmentwide Implementation Will Be Uneven, [hyperlink, http://www.gao.gov/products/GAO/GGD-97-109] (June 2, 1997); Managing for Results: Federal Managers' Views on Key Management Issues Vary Widely Across Agencies, [hyperlink, http://www.gao.gov/products/GAO-01- 592] (May 25, 2001); Results-Oriented Government: GPRA Has Established a Solid Foundation for Achieving Greater Results, [hyperlink, http://www.gao.gov/products/GAO-04-38] (Mar. 10, 2004); and most recently, Government Performance: Lessons Learned for the Next Administration on Using Performance Information to Achieve Results, [hyperlink, http://www.gao.gov/products/GAO-08-1026T] (Washington, D.C.: July 24, 3008) [38] GAO, Managing for Results: Enhancing Agency Use of Performance Information for Management Decision Making, [hyperlink, http://www.gao.gov/products/GAO-05-927] (Washington, D.C.: Sept. 9, 2005). [39] Although the Forest Service had the lowest ranking among all federal agencies, our recent work at this agency had already resulted in recommendations to address key management issues that we will continue to monitor. [40] See GAO, Government Performance: 2007 Federal Managers Survey on Performance and Management Issues, an E-supplement to [hyperlink, http://www.gao.gov/products/GAO-08-1026T], [hyperlink, http://www.gao.gov/products/GAO-08-1036SP] (Washington, D.C.: July 24, 2008), for the wording of the nine items used in computing the average agency change. These items were 8b, 8c, 8f, 8g, 8h, 8i, 8j, 8k, and 8l. We could not use the core-uses index items for this change analysis since it incorporated new items that had been added since the 2000 survey. [41] At the time of the 2000 survey, CMS was known as the Health Care Financing Administration. [42] For all five items, fewer than a fifth of CMS managers reported having any type of measure to a great or very great extent. [43] See GAO, Managing for Results: Enhancing Agency Use of Performance Information for Management Decision Making, [hyperlink, http://www.gao.gov/products/GAO-05-927] (Sept. 9, 2005). See the online e-supplement, GAO, Government Performance: 2007 Federal Managers Survey on Performance and Management Issues, an E-supplement to [hyperlink, http://www.gao.gov/products/GAO-08-1026T], [hyperlink, http://www.gao.gov/products/GAO-08-1036SP] (Washington, D.C.: July 24, 2008) for the wording of the items. The nine items constituting the index are questions 8a, 8c, 8d, 8e, 8k, 8m, 10d, 10m, and 11b. [44] For example, index score values between 1 and 2.99 were viewed as covering the two categories of "small" or "to no extent" while values of 3 to 3.99 fit the category "moderate extent" and values between 4 and 5 encompassed the categories of "great" or "very great" extent. Agency averages ranged from 2.94 to 3.64. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO‘s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO‘s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.