Public Health and Hospital Emergency Preparedness Programs

Evolution of Performance Measurement Systems to Measure Progress Gao ID: GAO-07-485R March 23, 2007

The September 11, 2001, terrorist attacks, the anthrax incidents during the fall of 2001, Hurricane Katrina, and concerns about the possibility of an influenza pandemic have raised public awareness and concerns about the nation's public health and medical systems' ability to respond to bioterrorist events and other public health emergencies. From 2002 to 2006, the Congress appropriated about $6.1 billion to the Department of Health and Human Services (HHS) to support activities to strengthen state and local governments' emergency preparedness capabilities under the Public Health Security and Bioterrorism Preparedness and Response Act of 2002 (Preparedness and Response Act). HHS has distributed funds annually to 62 recipients, including all 50 states and 4 large municipalities, through cooperative agreements under two programs--the Centers for Disease Control and Prevention's (CDC) Public Health Emergency Preparedness Program, and the Health Resources and Services Administration's (HRSA) National Bioterrorism Hospital Preparedness Program. The common goal of CDC's and HRSA's preparedness programs is to improve state and local preparedness to respond to bioterrorism and other large-scale public health emergencies, such as natural disasters or outbreaks of infectious disease. Annually, both CDC and HRSA develop and issue program guidance for recipients that describes activities necessary to improve their ability to respond to bioterrorism and other public health emergencies and sets out requirements for measuring their performance. Each recipient is required to submit periodic reports that track progress in improving their preparedness. As a result of the nation's ineffective response to Hurricane Katrina and the need to prepare for a possible influenza pandemic, members of the Congress have raised questions about CDC's and HRSA's efforts to monitor the progress of their preparedness programs. Because of these questions, we are reporting on (1) how CDC's and HRSA's performance measurement systems have evolved and (2) how CDC and HRSA are using these systems to measure the progress of their preparedness programs.

Since 2002, CDC's and HRSA's performance measurements have evolved from measuring capacity to assessing capability. Early in their programs, both agencies used markers or values that they called benchmarks to measure capacity-building efforts, such as purchasing equipment and supplies and acquiring personnel. These benchmarks were developed from activities authorized in the Preparedness and Response Act. In 2002, CDC established 14 benchmarks, such as requiring each recipient to designate an executive director of the bioterrorism and response program, establish a bioterrorism advisory committee, and develop a statewide response plan. From 2003 to 2005, CDC further developed its performance measurements by obtaining input from stakeholders to make a transition from using benchmarks focused on capacities to using performance measures focused on capabilities, such as whether personnel have been trained and can appropriately use equipment. In 2006, CDC continued to work with stakeholders to refine its performance measures. At the beginning of its program in 2002, HRSA established 5 benchmarks, such as requiring each recipient to designate a coordinator for bioterrorism planning, establish a hospital preparedness committee, and develop a plan for hospitals to respond to a potential epidemic. From 2003 to 2005, HRSA modified existing benchmarks and added new ones, such as training benchmarks, based on the existing legislation and input from stakeholders. In 2006, HRSA convened an expert panel to propose a set of performance measures focused on capabilities. CDC and HRSA officials told us they will continue to face challenges as their performance measures evolve, such as gaining consensus among stakeholders in light of minimal scientific data about public health and hospital emergency preparedness. CDC and HRSA use data from recipients' reports and site visits to monitor recipients' progress in improving their ability to respond to bioterrorism events and other public health emergencies. CDC and HRSA project officers use performance measurement data from recipients' required progress reports, along with site visits, to monitor progress and provide feedback about whether individual recipients have accomplished activities related to their ability to respond to bioterrorism events and other public health emergencies. Currently, there are no standard analyses or reports that enable CDC and HRSA to compare data across recipients to measure collective progress, compare progress across recipients' programs, or provide consistent feedback to recipients. However, in mid to late 2006 both agencies began developing formal data analysis programs that are intended to validate recipient-reported data and assist in generating standardized reports. According to CDC officials, CDC plans to finish validation projects by August 2007 and then develop routine reports summarizing individual recipient and national progress. In addition, CDC plans to issue a report by the end of 2007 providing a "snapshot" of the progress recipients have made in building emergency readiness capacity and addressing how CDC will measure capability in the future. However, because of the expected move of HRSA's program to a different HHS office in 2007, its schedule for finishing data validation was tentative at the time we briefed your staff. Furthermore, due to the expected move, HRSA officials said at that time that decisions about whether to issue a report in 2007 on recipients' progress also had not been made.



GAO-07-485R, Public Health and Hospital Emergency Preparedness Programs: Evolution of Performance Measurement Systems to Measure Progress This is the accessible text file for GAO report number GAO-07-485R entitled 'Public Health and Hospital Emergency Preparedness Programs: Evolution of Performance Measurement Systems to Measure Progress' which was released on April 25, 2007. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. March 23, 2007: The Honorable Bennie G. Thompson: Chairman: Committee on Homeland Security: House of Representatives: The Honorable Judd Gregg: Ranking Minority Member: Committee on the Budget: United States Senate: The Honorable Charles E. Grassley: Ranking Minority Member: Committee on Finance: United States Senate: The Honorable Henry A. Waxman: Chairman: Committee on Oversight and Government Reform: House of Representatives: The Honorable Edward J. Markey: House of Representatives: Subject: Public Health and Hospital Emergency Preparedness Programs: Evolution of Performance Measurement Systems to Measure Progress: The September 11, 2001, terrorist attacks, the anthrax incidents during the fall of 2001, Hurricane Katrina, and concerns about the possibility of an influenza pandemic have raised public awareness and concerns about the nation's public health and medical systems' ability to respond to bioterrorist events and other public health emergencies. From 2002 to 2006, the Congress appropriated about $6.1 billion to the Department of Health and Human Services (HHS) to support activities to strengthen state and local governments' emergency preparedness capabilities under the Public Health Security and Bioterrorism Preparedness and Response Act of 2002 (Preparedness and Response Act).[Footnote 1] HHS has distributed funds annually to 62 recipients, including all 50 states and 4 large municipalities, through cooperative agreements under two programs--the Centers for Disease Control and Prevention's (CDC) Public Health Emergency Preparedness Program,[Footnote 2] and the Health Resources and Services Administration's (HRSA) National Bioterrorism Hospital Preparedness Program. The common goal of CDC's and HRSA's preparedness programs is to improve state and local preparedness to respond to bioterrorism and other large-scale public health emergencies, such as natural disasters or outbreaks of infectious disease. To guide efforts by federal, state, and local departments and agencies to prepare and respond to terrorism and other major emergencies, the federal government has developed a number of national strategies, including a National Strategy for Homeland Security, which was issued in July 2002.[Footnote 3] Among other things, the National Strategy for Homeland Security requires federal government departments and agencies to create performance measures to evaluate progress in achieving homeland security initiatives, including national preparedness and emergency response, and to allocate future resources. Annually, both CDC and HRSA develop and issue program guidance for recipients that describes activities necessary to improve their ability to respond to bioterrorism and other public health emergencies and sets out requirements for measuring their performance. Each recipient is required to submit periodic reports that track progress in improving their preparedness. As a result of the nation's ineffective response to Hurricane Katrina and the need to prepare for a possible influenza pandemic, members of the Congress have raised questions about CDC's and HRSA's efforts to monitor the progress of their preparedness programs. Because of these questions, we are reporting on (1) how CDC's and HRSA's performance measurement systems have evolved and (2) how CDC and HRSA are using these systems to measure the progress of their preparedness programs. Enclosure I contains the information we provided to your staff at our February 28, 2007, briefing. To do our work, we reviewed and analyzed federal government documents related to national security and emergency preparedness. We also obtained reports and interviewed officials from federal agencies[Footnote 4] that had evaluated CDC's and HRSA's public health and hospital preparedness programs, professional associations involved in emergency preparedness, and policy research organizations that had published assessments or evaluations of public health and hospital preparedness programs. We analyzed CDC and HRSA documents and interviewed officials to determine how they have developed and implemented performance management systems for their cooperative agreement programs, including recipient reporting requirements, and systems for collecting data from recipients. Additionally, we analyzed other CDC and HRSA documents to identify procedures in place for management review of program progress and for providing feedback and suggestions for program improvements to recipients. We did not evaluate the actual performance measures adopted by CDC or HRSA or examine the accuracy or completeness of recipients' self-reported data as contained in the progress reports they are required to submit to CDC or HRSA. See enclosure II for detailed information on our scope and methodology. We conducted our work from June 2006 through March 2007 in accordance with generally accepted government auditing standards. Results in Brief: Since 2002, CDC's and HRSA's performance measurements have evolved from measuring capacity to assessing capability. Early in their programs, both agencies used markers or values that they called benchmarks to measure capacity-building efforts, such as purchasing equipment and supplies and acquiring personnel.[Footnote 5] These benchmarks were developed from activities authorized in the Preparedness and Response Act. In 2002, CDC established 14 benchmarks, such as requiring each recipient to designate an executive director of the bioterrorism and response program, establish a bioterrorism advisory committee, and develop a statewide response plan. From 2003 to 2005, CDC further developed its performance measurements by obtaining input from stakeholders to make a transition from using benchmarks focused on capacities to using performance measures focused on capabilities, such as whether personnel have been trained and can appropriately use equipment. In 2006, CDC continued to work with stakeholders to refine its performance measures. At the beginning of its program in 2002, HRSA established 5 benchmarks, such as requiring each recipient to designate a coordinator for bioterrorism planning, establish a hospital preparedness committee, and develop a plan for hospitals to respond to a potential epidemic. From 2003 to 2005, HRSA modified existing benchmarks and added new ones, such as training benchmarks, based on the existing legislation and input from stakeholders. In 2006, HRSA convened an expert panel to propose a set of performance measures focused on capabilities. CDC and HRSA officials told us they will continue to face challenges as their performance measures evolve, such as gaining consensus among stakeholders in light of minimal scientific data about public health and hospital emergency preparedness. CDC and HRSA use data from recipients' reports and site visits to monitor recipients' progress in improving their ability to respond to bioterrorism events and other public health emergencies. CDC and HRSA project officers use performance measurement data from recipients' required progress reports, along with site visits, to monitor progress and provide feedback about whether individual recipients have accomplished activities related to their ability to respond to bioterrorism events and other public health emergencies. Currently, there are no standard analyses or reports that enable CDC and HRSA to compare data across recipients to measure collective progress, compare progress across recipients' programs, or provide consistent feedback to recipients. However, in mid to late 2006 both agencies began developing formal data analysis programs that are intended to validate recipient- reported data and assist in generating standardized reports. According to CDC officials, CDC plans to finish validation projects by August 2007 and then develop routine reports summarizing individual recipient and national progress. In addition, CDC plans to issue a report by the end of 2007 providing a "snapshot" of the progress recipients have made in building emergency readiness capacity and addressing how CDC will measure capability in the future. However, because of the expected move of HRSA's program to a different HHS office in 2007, its schedule for finishing data validation was tentative at the time we briefed your staff. Furthermore, due to the expected move, HRSA officials said at that time that decisions about whether to issue a report in 2007 on recipients' progress also had not been made. Agency Comments: We requested comments on a draft of this report from HHS. The department provided written comments that are reprinted in enclosure III. In commenting on this draft, HHS provided additional information about the transfer on March 5, 2007, of the National Bioterrorism Hospital Preparedness Program from HRSA to the new HHS Office of the Assistant Secretary for Preparedness and Response. According to HHS, it has made a number of changes that it believes will improve its ability to monitor performance at the individual recipient level and for the program overall. HHS is also planning to conduct an analysis of the performance data for existing recipients for fiscal years 2002-2006 in order to develop a more complete picture of levels of preparedness from all National Bioterrorism Hospital Preparedness Program recipients. Many of the initiatives outlined in HHS' comments were begun after our briefings to your staff on February 28, 2007, and are still being implemented; we are unable to comment on their effectiveness. As we continue to evaluate emergency preparedness programs we will review the results of their continued efforts to develop measurable evidence-based benchmarks and objective standards and their ability to compare data across recipients to measure collective progress, compare progress across recipients' programs, or provide consistent feedback to recipients. As arranged with your offices, unless you release its content earlier, we plan no further distribution of this report until 30 days after its issuance date. At that time, we will send copies of this report to the Secretary of HHS and other interested parties. We will also make copies available to others on request. In addition, the report will be available at no charge on the GAO Web site at http://www.gao.gov. Contact points for our Office of Congressional Relations and Public Affairs may be found on the last page of this report. If you and your staff have any questions or need additional information, please contact me at (202) 512-7101, or bascettac@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff members who made major contributions to this report are listed in enclosure IV. Signed by: Cynthia A. Bascetta: Director, Health Care: Enclosures - 4: Enclosure I: Information Presented in Briefing on February 28, 2007: The information in this enclosure is taken directly from the slides used in the briefing presented to the staffs of the Honorable Judd Gregg, Ranking Minority Member, Senate Committee on the Budget; the Honorable Charles E. Grassley, Ranking Minority Member, Senate Committee on Finance; the Honorable Bennie G. Thompson, Chairman, House Committee on Homeland Security; the Honorable Henry A. Waxman, Chairman, House Committee on Oversight and Government Reform; and the Honorable Edward J. Markey, House of Representatives on February 28, 2007. Introduction (slides 3 through 6): The September 11, 2001, terrorist attacks, the anthrax incidents, Hurricane Katrina, and concerns about the possibility of an influenza pandemic have raised public awareness and concerns about the nation's public health and medical systems' ability to respond to bioterrorist events and other public health emergencies. In November 2002, the Congress passed legislation creating the Department of Homeland Security (DHS), giving it the overall responsibility for managing emergency preparedness. The Department of Health and Human Services (HHS) is designated as the primary agency for implementing activities relating to public health and hospital emergency preparedness. From 2002 to 2006, the Congress appropriated about $6.1 billion to support activities under the Public Health Security and Bioterrorism Preparedness and Response Act of 2002 (Preparedness and Response Act) to strengthen state and local governments' emergency readiness capabilities. HHS has distributed these funds annually to 62 recipients, including all 50 states and 4 large municipalities, through cooperative agreements under two programs: * Centers for Disease Control and Prevention's (CDC) Public Health Emergency Preparedness Program (formerly the Public Health Preparedness and Response for Bioterrorism Program), and: * Health Resources and Services Administration's (HRSA) National Bioterrorism Hospital Preparedness Program. In addition to bioterrorism, these programs also address other large- scale public health emergencies, such as natural disasters or outbreaks of infectious disease. This "all-hazards" approach recognizes that some aspects of response to bioterrorism, such as providing emergency medical services and managing mass casualties, can be the same as for response to other public health emergencies. Public Law 109-417, the Pandemic and All-Hazards Preparedness Act, enacted December 19, 2006, amended the Preparedness and Response Act and authorizes appropriations for CDC's and HRSA's public health and hospital preparedness programs through 2011. The legislation also creates a new Assistant Secretary for Preparedness and Response in HHS and transfers responsibility for HRSA's hospital preparedness program to this position. The program is expected to move some time in 2007. To guide preparedness and response for terrorism and other major emergencies, the federal government developed a number of national strategies, including a National Strategy for Homeland Security issued in July 2002.[Footnote 6] This national strategy requires federal government departments and agencies to create performance measures to evaluate progress in achieving homeland security initiatives, including national preparedness and emergency response, and to allocate future resources. Purpose and Questions (slide 7): As a result of the nation's ineffective response to Hurricane Katrina and the need to prepare for a possible influenza pandemic, members of the Congress have raised questions about CDC's and HRSA's efforts to monitor the progress of their preparedness programs. To assess CDC's and HRSA's systems to monitor these programs, we reviewed the following questions: 1. How have CDC's and HRSA's performance measurement systems evolved? 2. How are CDC and HRSA using these systems to measure the progress of their preparedness programs? Scope and Methodology (slides 8 through 10): To do our work, we interviewed officials from: * HHS's Office of Public Health Emergency Preparedness (OPHEP), Office of the Assistant Secretary for Planning and Evaluation, Office of the Inspector General (OIG), and Agency for Healthcare Research and Quality (AHRQ); * CDC's Coordinating Office for Terrorism Preparedness and Emergency Response; * HRSA's National Bioterrorism Hospital Preparedness Program; * Congressional Research Service; and: * professional associations involved in emergency preparedness and policy research organizations that had published assessments or evaluations of public health and hospital preparedness programs. We also reviewed and analyzed documents from: * The Executive Office of the President, including the National Strategy for Homeland Security and Homeland Security Presidential Directives; * DHS, including the National Response Plan, the Interim National Preparedness Goal, and the draft Target Capabilities List; * HHS's OIG and AHRQ; * Congressional Research Service; * Office of Management and Budget, including Program Assessment Rating Tool reviews; * CDC and HRSA on the development of performance management systems and recipients' annual applications and progress reports; and: * professional associations and policy research organizations. We did not evaluate the actual performance measures adopted by CDC or HRSA or examine the accuracy or completeness of recipients' self- reported data as contained in the progress reports they are required to submit to CDC or HRSA. Our review was conducted from June 2006 through March 2007 in accordance with generally accepted government auditing standards. Background (slides 11 through 17): CDC's and HRSA's Preparedness Programs: The common goal of CDC's and HRSA's preparedness programs is to improve state and local preparedness to respond to bioterrorism and other public health emergencies. * CDC's program focuses on public health preparedness. * HRSA's program focuses on hospital preparedness. CDC and HRSA annually distribute program funds to recipients. These funds are used to improve their ability to respond to bioterrorism and other public health emergencies, such as training volunteers to provide mass vaccinations or antibiotics in the event of a public health emergency. CDC and HRSA also develop program guidance for recipients that describes activities necessary to improve preparedness and sets out requirements for measuring recipients' performance. CDC's Preparedness Program: CDC distributes funds under its cooperative agreements on an annual basis. Each recipient: * must apply annually for these funds; * receives a base amount, plus an amount based on its proportional share of the national population; and: * has flexibility in how to distribute the funds to local public health agencies based on the workplan submitted to CDC with the recipient's application. Each recipient must submit reports that track progress in improving its ability to respond to bioterrorism and other public health emergencies. These have included quarterly, midyear, and annual reports. HRSA's Preparedness Program: HRSA distributes funds under its cooperative agreements on an annual basis. Each recipient: * receives a base amount, plus an amount based on its proportional share of the national population; and: * must allocate at least 75 percent of its funds to hospitals or other health care entities. - Recipients distribute most of the funds to hospitals, with a small portion going to other entities such as community health centers, emergency medical services, and poison control centers. - Recipients may use the remaining funds to support their administrative costs and needs assessments. Each recipient must submit midyear and annual reports that track progress in improving its ability to respond to bioterrorism and other public health emergencies. Prior Reviews of CDC's and HRSA's Preparedness Programs: Several government and private studies, including those conducted by GAO, HHS's OIG, and Rand, have noted weaknesses in CDC's and HRSA's preparedness programs. * In February 2004, we reported (GAO-04-360R) that although the states' progress fell short of 2002 goals and much remained to be accomplished, these programs enabled states to make needed improvements in public health and health care capabilities critical for preparedness. * Since December 2002, HHS's OIG has issued seven evaluation and inspection reports on program results. It found that all of the studied recipients had prepared bioterrorism responses and were working to strengthen their infrastructure, but barriers to preparedness remained, including problems with staffing, funding, and communication and the need for standards and guidance. * Since 2001, Rand has conducted many studies related to preparedness for public health emergencies. Rand studied how public health preparedness is transforming public health agencies and found: - the preparedness mission has raised challenges in terms of accountability among local health jurisdictions; - it is difficult to assess preparedness because measures to define and assess preparedness, and a strong evidence base to support those measures are lacking; and: - it is difficult to measure preparedness because it involves measuring the capacity to deal with situations that rarely happen. Under a contract with HHS, Rand currently is convening expert panels and performing literature searches to help define preparedness. Presidential Directive 8--National Preparedness: Homeland Security Presidential Directive 8 provides some guidance on implementing the National Strategy for Homeland Security. Consistent with the directive, DHS developed the Interim National Preparedness Goal and the draft Target Capabilities List and issued them in 2005.[Footnote 7] * The Interim National Preparedness Goal establishes preparedness priorities, targets, and standards for preparedness assessments and strategies to align efforts of federal, state, local, tribal, private- sector, and nongovernmental entities. * The draft Target Capabilities List identifies 37 capabilities that federal, state, local, tribal, private-sector, and nongovernmental entities need in order to prevent, protect against, respond to, and recover from a major event to minimize the impact on lives, property, and the economy. CDC's and HRSA's preparedness programs provide both funds and guidance to state and local entities and hospitals to help them develop these capabilities and meet these preparedness priorities. Performance Measurement Systems: Early in a program, performance measurement systems can focus on measuring capacity, such as equipment and supplies purchased and personnel hired. As programs mature and more data and scientific evidence are available, performance measurement systems can focus more on measuring capabilities, such as whether personnel are trained and can appropriately use equipment and supplies. Measurements can include: * type or level of program activities conducted (process), * direct products and services delivered (outputs), or: * results of those products and services (outcomes). Finding 1: CDC and HRSA Performance Measures Evolved from Measuring Capacity to Assessing Capability (slides 18 through 29): In 2002, CDC's and HRSA's efforts focused on measuring capacity, such as the type of staff hired and equipment needed to respond to a bioterrorism attack. To do this, CDC and HRSA identified markers or values against which recipients were expected to measure their performance. These initial markers or values, which they called benchmarks, were developed from emergency preparedness activities authorized in the Preparedness and Response Act. From 2003 to 2006, CDC and HRSA changed their approach from using benchmarks that measure capacity to using performance measures that focus on whether a program has met standards assessing capabilities. In 2004, CDC and HRSA increased their coordination and in 2005 began to coordinate with DHS to align their preparedness programs with the Interim National Preparedness Goal and draft Target Capabilities List. 2002--CDC's Initial Measurements Based on Legislation: In 2002, CDC initially established its performance measurement systems using benchmarks based on emergency preparedness activities authorized in the Preparedness and Response Act. CDC officials said these initial benchmarks measured program capacity- building efforts such as purchasing equipment and supplies and acquiring personnel.[Footnote 8] CDC established 14 critical benchmarks, such as requiring each recipient to designate an executive director of the bioterrorism and response program, establish a bioterrorism advisory committee, and develop a statewide response plan. 2003 to 2005--CDC's Transition from Measuring Capacity to Assessing Capability: From 2003 to 2005, CDC began to include the participation and input of stakeholders--other federal agencies, recipients of program funds, public health professional association officials, and industry experts- -as it further developed its performance measurements. This input resulted in modifications of the benchmarks and the transition from benchmarks to performance measures that address capabilities. * In 2003, an initial draft of over 100 proposed measures was developed from input by CDC internal subject matter experts. An external workgroup, including professional association representatives, reviewed and assessed the proposed measures. Some of the measures focused on new areas, such as exercising, drilling, and training. * In 2004, CDC convened a second CDC internal expert panel to conduct a literature search to identify evidence-based criteria to support the performance measures. The panel consolidated the over 100 performance measures into 47 interim performance measures. Subsequent field-testing eliminated one proposed measure. * In late 2004, CDC held teleconferences with selected recipients and professional association representatives to discuss these interim performance measures. This process reduced the number of performance measures to 34. * In 2005, CDC introduced the 34 performance measures in the 2005 cooperative agreement guidance and field tested the new measures in five locations. Example of the transition of a CDC benchmark into a performance measure that addresses capabilities: * 2002 benchmark: Recipients were required to develop a system to receive and evaluate urgent disease reports on a 24-hour-per-day, 7- day-per-week basis. * 2003/2004 benchmark: Recipients were required to complete development of and maintain a system to receive and evaluate urgent disease reports. * 2005 performance measure: Recipients were required to meet a target time of 15 minutes for a knowledgeable public health professional to respond to a call or a communication that appears to be of urgent public health consequence. 2005 to 2006--CDC's Refinement of Capability Assessment: In late 2005, CDC met with representatives from professional organizations and state and local public health laboratories and health departments to review and refine the performance measures. In 2006, CDC held further meetings with seven recipients and other stakeholders to discuss data collection efforts for performance measures and found that gathering some of the data would not be feasible. As a result, CDC further reduced the number of performance measures from 34 to 23. CDC's 2006 guidance with the 23 performance measures was issued in June 2006. Recipients were expected to comply with this guidance when implementing their 2006 programs, during the period from August 31, 2006, to August 30, 2007. 2002--HRSA's Initial Measurements Based on Legislation: In 2002, HRSA initially established its performance measurement systems using benchmarks based on emergency preparedness activities authorized in the Preparedness and Response Act. HRSA officials said these initial benchmarks measured program capacity- building efforts such as purchasing equipment and supplies and acquiring personnel. HRSA established five critical benchmarks, such as requiring each recipient to designate a coordinator for bioterrorism planning, establish a hospital preparedness committee, and develop a plan for hospitals to respond to a potential epidemic. 2003 to 2005--HRSA's Benchmarks Modified and Expanded: From 2003 to 2005, HRSA, like CDC, began to include the participation and input of stakeholders--federal agencies, cooperative agreement recipients, public health professional association officials, and industry experts--as it further developed its performance measurements. This input resulted in modifications of the benchmarks. * In 2003, HRSA added new benchmarks based on the existing legislation and meetings and discussions with stakeholders. The benchmarks focused on such things as exercising, drilling, and training. * In 2004, each of HRSA's benchmarks was divided into HRSA-identified "sentinel indicators," which are smaller component tasks that are intended to accomplish the larger benchmark activity. For example, for the benchmark "Surge Capacity: Beds," one of the sentinel indicators is the number of additional hospital beds for which a recipient could make patient care available within 24 hours. * In 2005, HRSA increased the number of sentinel indicators from 21 to 72 at HHS's request. For example, HHS asked for additional measures to identify bed capacity for trauma and burn victims. 2006--HRSA's Transition from Measuring Capacity to Assessing Capability: In early 2006, HRSA convened an expert panel that proposed a set of performance measures, which were then disseminated to stakeholders such as recipients, professional associations, industry experts, and federal agencies for feedback. This input resulted in adoption of 6 performance measures and 17 program measures (HRSA defined program measures as a mixture of program activities and process and outcome measures) that focus on capabilities. HRSA also maintained reporting requirements for 17 of its 72 sentinel indicators. HRSA's 2006 performance and program measures and sentinel indicators were not issued with its guidance in July 2006 because HRSA officials were awaiting final approval by HHS. These measures were issued in December 2006. However, according to HRSA officials, recipients were aware of the expectations contained in the guidance because they helped develop them. As such, it was HRSA's expectation that recipients would comply with them when implementing their 2006 programs, during the period from September 1, 2006, to August 31, 2007. Increased Coordination between CDC and HRSA; Coordination Initiated with DHS: In 2004, CDC and HRSA increased their coordination and in 2005 began to coordinate with DHS to align their preparedness programs with the Interim National Preparedness Goal and draft Target Capabilities List. For example, * CDC and HRSA project officers shared information in monthly conference calls. * CDC subject matter experts assisted HRSA's recipients. * CDC, HRSA, and DHS created a Joint Advisory Committee in 2005 to create common terminology for their respective programs and improve commonality in their guidance. * CDC and HRSA officials stated that in 2005 they had more closely aligned their performance measurements with the draft Target Capabilities List and the Interim National Preparedness Goal. Figure 1 provides an example of how CDC and HRSA have aligned their performance measurements with DHS's draft Target Capabilities List and the Interim National Preparedness Goal. Figure 1: Alignment of CDC and HRSA Performance Measures with DHS's Draft Target Capabilities List and the Interim National Preparedness Goal: [See PDF for Image] Source: CDC, HRSA, and DHS documents. [End of figure] CDC's and HRSA's Challenges: According to CDC and HRSA officials, they will continue to face challenges as their performance measures evolve, because gaining consensus among the various stakeholders--federal agencies, state and local governments, and professional associations--is difficult. These difficulties arise because: * minimal scientific data exist in this new area of public health and hospital emergency preparedness to guide performance measurement systems; and: * scientists, subject matter experts, and program officials can disagree as to what could and should be measured. Finding 2: CDC and HRSA Use Data from Recipients' Reports and Site Visits to Measure Progress (slides 30 through 36): CDC and HRSA project officers use performance measurement data from recipients' required reports, along with site visits, to monitor progress and provide feedback about whether individual recipients meet goals and accomplish activities related to their ability to respond to bioterrorism events and other public health emergencies. Both CDC and HRSA are making improvements to address the need for formal data analysis programs based on validated data and standardized procedures. Report and Site Visit Data: CDC and HRSA project officers are responsible for monitoring individual recipients' progress, providing technical assistance, and giving feedback on their emergency preparedness activities. Experts in areas such as epidemiology, laboratory testing, and surveillance assist project officers in providing technical assistance. * Project officers analyze and monitor individual recipients' progress from the information gathered through recipients' progress reports, phone calls, and e-mails and by conducting site visits. * Project officers use the information and their analyses of it to (1) provide recipients with technical assistance and feedback on their ability to respond to bioterrorism and other public health emergencies, (2) determine issues to discuss during future site visits, and (3) assist recipients in developing future cooperative agreement applications. * Project officers also collaborate with recipients to identify their specific needs for improving their emergency preparedness. For example, prior to site visits CDC project officers ask recipients what type of technical assistance they need and then include appropriate subject matter experts on the site visit. Providing Feedback: Both CDC and HRSA have various methods for providing feedback on progress to recipients: * Project officers determine the type and amount of feedback to provide each recipient on their progress. * CDC and HRSA periodically provide recipients with information about promising practices and lessons learned on improving their ability to respond to bioterrorism and other public health emergencies. * CDC and HRSA both hold annual conferences with all recipients to provide training, and other information such as changes to program guidance. Standard Analysis and Reports Currently Lacking: CDC and HRSA officials told us that project officers lack standard protocols, checklists, or procedures for analyzing recipients' reports that include both qualitative and quantitative data. Consequently, each project officer develops his or her own methods or procedures for analyzing and measuring recipients' progress. CDC and HRSA project officers have not generated standardized reports summarizing individual or collective recipients' progress and activities. Ongoing Improvements: However, both CDC and HRSA are making improvements in measuring progress: * In mid to late 2006, both CDC and HRSA began developing formal data analysis programs. They plan to generate standardized reports for management and other stakeholders as needed. * CDC and HRSA plan to put procedures in place to validate the accuracy, reasonableness, and completeness of selected data that recipients self-report. * Officials said validation is needed to: - ensure that reports based on recipients' data provide accurate information; - determine whether all recipients are comparably reporting the status of their preparedness; and: - allow managers to make informed decisions to improve the individual recipients' cooperative agreements and, ultimately, the nation's preparedness. Once the data validation projects are completed, CDC officials plan to develop routine reports with specific recipient information and reports that provide national summaries. CDC officials plan to finish the validation projects by August 2007. CDC officials said that in the interim they would continue to use many of the measurements from 2005 and 2006 to trace recipients' progress. HRSA's time frame to finish validation is tentative due to the hospital preparedness program's expected move to another office within HHS in 2007. As programs mature and more data become available, performance measures will continue to evolve to better measure outcomes. Because the process is iterative, the system allows for continuous improvements. Plans for Making Preparedness Information Public: CDC plans to issue a report by the end of 2007 providing a "snapshot" of the progress recipients have made in building emergency readiness capacity and addressing how CDC will measure capability in the future. HRSA officials said that decisions about whether to issue a report in 2007 on recipients' progress had not been made due to the hospital preparedness program's expected move to another office within HHS. Beginning in 2009, and every 4 years thereafter, the Pandemic and All- Hazards Preparedness Act requires that HHS report to the Congress on the status of public health emergency preparedness and response. * This includes a National Health Security Strategy and an implementation plan that includes an evaluation of progress made toward preparedness based on evidence-based benchmarks and objective standards that measure levels of preparedness. * The Act is generally silent on the type of information that is to be included in this evaluation other than an aggregate and recipient- specific breakdown of funding. [End of section] Enclosure II: Scope and Methodology: To determine how the Centers for Disease Control and Prevention's (CDC) and Health Resources and Services Administration's (HRSA) performance measurement systems have evolved, we reviewed and analyzed federal government documents related to national security and emergency preparedness, including the Executive Office of the President's National Strategy for Homeland Security and several Homeland Security Presidential Directives, and the Department of Homeland Security's (DHS) National Response Plan, Interim National Preparedness Goal, and draft Target Capabilities List. We interviewed officials from CDC and HRSA to identify and document how they have developed and implemented performance management systems for their cooperative agreement programs, including determining how standards were identified, indicators were selected, goals and targets were established, measures were defined, data systems were developed, and data were collected from recipients. We obtained and analyzed CDC and HRSA documents to identify the development of performance measures from program inception to the present, recipient reporting requirements, and systems for collecting data from cooperative agreement recipients. We also obtained reports and interviewed officials from federal agencies that had evaluated CDC's and HRSA's public health and hospital preparedness programs, including HHS's Office of Inspector General, HHS's Agency for Healthcare Research and Quality, and the Congressional Research Service. We also obtained reports and interviewed officials from professional associations involved in emergency preparedness and from policy research organizations that had published assessments or evaluations of public health and hospital preparedness programs. The professional associations included: * American Hospital Association, * Association of Professionals in Infection Control, * Association of Public Health Laboratories, * Association of State and Territorial Health Officials, * National Association of County and City Health Officials, * National Association of Public Hospitals, and: * The Joint Commission (formerly Joint Commission on Accreditation of Healthcare Organizations). The policy research organizations we contacted included: * Center for Studying Health System Changes, * National Center for Disaster Preparedness at Columbia University, * Public Health Foundation, * Rand Corporation, * The Century Foundation, and: * Trust for America's Health. We did not evaluate the actual performance measures adopted by CDC or HRSA. To determine how CDC and HRSA measure the progress of their preparedness programs, we interviewed CDC and HRSA officials to identify and document how they oversee and evaluate their cooperative agreement programs. To identify procedures used for reviewing recipient data and reporting results to applicable program managers, we obtained and analyzed documents and recipient-submitted progress reports from CDC and HRSA for program year 2004 and the first half of program year 2005 and interviewed CDC and HRSA project officers. Additionally, we analyzed documents to identify procedures in place for providing feedback and suggestions for program improvements to cooperative agreement recipients. We also reviewed documents and conducted interviews about the procedures used by project officers to provide recipients with feedback on their performance, share expertise on developing plans or conducting exercises, and disseminate "promising practices" information. We did not examine the accuracy or completeness of recipients' self-reported data in the progress reports submitted to CDC or HRSA. We conducted our work from June 2006 to March 2007 in accordance with generally accepted government auditing standards. [End of section] Enclosure III: Comments from Department of Health and Human Services: Office of the Assistant Secretary for Legislation: Department Of Health & Human Services: Washington, D.C. 20201: Mar 14 2001: Cynthia A. Bascetta: Director, Health Care: U.S. Government Accountability Office: Washington, DC 20548: Dear Ms. Bascetta: Enclosed are the Department's comments on the U.S. Government Accountability Office's (GAO) draft report entitled, "Public Health and Hospital Emergency Preparedness Programs: Evolution of Performance Measurement Systems to Measure Progress" (GAO-07-485R), before its publication. The Department appreciates the opportunity to comment on this draft report before its publication. Sincerely, Signed by: Vincent J. Ventimiglia, Jr. Assistant Secretary for Legislation: Comments On The Department Of Health And Human Services On The Government Accountability Office Draft Report Entitled "Public Health And Hospital Emergency Preparedness Programs Evolution Of Performance Measurement Systems To Measure Progress (GAO 07-485R): HHS Comments: On March 5, 2007, the Bioterrorism Hospital Preparedness Program (BHPP) transferred to the Office of the Assistant Secretary for Preparedness and Response. We maintained the existing staff assignments with the states to ensure continuity of support during the transition period and immediately afterward. We did, however, make some immediate and meaningful changes that we anticipate will greatly improve our ability to monitor performance at the individual awardee level and for the program overall. Most importantly, we have reassigned staff from the Office of the Assistant Secretary for Preparedness and Response to support the evaluation unit of the BHPP and reassigned former members of the evaluation unit staff to assignments that are a better fit for their skill set. This change will ensure that those individuals monitoring the performance of the awardees have the necessary analysis skills. In addition to strengthening the Program's evaluation unit, we have taken steps to establish partnerships with the Office of the Assistant Secretary for Evaluation and Policy here in the Department of Health and Human Services (HHS) and the Division of State and Local Readiness, Outcome Monitoring and Evaluation Branch, the evaluation unit at for the Public Health Emergency Program at CDC. Currently, these three units are working to develop the measurable, evidence-based benchmarks and objective standards for both programs as required by the Pandemic All-Hazards Preparedness Act (the Act). These benchmarks and standards will be vetted with our State and local stakeholders and finalized by the June 17, 2007 deadline specified in the Act. The establishment of these measures will allow us to monitor and track performance in a systematic and uniform manner during the upcoming BHPP project period, which includes Fiscal Years 2007-2011. Finally, we are preparing to conduct an analysis of performance data for existing BHPP awardees for the initial project period-Fiscal Years 2002-2006. We will utilize a contract currently in place to review and analyze the data and other information to be submitted by awardees through August 31, 2007. While we do not anticipate gleaning consistent information from all awardees, we do expect to develop a more complete picture of levels of preparedness for all awardees. We welcome an opportunity to share the results of our analysis during the next several months. [End of section] Enclosure IV: GAO Contact and Staff Acknowledgments: GAO Contact: Cynthia A. Bascetta at (202) 512-7101or bascettac@gao.gov: Acknowledgments: In addition to the contact name above, Karen Doran, Assistant Director; La Sherri Bush; Jeffrey Mayhew; Roseanne Price; Lois Shoemaker; and Cherie' Starck. (290537): FOOTNOTES [1] Pub. L. No. 107-188, 116 Stat. 594. The Pandemic and All-Hazards Preparedness Act, Pub. L. No. 109-417, 120 Stat. 2831 enacted December 19, 2006, reauthorized and amended the Preparedness and Response Act and authorized appropriations for HHS's Centers for Disease Control and Prevention's and Health Resources and Services Administration's public health and hospital preparedness programs through 2011. [2] CDC's program was formerly known as the Public Health Preparedness and Response for Bioterrorism Program. [3] These strategies also include the National Strategy for Pandemic Influenza and the National Security Strategy. [4] The federal agencies include HHS's Office of Inspector General (OIG), HHS's Agency for Healthcare Research and Quality (AHRQ), and the Congressional Research Service (CRS). [5] According to CDC officials, acquisition of personnel was necessary in order to develop and implement the activities authorized in the Preparedness and Response Act. [6] These strategies also include the National Strategy for Pandemic Influenza and the National Security Strategy. [7] Homeland Security Presidential Directives record and communicate presidential decisions about homeland security policies of the United States. [8] According to CDC officials, acquisition of personnel was necessary in order to develop and implement emergency preparedness activities authorized by the Preparedness and Response Act. GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site (www.gao.gov). Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to www.gao.gov and select "Subscribe to Updates." Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000 TDD: (202) 512-2537 Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Gloria Jarmon, Managing Director, JarmonG@gao.gov (202) 512-4400 U.S. Government Accountability Office, 441 G Street NW, Room 7125 Washington, D.C. 20548: Public Affairs: Paul Anderson, Managing Director, AndersonP1@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.