Results-Oriented Government
GPRA Has Established a Solid Foundation for Achieving Greater Results
Gao ID: GAO-04-38 March 10, 2004
Now that the Government Performance and Results Act (GPRA) has been in effect for 10 years, GAO was asked to address (1) the effect of GPRA in creating a governmentwide focus on results and the government's ability to deliver results to the American public, (2) the challenges agencies face in measuring performance and using performance information in management decisions, and (3) how the federal government can continue to shift toward a more results-oriented focus.
GPRA's requirements have established a solid foundation of results-oriented performance planning, measurement, and reporting in the federal government. Federal managers surveyed by GAO reported having significantly more of the types of performance measures called for by GPRA. GPRA has also begun to facilitate the linking of resources to results, although much remains to be done in this area to increase the use of performance information to make decisions about resources. We also found agency strategic and annual performance plans and reports we reviewed have improved over initial efforts. Although a foundation has been established, numerous significant challenges to GPRA implementation still exist. Inconsistent top leadership commitment to achieving results within agencies and Office of Management and Budget (OMB) can hinder the development of results-oriented cultures in agencies. Furthermore, in certain areas, federal managers continue to have difficulty setting outcome-oriented goals, collecting useful data on results, and linking institutional, program, unit, and individual performance measurement and reward systems. Finally, there is an inadequate focus on addressing issues that cut across federal agencies. OMB, as the focal point for management in the federal government, is responsible for overall leadership and direction in addressing these challenges. OMB has clearly placed greater emphasis on management issues during the past several years. However, it has showed less commitment to GPRA implementation in its guidance to agencies and in using the governmentwide performance plan requirement of GPRA to develop an integrated approach to crosscutting issues. In our view, governmentwide strategic planning could better facilitate the integration of federal activities to achieve national goals.
Recommendations
Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.
Director:
Team:
Phone:
GAO-04-38, Results-Oriented Government: GPRA Has Established a Solid Foundation for Achieving Greater Results
This is the accessible text file for GAO report number GAO-04-38
entitled 'Results-Oriented Government: GPRA Has Established a Solid
Foundation for Achieving Greater Results' which was released on March
10, 2004.
This text file was formatted by the U.S. General Accounting Office
(GAO) to be accessible to users with visual impairments, as part of a
longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to Congressional Requesters:
March 2004:
RESULTS-ORIENTED GOVERNMENT:
GPRA Has Established a Solid Foundation for Achieving Greater Results:
GAO-04-38:
GAO Highlights:
Highlights of GAO-04-38, a report to congressional requesters
Why GAO Did This Study:
Now that the Government Performance and Results Act (GPRA) has been in
effect for 10 years, GAO was asked to address (1) the effect of GPRA in
creating a governmentwide focus on results and the government‘s ability
to deliver results to the American public, (2) the challenges agencies
face in measuring performance and using performance information in
management decisions, and (3) how the federal government can continue
to shift toward a more results-oriented focus.
What GAO Found:
GPRA‘s requirements have established a solid foundation of results-
oriented performance planning, measurement, and reporting in the
federal government. Federal managers surveyed by GAO reported having
significantly more of the types of performance measures called for by
GPRA (see figure below). GPRA has also begun to facilitate the linking
of resources to results, although much remains to be done in this area
to increase the use of performance information to make decisions about
resources. We also found agency strategic and annual performance plans
and reports we reviewed have improved over initial efforts.
Although a foundation has been established, numerous significant
challenges to GPRA implementation still exist. Inconsistent top
leadership commitment to achieving results within agencies and OMB can
hinder the development of results-oriented cultures in agencies.
Furthermore, in certain areas, federal managers continue to have
difficulty setting outcome-oriented goals, collecting useful data on
results, and linking institutional, program, unit, and individual
performance measurement and reward systems. Finally, there is an
inadequate focus on addressing issues that cut across federal agencies.
OMB, as the focal point for management in the federal government, is
responsible for overall leadership and direction in addressing these
challenges. OMB has clearly placed greater emphasis on management
issues during the past several years. However, it has showed less
commitment to GPRA implementation in its guidance to agencies and in
using the governmentwide performance plan requirement of GPRA to
develop an integrated approach to crosscutting issues. In our view,
governmentwide strategic planning could better facilitate the
integration of federal activities to achieve national goals.
What GAO Recommends:
GAO recommends that the Office of Management and Budget (OMB) improve
its guidance and oversight of GPRA implementation, as well as develop a
governmentwide performance plan. GAO also believes Congress should
consider amending GPRA to require that (1) agencies update their
strategic plans at least once every four years, consult with
congressional stakeholders at least once every new Congress, and make
interim updates to strategic and performance plans as appropriate; and
(2) the President develop a governmentwide strategic plan. OMB
generally agreed with our recommendations, but stated that the
President‘s Budget can serve as both a governmentwide strategic and
annual plan. However, we believe the budget provides neither a long-
term nor an integrated perspective on the federal government‘s
performance.
www.gao.gov/cgi-bin/getrpt?GAO-04-38.
To view the full product, including the scope and methodology, click on
the link above. For more information, contact Patricia A. Dalton at
(202) 512-6806 or daltonp@gao.gov.
[End of section]
Contents:
Transmittal Letter:
Executive Summary:
Purpose:
Background:
Results in Brief:
Principal Findings:
Recommendations for Executive Action:
Matters for Congressional Consideration:
Agency Comments:
Chapter 1: Introduction:
Impact of Emerging Trends and Fiscal Challenges:
GPRA Background:
Scope and Methodology:
Chapter 2: GPRA Established the Foundation for a More Results-Oriented
Federal Government:
GPRA Statutory Requirements Laid a Foundation for Agencywide Results-
Oriented Management:
Chapter 3: Agencies Have Addressed Many Critical Performance Planning
and Reporting Challenges, but Weaknesses Persist:
Quality of Selected Strategic Plans Reflects Improvements over Initial
Drafts:
Fiscal Year 2004 Annual Performance Plans Addressed Some Weaknesses of
Earlier Plans, but Still Have Room for Significant Improvement:
Strengths and Weaknesses of Selected Agencies' Fiscal Year 2002 Annual
Performance Reports:
Chapter 4: Challenges to GPRA Implementation Persist:
Top Leadership Does Not Consistently Show Commitment to Achieving
Results:
Managers Report Mixed Results in Use of Performance Information:
Managers Continue to Confront a Range of Human Capital Management
Challenges:
Persistent Challenges in Setting Outcome-Oriented Goals, Measuring
Performance, and Collecting Useful Data:
Crosscutting Issues Hinder Successful GPRA Implementation:
Managers View Congress' Use of Performance Information as Limited:
Chapter 5: Conclusions and Recommendations:
Agenda for Achieving a Sustainable, Governmentwide Focus on Results:
Recommendations for Executive Action:
Matters for Congressional Consideration:
Agency Comments:
Appendixes:
Appendix I: Objectives, Scope, and Methodology:
Methodology for Governmentwide Survey:
Methodology for Focus Groups:
Methodology for Interviews with Political Appointees:
Methodology for Selecting Agencies to Review for Changes in the Quality
of Their Strategic Plans, Annual Performance Plans, and Annual
Performance Reports:
Appendix II: Focus Group Participants Agreed GPRA Provides a Framework
for Federal Agencies to Become More Results Oriented:
GPRA Accomplishments:
Views on Delivering Results to the American Public Were Mixed:
Alternate Views on GPRA's Effect:
Challenges in Implementing and Overseeing GPRA Activities:
Suggestions to Address GPRA Challenges:
Appendix III: Observations on Agencies' Strategic Plans:
Required Elements of Agency Strategic Plans:
Observations on Changes in the Quality of Education's Strategic Plan:
Observations on Changes in the Quality of DOE's Strategic Plan:
Observations on Changes in the Quality of HUD's Strategic Plan:
Observations on Changes in the Quality of SBA's Strategic Plan:
Observations on Changes in the Quality of SSA's Strategic Plan:
Observations on Changes in the Quality of DOT's Strategic Plan:
Appendix IV: Observations on Agencies' Annual Performance Plans:
Key Elements of Information for Annual Performance Plans:
Observations on Changes in the Quality of Education's Annual
Performance Plan:
Observations on Changes in the Quality of DOE's Annual Performance
Plan:
Observations on Changes in the Quality of HUD's Annual Performance
Plan:
Observations on Changes in the Quality of SBA's Annual Performance
Plan:
Observations on Changes in the Quality of SSA's Annual Performance
Plan:
Observations on Changes in the Quality of DOT's Annual Performance
Plan:
Appendix V: Observations on Agencies' Annual Performance and
Accountability Reports:
Observations on the Quality of Education's Fiscal Year 2002 Performance
and Accountability Report:
Observations on the Quality of DOE's Fiscal Year 2002 Annual
Performance and Accountability Report:
Observations on the Quality of HUD's Fiscal Year 2002 Annual
Performance and Accountability Report:
Observations on the Quality of SBA's Fiscal Year 2002 Annual
Performance and Accountability Report:
Observations on the Quality of SSA's Fiscal Year 2002 Annual
Performance and Accountability Report:
Observations on the Quality of DOT's Fiscal Year 2002 Annual
Performance and Accountability Report:
Appendix VI: GAO Federal Managers' Survey Data:
Appendix VII: Agencies Subject to the Chief Financial Officers Act:
Appendix VIII: Comments from the Office of Management and Budget:
Appendix IX: Comments from the Department of Energy:
GAO Comments:
Appendix X: Comments from the Department of Housing and Urban
Development:
GAO Comments:
Appendix XI: Comments from the Social Security Administration:
GAO Comments:
Appendix XII: GAO Contact and Staff Acknowledgments:
GAO Contact:
Acknowledgments:
Related GAO Products:
GPRA/Managing for Results:
Strategic Human Capital Management:
Linking Resources to Results:
Measuring Performance:
Data Credibility:
Using Performance Information:
Tables:
Table 1: Agencies' Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Table 2: Characterizations of Agencies' Fiscal Year 1999 and 2004
Annual Performance Plans:
Table 3: Characterizations of Agencies' 2002 Annual Performance Reports:
Table 4: Summary of Characteristics of Agencies Selected for Review of
Strategic Plans, Annual Performance Plans, and Annual Performance
Reports:
Table 5: Agencies' Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Table 6: Education's Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Table 7: DOE's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Table 8: HUD's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Table 9: SBA's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Table 10: SSA's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Table 11: DOT's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Table 12: Characterizations of Agencies' Annual Performance Plans:
Table 13: Characterizations of Agencies' Fiscal Year 2002 Annual
Performance and Accountability Reports:
Figures:
Figure 1: Composition of Spending as a Share of GDP Assuming
Discretionary Spending Grows with GDP after 2003 and All Expiring Tax
Provisions Are Extended:
Figure 2: Percentage of Federal Managers Who Reported That There Were
Performance Measures for the Programs with Which They Were Involved:
Figure 3: Percentage of Federal Managers Who Reported Having Specific
Types of Performance Measures to a Great or Very Great Extent:
Figure 4: Percentage of Federal Managers Who Reported Their Awareness
of GPRA:
Figure 5: Percentage of Federal Managers Who Reported Hindrances to
Measuring Performance or Using the Performance Information to a Great
or Very Great Extent:
Figure 6: Percentage of Federal Managers and SES Managers Who Reported
That OMB Paid Attention to Their Agency's Efforts under GPRA to a Great
or Very Great Extent:
Figure 7: Percentage of Federal Managers Who Reported They Considered
Strategic Goals to a Great or Very Great Extent When Allocating
Resources:
Figure 8: Percentage of Federal Managers Who Reported They Considered
Performance Information to a Great or Very Great Extent When Allocating
Resources:
Figure 9: Percentage of Federal Managers Who Reported That Funding
Decisions Were Based on Results or Outcome-Oriented Performance
Information to a Great or Very Great Extent:
Figure 10: Percentage of Federal Managers Who Reported to a Great or
Very Great Extent Their Top Leadership Has a Strong Commitment to
Achieving Results:
Figure 11: Percentage of SES and Non-SES Managers Who Reported to a
Great or Very Great Extent Their Agency Top Leadership Demonstrated
Strong Commitment to Achieving Results:
Figure 12: Percentage of Federal Managers Who Reported Using
Information Obtained from Performance Measurement to a Great or Very
Great Extent for Various Management Activities:
Figure 13: Percentage of Federal Managers Responding "Yes" about Being
Involved in the Following Activities:
Figure 14: Percentage of Federal Managers Reporting to a Great or Very
Great Extent That Managers/Supervisors at Their Levels Had the
Decision-Making Authority They Needed to Help the Agency Accomplish Its
Strategic Goals:
Figure 15: Percentage of Federal Managers, SES, and Non-SES in 2003
Reporting to a Great or Very Great Extent That They Were Held
Accountable for the Accomplishment of Agency Strategic Goals:
Figure 16: Percentage of Federal Managers in Each Survey Year Who
Reported That during the Past 3 Years Their Agencies Provided,
Arranged, or Paid for Training That Would Help Them Accomplish Specific
Tasks:
Figure 17: Percentage of Federal Managers Who Reported to a Great or
Very Great Extent That Employees in Their Agencies Received Positive
Recognition for Helping Their Agencies Accomplish Their Strategic
Goals:
Figure 18: Percentage of Federal Managers Reporting to a Great or Very
Great Extent That a Lack of Ongoing Congressional Commitment or Support
for Using Performance Information in Making Program/Funding Decisions
Is a Hindrance:
Figure 19: Summary of Education's Performance Indicators for Fiscal
Year 2002:
Figure 20: Inputs: Allocating Funds for Education's Objective to Ensure
That All Students Read on Grade Level by the Third Grade:
Figure 21: Summary of DOE's Performance Indicators for Fiscal Year
2002:
Figure 22: Summary of HUD's Performance Indicators for Fiscal Year
2002:
Figure 23: Summary of SBA's Performance Goals for Fiscal Year 2002:
Figure 24: Summary of SSA's Performance Goals for Fiscal Year 2002:
Figure 25: Summary of DOT's Performance Indicators for Fiscal Year
2002:
Abbreviations:
AP: advanced placement:
CDBG: Community Development Block Grants:
CFO: Chief Financial Officer:
CRS: Congressional Research Service:
DOE: Department of Energy:
DOT: Department of Transportation:
EPA: Environmental Protection Agency:
FAA: Federal Aviation Administration:
FSA: Federal Student Assistance:
FTE: full-time employee:
GM: general management:
GPRA: Government Performance and Results Act of 1993:
GS: general schedule:
HHS: Department of Health and Human Services:
HOME: Home Investment Partnership Program:
HUD: Department of Housing and Urban Development:
ICH: Interagency Council on the Homeless:
IG: Inspector General:
IT: information technology:
IRS: Internal Revenue Service:
JARC: Job Access and Reverse Commute:
NAEP: National Assessment for Educational Progress:
NASA: National Aeronautics and Space Administration:
OASI: Old Age and Survivors Insurance:
OMB: Office of Management and Budget:
OPM: Office of Personnel Management:
PART: Program Assessment Rating Tool:
PMA: President‘s Management Agenda:
SBA: Small Business Administration:
SBDC: Small Business Development Centers:
SES: Senior Executive Service:
SSA: Social Security Administration:
SSI: Supplemental Security Income:
SSN: Social Security number:
VA: Department of Veterans Affairs:
Transmittal Letter March 10, 2004:
Congressional Requesters:
As you requested, we have assessed the effectiveness of the Government
Performance and Results Act (GPRA), in light of its 10-year anniversary
in 2003. Our review focused on GPRA's accomplishments, challenges to
its continued implementation, and an agenda for achieving a
sustainable, governmentwide focus on results.
Upon issuance, we will send copies to the Director of the Office of
Management and Budget and executive branch agencies (see appendix VII
for a list). We will also make copies available to others upon request.
In addition, this report will be available at no charge on the GAO web
site at [Hyperlink, http://www.gao.gov].
If you have any questions concerning this report, please contact
Patricia A. Dalton at (202) 512-6806 or daltonp@gao.gov. The major
contributors to this report are listed in appendix XII.
Signed by:
David M. Walker:
Comptroller General of the United States:
List of Requesters:
The Honorable Susan M. Collins:
Chairman:
The Honorable Joe Lieberman:
Ranking Minority Member:
Committee on Governmental Affairs:
United States Senate:
The Honorable George V. Voinovich:
Chairman:
The Honorable Richard Durbin:
Ranking Minority Member:
Subcommittee on Oversight of Government Management, the Federal
Workforce, and the District of Columbia:
Committee on Governmental Affairs:
United States Senate:
The Honorable Peter G. Fitzgerald:
Chairman:
The Honorable Daniel K. Akaka:
Ranking Minority Member:
Subcommittee on Financial Management, the Budget, and International
Security:
Committee on Governmental Affairs:
United States Senate:
The Honorable Tom Davis:
Chairman:
The Honorable Henry A. Waxman:
Ranking Minority Member:
Committee on Government Reform:
House of Representatives:
The Honorable Todd Russel Platts:
Chairman:
The Honorable Edolphus Towns:
Ranking Minority Member:
Subcommittee on Government Efficiency and Financial Management:
Committee on Government Reform:
House of Representatives:
[End of section]
Executive Summary:
Purpose:
From defending the homeland against terrorists, to preventing the
spread of infectious diseases, to providing a reliable stream of social
security income to retirees and supporting the transition from welfare
to work, the federal government provides funding and services to the
American public that can affect their lives in critical ways every day.
However, the federal government is in a period of profound transition
and faces an array of challenges and opportunities to enhance
performance, ensure accountability, and position the nation for the
future. A number of overarching trends, such as diffuse security
threats and homeland security needs, increasing global interdependency,
the shift to knowledge-based economies, and the looming fiscal
challenges facing our nation drive the need to reconsider the proper
role for the federal government in the 21st century, how the government
should do business (including how it should be structured), and in some
instances, who should do the government's business.
Without effective short-and long-term planning, which takes into
account the changing environment and needs of the American public and
the challenges they face and establishes goals to be achieved, federal
agencies risk delivering programs and services that may or may not meet
society's most critical needs. At a cost to taxpayers of over $2
trillion annually, the federal government should be able to demonstrate
to the American public that it can anticipate emerging issues, develop
sound strategies and plans to address them, and be accountable for the
results that have been achieved.
Concerned that the federal government was more focused on program
activities and processes than the results to be achieved, Congress
passed the Government Performance and Results Act of 1993
(GPRA).[Footnote 1] The act required federal agencies to develop
strategic plans with long-term, outcome-oriented goals and objectives,
annual goals linked to achieving the long-term goals, and annual
reports on the results achieved. Now that GPRA has been in effect for
10 years, you asked us to assess the effectiveness of GPRA in creating
a focus on results in the federal government. Specifically, this report
discusses (1) the effect of GPRA over the last 10 years in creating a
governmentwide focus on results and the government's ability to deliver
results to the American public, including an assessment of the changes
in the overall quality of agencies' strategic plans, annual performance
plans, and annual performance reports; (2) the challenges agencies face
in measuring performance and using performance information in
management decisions; and (3) how the federal government can continue
to shift toward a more results-oriented focus.
To meet our objectives, we reviewed our extensive prior work on GPRA
best practices and implementation and collected governmentwide data to
assess the government's overall focus on results. We conducted a
random, stratified, governmentwide survey of federal managers
comparable to surveys we conducted in 1997 and 2000. We also held eight
in-depth focus groups--seven comprised of federal managers from 23
federal agencies and one with GPRA experts. We also interviewed top
appointed officials from the current and previous administrations.
Finally, we judgmentally selected a sample of six agencies to review
for changes in the quality of their strategic plans, performance plans,
and performance reports since their initial efforts. The agencies we
selected included the Departments of Education (Education), Energy
(DOE), Housing and Urban Development (HUD), and Transportation (DOT)
and the Small Business (SBA) and Social Security Administrations (SSA).
In making this selection, we chose agencies that collectively
represented the full range of characteristics in the following four
areas: (1) agency size (small, medium, large); (2) primary program type
(direct service, research, regulatory, transfer payments, and contracts
or grants); (3) quality of fiscal year 2000 performance plans based on
our previous review;[Footnote 2] and (4) type of agency (cabinet
department and independent agency). Appendix I contains a more detailed
discussion of our scope and methodology. We performed our work in
Washington, D.C., from January through November 2003 in accordance with
generally accepted government auditing standards.
Background:
GPRA is the centerpiece of a statutory framework that Congress put in
place during the 1990s to help resolve the long-standing management
problems that have undermined the federal government's efficiency and
effectiveness and to provide greater accountability for results. GPRA
was intended to address several broad purposes, including strengthening
the confidence of the American people in their government; improving
federal program effectiveness, accountability, and service delivery;
and enhancing congressional decision making by providing more objective
information on program performance.
GPRA requires executive agencies to complete strategic plans in which
they define their missions, establish results-oriented goals, and
identify the strategies that will be needed to achieve those goals.
GPRA requires agencies to consult with Congress and solicit the input
of others as they develop these plans. Through this strategic planning
requirement, GPRA has required federal agencies to reassess their
missions and long-term goals as well as the strategies and resources
they will need to achieve their goals. Agencies developed their first
strategic plans in fiscal year 1997, and are required to update the
plans every 3 years since then.
GPRA also requires executive agencies to prepare annual performance
plans that articulate goals for the upcoming fiscal year that are
aligned with their long-term strategic goals. These performance plans
are to include results-oriented annual goals linked to the program
activities displayed in budget presentations as well as the indicators
the agency will use to measure performance against the results-oriented
goals. Agencies developed their first annual performance plans in
fiscal year 1999 and are required to issue plans annually thereafter to
correspond with budget submissions to Congress.
Finally, GPRA requires agencies to measure performance toward the
achievement of the goals in the annual performance plan and report
annually on their progress in program performance reports. If a goal
was not met, the report is to provide an explanation and present the
actions needed to meet any unmet goals in the future. These reports are
intended to provide important information to agency managers,
policymakers, and the public on what each agency accomplished with the
resources it was given. Agencies issued their first annual performance
reports on their fiscal year 1999 performance in fiscal year 2000 and
are required to issue a report on each subsequent performance plan.
The Office of Management and Budget (OMB) plays an important role in
the management of federal government performance and specifically GPRA
implementation. Part of OMB's overall mission is to ensure that agency
plans and reports are consistent with the President's budget and
administration policies. OMB is responsible for receiving and reviewing
agencies' strategic plans, annual performance plans, and annual
performance reports. To improve the quality and consistency of these
documents, OMB issues annual guidance to agencies for their
preparation, including guidelines on format, required elements, and
submission deadlines. GPRA requires OMB to prepare a governmentwide
performance plan, based on agencies' annual performance plan
submissions. OMB also played an important role in the pilot phase of
GPRA implementation by designating agencies for pilot projects in
performance measurement, managerial accountability and flexibility,
and performance budgeting, and assessing the results of the pilots.
Finally, GPRA provides OMB with authority to grant agencies waivers to
certain administrative procedures and controls.
Recent OMB guidance--OMB Circular A-11, July 2003--requires agencies to
submit "performance budgets" in lieu of annual performance plans for
their fiscal year 2005 budget submission to OMB and Congress. According
to OMB, performance budgets should satisfy all the statutory
requirements of GPRA for annual performance plans. In addition,
agencies are to include all performance goals used in the assessment of
program performance done under OMB's Program Assessment Rating Tool
(PART) process.[Footnote 3] Moreover, the guidance states that until
all programs have been assessed by PART, the performance budget will
also for a time include performance goals for agency programs that have
not yet been assessed using PART. The expectation is that agencies are
to substitute new or revised performance goals resulting from OMB's
review for goals it deemed unacceptable.
Results in Brief:
Among the purposes of GPRA cited by Congress was to improve federal
program effectiveness and service delivery by promoting a new focus on
results, service quality, and customer satisfaction by setting program
goals measuring performance against goals, and reporting publicly on
progress. Furthermore, GPRA was to improve congressional decision
making by providing more objective information on achieving objectives,
and on the relative effectiveness and efficiency of federal programs
and spending. Ten years after enactment, GPRA's requirements have laid
a solid foundation of results-oriented agency planning, measurement,
and reporting that have begun to address these purposes. Focus group
participants and high-level political appointees, as well as OMB
officials we interviewed, cited positive effects of GPRA that they
generally attributed to GPRA's statutory requirements for planning and
reporting. Performance planning and measurement have slowly yet
increasingly become a part of agencies' cultures. The results of our
stratified, random sample survey of federal managers indicate that
since GPRA went into effect governmentwide in 1997, federal managers
reported having significantly more of the types of performance measures
called for by GPRA--particularly outcome-oriented performance
measures. Survey data also suggested that more federal managers,
especially at the Senior Executive Service (SES) level, believed that
OMB was paying attention to their agencies' efforts under GPRA. GPRA
has also begun to facilitate the linking of resources to results,
although much remains to be done in this area.
Beginning with agencies' initial efforts to develop effective strategic
plans in 1997 and annual performance plans and reports for fiscal year
1999, Congress, GAO, and others have commented on the quality of those
efforts and provided constructive feedback on how agency plans and
reports could be improved. According to our current review of the
strategic plans, annual performance plans, and annual performance
reports of six selected agencies, these documents reflect much of the
feedback that was provided. For example, goals are more quantifiable
and results oriented, and agencies are providing more information about
goals and strategies to address performance and accountability
challenges and the limitations to their performance data. However,
certain serious weaknesses persist, such as lack of detail on how
annual performance goals relate to strategic goals and how agencies are
coordinating with other entities to address common challenges and
achieve common objectives.
While a great deal of progress has been made in making federal agencies
more results oriented, numerous challenges still exist. As we have
noted before, top leadership commitment and sustained attention to
achieving results, both within the agencies and at OMB, is essential to
GPRA implementation. While one might expect an increase in agency
leadership commitment since GPRA was implemented governmentwide
beginning in fiscal year 1997, federal managers reported that such
commitment has not significantly increased. Furthermore, although OMB
has recently demonstrated leadership in its review of performance
information from a budgetary perspective using the PART tool, it is
unclear whether the results of those reviews, such as changes in
program performance measures, will complement and be integrated with
the long-term, strategic focus of GPRA. OMB provided significantly less
guidance on GPRA implementation for the fiscal year 2005 budget,
compared to the very detailed guidance provided in prior years. Without
consistent guidance from OMB on meeting GPRA requirements and following
best practices, it may be difficult to maintain the improvements in
agency performance plans and reports or bring about improvements in
areas where weaknesses remain. The commitment of top leadership within
agencies, OMB, and Congress is critical to the success of strategic
planning efforts. However, GPRA specifies time frames for updating
strategic plans that do not correspond to presidential or congressional
terms. As a result, an agency may be required to update its strategic
plan a year before a presidential election and without input from a new
Congress. A strategic plan should reflect the policy priorities of an
organization's leaders and the input of key stakeholders if it is to be
an effective management tool.
Managers reported they had more performance measures, but indications
that managers are making greater use of this information to improve
performance are mixed. Additionally, managers reported several human
capital-related challenges that impede results-oriented management,
including a lack of authority and training to carry out GPRA
requirements, as well as a lack of recognition for completing these
tasks. Unfortunately, most existing federal performance appraisal
systems are not designed to support a meaningful performance-based pay
system in that they fail to link institutional, program, unit, and
individual performance measurement and reward systems. Fewer than half
of federal managers reported receiving relevant training in critical
results-oriented management-related tasks. Managers also reported
significant challenges persist in setting outcome-oriented goals,
measuring performance, and collecting useful data. In some agencies,
particularly those that have a research and development component,
managers reported difficulties in establishing meaningful outcome
measures. Managers also identified difficulties in distinguishing
between the results produced by the federal program and results caused
by external factors or nonfederal actors, such as with grant programs.
Timely and useful performance information is not always available to
federal agencies, making it more difficult to assess and report on
progress achieved. Finally, agency officials believe that Congress
could make greater use of performance information to conduct oversight
and to inform appropriations decisions. GPRA provides a vehicle for
Congress to explicitly state its performance expectations in outcome-
oriented terms when establishing new programs or in exercising
oversight of existing programs that are not achieving desired results.
Mission fragmentation and overlap contribute to difficulties in
addressing crosscutting issues, particularly when those issues require
a national focus, such as homeland security, drug control, and the
environment. GPRA requires a governmentwide performance plan, where
these issues could be addressed in a centralized fashion, but OMB has
not issued a distinct plan since 1999. Most recently, the President's
fiscal year 2004 budget focused on describing agencies' progress in
addressing the President's Management Agenda (PMA) and the results of
PART reviews of agency programs. Such information is important and
useful, but is not adequate alone to provide a broader and more
integrated perspective of planned performance on governmentwide
outcomes. GAO has previously reported on a variety of barriers to
interagency cooperation, such as conflicting agency missions,
jurisdiction issues, and incompatible procedures, data, and processes.
A strategic plan for the federal government, supported by a set of key
national indicators to assess the government's performance, position,
and progress, could provide an additional tool for governmentwide
reexamination of existing programs, as well as proposals for new
programs. Such a plan could be of particular value in linking agencies'
long-term performance goals and objectives horizontally across the
government. In addition, it could provide a basis for integrating,
rather than merely coordinating, a wide array of federal activities.
To address these challenges, continued and sustained commitment and
leadership are needed. OMB, as the primary focal point for overall
management in the federal government, can provide this leadership and
direction working with the various management councils and work groups
of the government. Also, governmentwide planning could better
facilitate the integration of federal activities to achieve national
goals.
GAO recommends that the Director of OMB (1) fully implement GPRA's
requirement to develop a governmentwide performance plan; (2)
articulate and implement an integrated, complementary relationship
between GPRA and PART; (3) provide clearer and consistent guidance to
executive branch agencies on how to implement GPRA; (4) continue to
maintain a dialogue with agencies about their performance measurement
practices with a particular focus on grant-making, research and
development, and regulatory functions to identify and replicate
successful approaches agencies are using to measure and report on their
outcomes, including the use of program evaluation tools; and, work with
executive branch agencies to identify the barriers to obtaining timely
data to show progress against performance goals and the best ways to
report information when there are unavoidable lags in data
availability; and (5) work with agencies to ensure they are making
adequate investments in training on performance planning and
measurement, with a particular emphasis on how to use performance
information to improve program performance.
We also suggest that Congress consider amending GPRA to require that
updates to agency strategic plans be submitted at least once every 4
years, 12-18 months after a new administration begins its term.
Additionally, consultations with congressional stakeholders on
existing strategic plans should be held at least once every new
Congress and revisions should be made as needed. Further, we suggest
Congress use these consultations and its oversight role to clarify its
performance expectations for agencies. Congress should also consider
amending GPRA to require the President to develop a governmentwide
strategic plan.
In commenting on a draft of this report, OMB generally agreed with our
findings and conclusions. OMB agreed to implement most of our
recommendations, but stated that the President's Budget represents the
executive branch's governmentwide performance plan and could also serve
as a governmentwide strategic plan. However, because of the budget's
focus on agency-level expenditures for the upcoming fiscal year, we
believe that the President's Budget provides neither a long-term nor an
integrated perspective on the federal government's performance. OMB's
comments appear in appendix VIII. Our response appears in chapter 5. We
also provided relevant sections of the draft to the six agencies whose
plans and reports we reviewed. DOE, HUD, and SSA disagreed with some of
our observations, and we changed or clarified relevant sections of the
report, as appropriate. Written comments from DOE, HUD, and SSA are
reprinted in appendixes IX, X, and XI, respectively, along with our
responses.
Principal Findings:
GPRA Laid the Foundation for a More Results-Oriented Federal
Government:
Prior to enactment of GPRA, our 1992 review of the collection and use
of performance data by federal agencies revealed that, although many
agencies collected performance information at the program level, few
agencies had results-oriented performance information to manage or
make strategic policy decisions for the agency as a whole.[Footnote 4]
GPRA addressed agencies' shortcomings by creating a comprehensive and
consistent statutory foundation of required agencywide strategic plans,
annual performance plans, and annual performance reports. Participants
in eight focus groups comprised of experts on GPRA and federal managers
from 23 agencies cited the creation of this statutory foundation as one
of the key accomplishments of GPRA. One of the premises of GPRA is that
both congressional and executive branch oversight of federal agency
performance were seriously hampered by a lack of adequate results-
oriented goals and performance information. As noted above, prior to
the enactment of GPRA few agencies reported their performance
information externally. OMB officials we interviewed as part of our
current review suggested that OMB has been a key consumer of the
performance information produced under GPRA and that it has provided a
foundation for their efforts to oversee agency performance.
Federal managers' views of GPRA's effect on the federal government's
ability to deliver results to the American public were mixed. When
asked about the direct effects of GPRA on the public, 23 percent of the
federal managers surveyed agreed to a moderate or greater extent that
GPRA improved their agency's ability to deliver results to the American
public. High-level political appointees we interviewed cited a number
of examples of how the structure of GPRA created a greater focus on
results in their agencies. Participants in our focus groups had mixed
perceptions of GPRA's effect on their agency's ability to deliver
results to the American public. Participants indicated GPRA has had a
positive effect by shifting the focus of federal management from
program activities and processes to achieving the intended results of
those programs. Another major accomplishment of GPRA cited by focus
group participants is that GPRA improved the transparency of government
results to the American public. Other focus group participants had
difficulty attributing the results their agencies achieved directly to
GPRA's requirements.
Focus group and survey results suggest that performance planning and
measurement have slowly, but increasingly, become a part of agencies'
cultures. Compared to the results of our 1997 governmentwide survey of
federal managers, in our 2003 governmentwide survey more managers
reported having performance measures for their programs. When we asked
managers who said they had performance measures which of the five types
of measures they had to a great or very great extent, they reported
increases in all five types of measures between 1997 and 2003,[Footnote
5] all of which were statistically significant.
Similarly, focus group participants commented on certain cultural
changes that had taken place within their agencies since the passage of
GPRA in which the "vocabulary" of performance planning and measurement-
-e.g., a greater focus on performance measurement, orientation toward
outcomes over inputs and outputs, and an increased focus on program
evaluation--had become more pervasive. This perception is partly born
out by our survey results. Consistent with our survey results
indicating increases in results-oriented performance measures, we also
observed a significant decline in the percentage of federal managers
who agreed that certain factors hindered measuring performance or using
the performance information. Finally, our survey data suggested that
more federal managers, especially at the SES level, believed that OMB
was paying attention to their agencies' efforts under GPRA, but with no
corresponding increase in their concern that OMB would micromanage the
programs in their agencies.
Agencies have begun to establish a link between results and resources.
Our 1998 assessment of fiscal year 1999 performance plans found that
agencies generally covered the program activities in their budgets, but
most plans did not identify how the funding for those program
activities would be allocated to performance goals.[Footnote 6]
However, our subsequent reviews of performance plans indicate that
agencies have made progress in demonstrating how their performance
goals and objectives relate to program activities in the budget.
We reviewed a sample of six agencies' strategic plans (Education, DOE,
HUD, DOT, SBA, and SSA) and found the quality of the selected plans
reflected improvements over these agencies' initial strategic plans.
Our 1997 review of agencies' draft strategic plans found that a
significant amount of work remained to be done by executive branch
agencies if their strategic plans were to fulfill the requirements of
GPRA, serve as a basis for guiding agencies, and help congressional and
other policymakers make decisions about agency activities and
programs.[Footnote 7] The six strategic plans we looked at for this
2003 review reflected many new and continuing strengths as well as
improvements over the 1997 initial draft plans, but we continued to
find certain persistent weaknesses. Of the six elements required by
GPRA, the plans generally discussed all but one--program evaluation, an
area in which we have found agencies often lack capacity. Although the
strategic plans listed the program evaluations agencies intended to
complete over the planning period, they generally did not address how
the agencies planned to use their evaluations to establish new or
revise existing strategic goals, as envisioned by GPRA. Finally,
although not required by GPRA, the strategic plans would have benefited
from more complete discussions of how agencies planned to coordinate
and collaborate with other entities to address common challenges and
achieve common or complementary goals and objectives.
The six selected agencies' fiscal year 2004 annual performance plans
addressed some weaknesses of earlier plans, but there is still
significant room for improvement. During our review of agencies' first
annual performance plans, which presented agencies' annual performance
goals for fiscal year 1999,[Footnote 8] we found that substantial
further development was needed for these plans to be useful in a
significant way to congressional and other decision makers. Most of the
2004 plans that we reviewed showed meaningful improvements over the
fiscal year 1999 plans by showing a clearer picture of intended
performance, providing strategies and resources that were more
specifically related to achieving agency goals, and providing a greater
level of confidence that performance data would be credible. But these
plans also contained a number of serious weaknesses, such as inadequate
discussion of coordination and collaboration and inconsistent or
limited discussions of procedures used to verify and validate
performance data, which limited their quality and undermined their
usefulness.
Our review of the six agencies' fiscal year 2002 performance reports
showed a number of strengths and improvements over their fiscal year
1999 performance reports, as well as areas that needed improvement. As
we found in our earlier reviews, the six agencies' fiscal year 2002
reports generally allowed for an assessment of progress made in
achieving agency goals. In addition, the majority of agencies discussed
the progress achieved in addressing performance and accountability
challenges identified by agency inspectors general and GAO. However, as
with the fiscal year 1999 reports, many of the weaknesses we identified
in the agencies' fiscal year 2002 reports were related to the
significant number of performance goals not achieved or for which
performance data were unavailable. In addition, the majority of the
reports we reviewed did not include other GPRA requirements, such as a
summary of the findings from program evaluations. Finally, only one of
the six agencies clearly linked its costs to the achievement of
performance goals or objectives.
Challenges to GPRA Implementation Exist:
While a great deal of progress has been made in making federal agencies
more results oriented, numerous challenges still exist to effective
implementation of GPRA. We observed in our 1997 report that we would
expect to see managers' positive perceptions on items, such as the
extent to which top leadership is committed to achieving results,
become more prevalent and the gap between SES and non-SES managers
begin to narrow as GPRA and related reforms are implemented. However,
these changes do not appear to be happening to the extent anticipated.
The need for strong, committed, and sustained leadership extends to OMB
as well. OMB has shown a commitment to improving the management of
federal programs, both through its leadership in reviewing agency
program performance using the PART tool as well as through the PMA. As
part of the President's budget preparation, PART clearly must serve the
President's interests. However, it is not well suited to addressing
crosscutting (or horizontal) issues or to looking at broad program
areas in which several programs address a common goal. GPRA was
designed to address the needs of many users of performance information,
including (1) Congress to provide oversight and inform funding
decisions, (2) agency managers to manage programs and make internal
resource decisions, and (3) the public to provide greater
accountability. It is not yet clear the extent to which PART
performance goals and measures will compete with agencies' long-term,
strategic GPRA goals and objectives that were established in
consultation with Congress and other stakeholders.
We also found that, while the quality of agency plans and reports have
improved overall since their initial efforts, they continue to suffer
from certain persistent weaknesses as noted above. However, OMB's July
2003 guidance for preparation and submission of annual performance
plans is significantly shorter and less detailed than its 2002
guidance. Consistent, more explicit OMB guidance on preparing GPRA
documents can help ensure that gains in the quality of GPRA documents
are maintained and provide a resource for agency managers to make
further improvements in those documents.
We also found that timing issues may affect the development of agency
strategic plans that are meaningful and useful to top leadership. The
commitment and sustained attention of top leadership within agencies,
OMB, and Congress is critical to the success of strategic planning
efforts. A strategic plan should reflect the policy priorities of an
organization's leaders and the input of key stakeholders if it is to be
an effective management tool. However, GPRA specifies time frames for
updating strategic plans that do not correspond to presidential or
congressional terms. As a result, an agency may be required to update
its strategic plan a year before a presidential election and without
input from a new Congress. If a new president is elected, the updated
plan is essentially moot and agencies must spend additional time and
effort revising it to reflect new priorities. Our focus group
participants, including GPRA experts, strongly agreed that this timing
issue should be addressed by adjusting time frames to correspond better
with presidential and congressional terms.
The benefit of collecting performance information is only fully
realized when this information is actually used by managers to bring
about desired results. However, federal managers reported mixed results
in the use of performance information. Focus group participants and
survey respondents noted that although many federal managers understand
and use results-oriented management concepts in their day-to-day
activities, such as strategic planning and performance measurement,
they do not always connect these concepts to the requirements of GPRA.
According to our 2003 survey results, the reported use of performance
information to a great or very great extent for nine management
activities, such as setting program priorities or setting individual
job expectations for staff, ranging from 41 to 66 percent, has not
changed significantly since our first survey in 1997. One exception was
the reported use to a great or very great extent of performance
information to adopt new program approaches or change work processes,
which was significantly lower than the 1997 results. GPRA's usefulness
to agency leaders and managers as a tool for management and
accountability was cited as a key accomplishment numerous times by
focus group participants. However, a number of alternative views
indicated that the usefulness of GPRA as a management tool has been
limited. Our survey data also indicate that managers' perceive their
participation in activities related to the development and use of
performance information has been limited.
Federal managers continue to confront a range of important human
capital management challenges. These managers report that they are held
accountable for program results, but may not have the decision-making
authority they need to accomplish agency goals. Moreover, fewer than
half of managers reported receiving relevant training. Managers also
perceive a lack of positive recognition for helping agencies achieve
results. Unfortunately, most existing federal performance appraisal
systems are not designed to support a meaningful performance-based pay
system in that they fail to link institutional, program, unit, and
individual performance measurement and reward systems. In our view, one
key need is to modernize performance management systems in executive
agencies so that they link to the agency's strategic plan, related
goals, and desired outcomes and are therefore capable of adequately
supporting more performance-based pay and other personnel decisions.
Managers reported persistent challenges in setting outcome-oriented
goals, measuring performance, and collecting useful data. Focus group
participants and survey respondents noted that outcome-oriented
performance measures were especially difficult to establish when the
program or line of effort was not easily quantifiable. In some
agencies, particularly those that have a research and development
component, managers reported difficulties in establishing meaningful
outcome measures. Managers also identified difficulties in
distinguishing between the results produced by the federal program and
results caused by external factors or nonfederal actors, such as with
grant programs. Finally, managers reported that timely and useful
performance information is not always available.
Crosscutting issues continue to be a challenge to GPRA implementation.
Our review of six agencies' strategic and annual performance plans
showed some improvement in addressing their crosscutting program
efforts, but a great deal of improvement is still necessary. We have
previously reported and testified that GPRA could provide OMB,
agencies, and Congress with a structured framework for addressing
crosscutting policy initiatives and program efforts. OMB could use the
provision of GPRA that calls for OMB to develop a governmentwide
performance plan to integrate expected agency-level performance. It
could also be used to more clearly relate and address the contributions
of alternative federal strategies. Unfortunately, this provision has
not been fully implemented. Instead, OMB has used the President's
Budget to present high-level information about agencies and certain
program performance issues. The current agency-by-agency focus of the
budget does not provide the integrated perspective of government
performance envisioned by GPRA. For example, the fiscal year 2004
budget identified budget requests and performance objectives by agency,
such as the U.S. Department of Defense, as opposed to crosscutting
governmentwide themes. From this presentation, one could assume that
the only activities the U.S. government planned to carry out in support
of national defense were those listed under the chapter "Department of
Defense." However, the chapter on the fiscal year 2004 budget
discussing "the Department of State and International Assistance
Programs," contains a heading titled, "Countering the Threat from
Weapons of Mass Destruction." And while OMB may have a technical reason
for not classifying this task as being related to national defense or
homeland security, it is unclear that a lay reader could make that
distinction. The fiscal year 2005 budget also identified budget
requests and performance objectives by agency, not by crosscutting
theme.
A strategic plan for the federal government could provide an additional
tool for governmentwide reexamination of existing programs, as well as
proposals for new programs. If fully developed, a governmentwide
strategic plan could potentially provide a cohesive perspective on the
long-term goals of the federal government and provide a much needed
basis for fully integrating, rather than merely coordinating, a wide
array of federal activities. Successful strategic planning requires the
involvement of key stakeholders. Thus, it could serve as a mechanism
for building consensus. Further, it could provide a vehicle for the
President to articulate long-term goals and a road map for achieving
them. In addition, a strategic plan could provide a more comprehensive
framework for considering organizational changes and making resource
decisions. The development of a set of key national indicators could be
used as a basis to inform the development of governmentwide strategic
and annual performance plans. The indicators could also link to and
provide information to support outcome-oriented goals and objectives in
agency-level strategic and annual performance plans.
Focus group members believed that one of the main challenges to GPRA
implementation was the reluctance of Congress to use performance
information when making decisions, especially appropriations
decisions. However, less than one quarter of federal managers in the
2003 survey shared that concern. Further, a recent Congressional
Research Service review suggests that Congress uses performance
information to some extent, as evidenced by citations in legislation
and committee reports. While there is concern regarding Congress' use
of performance information, it is important to make sure that this
information is useful. In other words, the information presented and
its presentation must meet the needs of the user. Regular consultation
with Congress about both the content and format of performance plans
and reports is critical.
As a key user of performance information, Congress also needs to be
considered a partner in shaping agency goals at the outset. GPRA
provides a vehicle for Congress to explicitly state its performance
expectations in outcome-oriented terms when consulting with agencies on
their strategic plans or when establishing new programs or exercising
oversight of existing programs that are not achieving desired results.
This would provide important guidance to agencies that could then be
incorporated in agency strategic and annual performance plans.
Recommendations for Executive Action:
GAO recommends that the Director of OMB implement five suggestions to
improve its guidance and oversight of GPRA implementation.
To provide a broader perspective and more cohesive picture of the
federal government's goals and strategies to address issues that cut
across executive branch agencies, we recommend that the Director of OMB
fully implement GPRA's requirement to develop a governmentwide
performance plan.
To achieve the greatest benefit from both GPRA and PART, we recommend
that the Director of OMB articulate and implement an integrated and
complementary relationship between the two. GPRA is a broad legislative
framework that was designed to be consultative with Congress and other
stakeholders, and allows for varying uses of performance information.
PART looks through a particular lens for a particular use--the
executive budget formulation process.
To improve the quality of agencies' strategic plans, annual performance
plans, and performance reports and help agencies meet the requirements
of GPRA, we recommend that the Director of OMB provide clearer and
consistent guidance to executive branch agencies on how to implement
GPRA. Such guidance should include standards for communicating key
performance information in concise as well as longer formats to better
meet the needs of external users who lack the time or expertise to
analyze lengthy, detailed documents.
To help address agencies' performance measurement challenges, we
recommend the Director of OMB engage in a continuing dialogue with
agencies about their performance measurement practices with a
particular focus on grant-making, research and development, and
regulatory functions to identify and replicate successful approaches
agencies are using to measure and report on their outcomes, including
the use of program evaluation tools. Additionally, we recommend that
the Director of OMB work with executive branch agencies to identify the
barriers to obtaining timely data to show progress against performance
goals and the best ways to report information where there are
unavoidable lags in data availability. Governmentwide councils, such as
the President's Management Council and the Chief Financial Officers
Council, may be effective vehicles for working on these issues.
To facilitate the transformation of agencies' management cultures to be
more results oriented, we recommend that the Director of OMB work with
agencies to ensure they are making adequate investments in training on
performance planning and measurement, with a particular emphasis on how
to use performance information to improve program performance.
Matters for Congressional Consideration:
GAO also identified two matters for congressional consideration to
improve the governmentwide focus on results.
To ensure that agency strategic plans more closely align with changes
in the federal government leadership, Congress should consider amending
GPRA to require that updates to agency strategic plans be submitted at
least once every 4 years, 12-18 months after a new administration
begins its term. Additionally, consultations with congressional
stakeholders should be held at least once every new Congress and
interim updates made to strategic and performance plans as warranted.
Congress should consider using these consultations along with its
traditional oversight role and legislation as opportunities to clarify
its performance expectations for agencies. This process may provide an
opportunity for Congress to develop a more structured oversight agenda.
To provide a framework to identify long-term goals and strategies to
address issues that cut across federal agencies, Congress should
consider amending GPRA to require the President to develop a
governmentwide strategic plan.
Agency Comments:
We provided a copy of the draft report to OMB for comment. OMB's
written comments are reprinted in appendix VIII. In general, OMB agreed
with our findings and conclusions. OMB agreed to implement most of our
recommendations, noting that these recommendations will enhance their
efforts to make the government more results oriented. OMB agreed to (1)
work with agencies to ensure they are provided adequate training in
performance management, (2) revise its guidance to clarify the
integrated and complementary relationship between GPRA and PART, and
(3) continue to use PART to improve agency performance measurement
practices and share those practices across government.
In response to our recommendation that OMB fully implement GPRA's
requirement to develop a governmentwide performance plan, OMB stated
that the President's Budget represents the executive branch's
governmentwide performance plan. However, according to GAO's review,
the agency-by-agency focus of the budget over the past few years does
not provide an integrated perspective of government performance, and
thus does not meet GPRA's requirement to provide a "single cohesive
picture of the annual performance goals for the fiscal year." To
clarify this point, we added an example that illustrates the lack of
integration between crosscutting issues in the budget.
In response to our matter for congressional consideration that Congress
should consider amending GPRA to require the President to develop a
governmentwide strategic plan, OMB noted that the budget serves as the
governmentwide strategic plan. However, the President's Budget focuses
on establishing agency budgets for the upcoming fiscal year. Unlike a
strategic plan, it provides neither a long-term nor an integrated
perspective on the federal government's activities. A governmentwide
strategic plan should provide a cohesive perspective on the long-term
goals of the federal government and provide a basis for fully
integrating, rather than primarily coordinating, a wide array of
existing and relatively short-term federal activities.
We provided relevant sections of the draft report to Education, DOE,
HUD, SBA, SSA, and DOT. Education and SBA did not provide any comments,
while DOT provided minor technical comments. DOE, HUD, and SSA
disagreed with some of our observations on their strategic plans,
performance plans, and performance reports; we changed or clarified
relevant sections of the report, as appropriate. Written comments from
DOE, HUD, and SSA are reprinted in appendixes IX, X, and XI,
respectively, along with our responses.
[End of section]
Chapter 1: Introduction:
From defending the homeland against terrorists, to preventing the
spread of infectious diseases, to providing a reliable stream of social
security income to retirees and supporting the transition from welfare
to work, the federal government provides funding and services to the
American public that can affect their lives in critical ways every day.
However, the federal government is in a period of profound transition
and faces an array of challenges and opportunities to enhance
performance, ensure accountability, and position the nation for the
future. A number of overarching trends, such as diffuse security
threats and homeland security needs, increasing global interdependency,
the shift to knowledge-based economies, and the looming fiscal
challenges facing our nation, drive the need to reconsider the proper
role for the federal government in the 21st century, how the government
should do business (including how it should be structured), and in some
instances, who should do the government's business.
Without effective short-and long-term planning, which takes into
account the changing environment and needs of the American public and
the challenges they face and establishes goals to be achieved, federal
agencies risk delivering programs and services that may or may not meet
society's most critical needs. At a cost to taxpayers of over $2
trillion annually, the federal government should be able to demonstrate
to the American public that it can anticipate emerging issues, develop
sound strategies and plans to address them, and be accountable for the
results that have been achieved.
Concerned that the federal government was more focused on program
activities and processes than the results to be achieved, Congress
passed the Government Performance and Results Act of 1993
(GPRA).[Footnote 9] The act required federal agencies to develop
strategic plans with long-term strategic goals, annual goals linked to
achieving the long-term goals, and annual reports on the results
achieved. Now that GPRA has been in effect for 10 years, you asked us
to assess the effectiveness of GPRA in creating a focus on results in
the federal government. Specifically, this report discusses (1) the
effect of GPRA over the last 10 years in creating a governmentwide
focus on results and the government's ability to deliver results to the
American public, including an assessment of the changes in the overall
quality of agencies' strategic plans, annual performance plans, and
annual performance reports; (2) the challenges agencies face in
measuring performance and using performance information in management
decisions; and (3) how the federal government can continue to shift
toward a more results-oriented focus.
Impact of Emerging Trends and Fiscal Challenges:
With the 21st century challenges we are facing, it is more vital than
ever to maximize the performance of federal agencies in achieving their
long-term goals. The federal government must address and adapt to major
trends in our country and around the world. At the same time, our
nation faces serious long-term fiscal challenges. Increased pressure
also comes from world events: both from the recognition that we cannot
consider ourselves "safe" between two oceans--which has increased
demands for spending on homeland security--and from the United States
(U.S.) role in combating terrorism in an increasingly interdependent
world. To be able to assess federal agency performance and hold agency
managers accountable for achieving their long-term goals, we need to
know what the level of performance is. GPRA planning and reporting
requirements can provide this essential information.
Our country's transition into the 21st century is characterized by a
number of key trends, including:
* the national and global response to terrorism and other threats to
our personal and national security;
* the increasing interdependence of enterprises, economies, markets,
civil societies, and national governments, commonly referred to as
globalization;
* the shift to market-oriented, knowledge-based economies;
* an aging and more diverse U.S. population;
* rapid advances in science and technology and the opportunities and
challenges created by these changes;
* challenges and opportunities to maintain and improve the quality of
life for the nation, communities, families, and individuals; and:
* the changing and increasingly diverse nature of governance structures
and tools.
As the nation and government policymakers grapple with the challenges
presented by these evolving trends, they do so in the context of
rapidly building fiscal pressures. GAO's long-range budget simulations
show that this nation faces a large and growing structural deficit due
primarily to known demographic trends and rising health care costs. The
fiscal pressures created by the retirement of the baby boom generation
and rising health costs threaten to overwhelm the nation's fiscal
future. As figure 1 shows, by 2040, absent reform or other major tax or
spending policy changes, projected federal revenues will likely be
insufficient to pay more than interest on publicly held debt. Further,
our recent shift from surpluses to deficits means the nation is moving
into the future in a more constrained fiscal position.
Figure 1: Composition of Spending as a Share of GDP Assuming
Discretionary Spending Grows with GDP after 2003 and All Expiring Tax
Provisions Are Extended:
[See PDF for image]
Notes: Although all expiring tax cuts are extended, revenue as a share
of gross domestic product (GDP) increases through 2013 due to (1) real
bracket creep, (2) more taxpayers becoming subject to the Alternative
Minimum Tax, and (3) increased revenue from tax-deferred retirement
accounts. After 2013, revenue as a share of GDP is held constant. This
simulation assumes that currently scheduled Social Security benefits
are paid in full throughout the simulation period.
[End of figure]
The United States has had a long-range budget deficit problem for a
number of years, even during recent years in which we had significant
annual budget surpluses. Unfortunately, the days of surpluses are gone,
and our current and projected budget situation has worsened
significantly. The bottom line is that our projected budget deficits
are not manageable without significant changes in "status quo"
programs, policies, processes, and operations.
GPRA Background:
GPRA is the centerpiece of a statutory framework that Congress put in
place during the 1990s to help resolve the long-standing management
problems that have undermined the federal government's efficiency and
effectiveness and to provide greater accountability for results. In
addition to GPRA, the framework comprises the Chief Financial Officers
Act of 1990, as amended by the Government Management Reform Act of
1994, and information technology reform legislation, including the
Paperwork Reduction Act of 1995 and the Clinger-Cohen Act of 1996.
Together, these laws provide a powerful framework for developing and
integrating information about agencies' missions and strategic
priorities, the results-oriented performance goals that flow from those
priorities, performance data to show the level of achievement of those
goals, and the relationship of reliable and audited financial
information and information technology investments to the achievement
of those goals.
GPRA was intended to address several broad purposes, including
strengthening the confidence of the American people in their
government; improving federal program effectiveness, accountability,
and service delivery; and enhancing congressional decision making by
providing more objective information on program performance.
The basic requirements of GPRA for the preparation of strategic plans,
annual performance plans, and annual program performance reports by
executive branch agencies are the following:
1. The agency‘s strategic plan must contain these six key elements:
* a comprehensive agency mission statement;
* agencywide long-term goals and objectives for all major functions and
operations;
* approaches (or strategies) and the various resources needed to
achieve the goals and objectives;
* a description of the relationship between the long-term goals and
objectives and the annual performance goals;
* an identification of key factors, external to the agency and beyond
its control, that could significantly affect the achievement of the
strategic goals; and
* a description of how program evaluations were used to establish or
revise strategic goals and a schedule for future program evaluations.
2. The agency must develop annual performance plans covering each
program activity set forth in the agencies‘ budgets:*
Building on the decisions made as part of the strategic planning
process, GPRA requires executive agencies to develop annual performance
plans covering each program activity set forth in the agencies‘
budgets. Annual performance plans, covering the upcoming fiscal year,
are to be submitted to Congress after the President‘s Budget is
submitted, which generally occurs in February. Each plan is to contain
an agency‘s annual performance goals and associated measures, which the
agency is to use in order to gauge its progress toward accomplishing
its strategic goals. OMB is to use the agencies‘ performance plans to
develop an overall federal government performance plan that is to be
submitted with the President‘s Budget. The performance plan for the
federal government is to present to Congress a single cohesive picture
of the federal government‘s annual performance goals for a given fiscal
year.
3. The agency must prepare annual reports on program performance for
the previous fiscal year, to be issued by March 31 each year:
GPRA requires executive agencies to prepare annual reports on program
performance for the previous fiscal year, to be issued by March 31 each
year. In each report, an agency is to compare its performance against
its goals, summarize the findings of program evaluations completed
during the year, and describe the actions needed to address any unmet
goals. Recent OMB guidance states that executive agencies must combine
their program performance report with their accountability report and
transmit the combined report for fiscal year 2003 by January 30, 2004,
and the combined report for fiscal year 2004 by November 15, 2004.
*Program activity refers to the lists of projects and activities in the
appendix portion of the Budget of the United States Government. Program
activity structures are intended to provide a meaningful representation
of the operations financed by a specific budget account.
The Office of Management and Budget (OMB) plays an important role in
the management of the federal government's performance, and
specifically GPRA implementation. Part of OMB's overall mission is to
ensure that agency plans and reports are consistent with the
President's Budget and administration policies. OMB is responsible for
receiving and reviewing agencies' strategic plans, annual performance
plans, and annual performance reports. To improve the quality and
consistency of these documents, OMB issues annual guidance to agencies
for their preparation, including guidelines on format, required
elements, and submission deadlines.[Footnote 10] GPRA requires OMB to
prepare the overall governmentwide performance plan, based on agencies'
annual performance plan submissions. OMB also played an important role
in the pilot phase of GPRA implementation by designating agencies for
pilot projects in performance measurement, managerial accountability
and flexibility, and performance budgeting, and assessing the results
of the pilots. Finally, GPRA provides OMB with authority to grant
agencies waivers to certain administrative procedures and controls.
Recent OMB guidance[Footnote 11] requires agencies to submit
"performance budgets" in lieu of annual performance plans for their
budget submission to OMB and Congress. Performance budgets are to meet
all the statutory requirements of GPRA for annual performance plans. In
addition, agencies are to include all performance goals used in the
assessment of program performance done under OMB's Program Assessment
Rating Tool (PART) process.[Footnote 12] Moreover, the guidance states
that until all programs have been assessed by PART, the performance
budget will also for a time include performance goals for agency
programs that have not yet been assessed using PART. The expectation is
that agencies are to substitute new or revised performance goals
resulting from OMB's review for goals it deemed unacceptable.
In crafting GPRA, Congress recognized that managerial accountability
for results is linked to managers having sufficient flexibility,
discretion, and authority to accomplish desired results. GPRA
authorizes agencies to apply for managerial flexibility waivers in
their annual performance plans beginning with fiscal year 1999. The
authority of agencies to request waivers of administrative procedural
requirements and controls is intended to provide federal managers with
more flexibility to structure agency systems to better support program
goals. The nonstatutory requirements that OMB can waive under GPRA
generally involve the allocation and use of resources, such as
restrictions on shifting funds among items within a budget account.
Agencies must report in their annual performance reports on the use and
effectiveness of any GPRA managerial flexibility waivers that they
receive.
OMB was to designate at least five agencies from the first set of pilot
projects to test managerial accountability and flexibility during
fiscal years 1995 and 1996. We previously reported on the results of
the pilot project to implement managerial flexibility waivers and found
that the pilot did not work as intended.[Footnote 13] OMB did not
designate any of the seven departments and one independent agency that
submitted a total of 61 waiver proposals as GPRA managerial
accountability and flexibility pilots. For about three-quarters of the
waiver proposals, OMB or other central management agencies determined
that the waivers were not allowable for statutory or other reasons or
that the requirement for which the waivers were proposed no longer
existed. For the remaining proposals, OMB or other central management
agencies approved waivers or developed compromises by using authorities
that were already available independent of GPRA.
Under GPRA, another set of pilot projects, which were scheduled for
fiscal years 1998 and 1999, were to test performance budgeting--i.e.,
the presentation of the varying levels of performance that would result
from different budget levels. We previously reported that OMB initially
deferred these pilots--originally to be designated in fiscal years 1998
and 1999--to give federal agencies time to develop the capability of
calculating the effects of marginal changes in cost or funding on
performance.[Footnote 14] When the pilots began in August 1999, OMB
designed them as case studies prepared by OMB staff to demonstrate how
performance information could be used to compare alternatives and to
develop funding recommendations for incorporation into the President's
fiscal year 2001 budget submission.
On January 18, 2001, OMB reported the results of five performance
budgeting pilots that explored agencies' capabilities of more formally
assessing the effects of different funding levels on performance goals.
OMB selected the pilots to reflect a cross section of federal functions
and capabilities so that a representative range of measurement and
reporting issues could be explored. In its report, OMB concluded that
legislative changes were not needed. OMB reported that the pilots
demonstrated that assuring further performance measurement
improvements and steadily expanding the scope and quality of
performance measures is paramount, and that the existing statute
provides sufficient latitude for such improvement.
Overall, OMB concluded that the pilots raised several key challenges
about performance budgeting at the federal level including, for
example, the following:
* In many instances, measuring the effects of marginal, annual budget
changes on performance is not precise or meaningful.
* While continuing to change from an almost total reliance on output
measures to outcome measures, it will be much more difficult to
associate specific resource levels with those outcomes, particularly
over short periods of time.
* Establishing clear linkages between funding and outcomes will vary by
the nature of the program and the number of external factors.
* Delays in the availability of performance data, sometimes caused by
agencies' reliance on nonfederal program partners for data collection,
will continue to present synchronization problems during budget
formulation.
Scope and Methodology:
To meet the three objectives stated earlier, we reviewed our extensive
prior work on GPRA best practices and implementation and collected
governmentwide data to assess the government's overall focus on
results. We conducted a random, stratified, governmentwide survey of
federal managers comparable to surveys we conducted in 1997 and 2000.
We also held eight in-depth focus groups--seven comprised of federal
managers from 23 federal agencies and one with GPRA experts. We also
interviewed top appointed officials from the current and previous
administrations. Finally, we judgmentally selected a sample of six
agencies to review for changes in the quality of their strategic plans,
performance plans, and performance reports since their initial efforts.
The agencies we selected were the Departments of Education (Education),
Energy (DOE), Housing and Urban Development (HUD), and Transportation
(DOT) and the Small Business (SBA) and Social Security Administrations
(SSA). In making this selection, we chose agencies that collectively
represented the full range of characteristics in the following four
areas: (1) agency size (small, medium, large); (2) primary program
types (direct service, research, regulatory, transfer payments, and
contracts or grants); (3) quality of fiscal year 2000 performance plan
based on our previous review (low, medium, high);[Footnote 15] and (4)
type of agency (cabinet department and independent agency). Appendix I
contains a more detailed discussion of our scope and methodology.
We performed our work in Washington, D.C., from January through
November 2003 in accordance with generally accepted government auditing
standards. Major contributors to this report are listed in appendix
XII.
[End of section]
Chapter 2: GPRA Established the Foundation for a More Results-Oriented
Federal Government:
Among the purposes of GPRA cited by Congress was to improve federal
program effectiveness and service delivery by promoting a new focus on
results, service quality, and customer satisfaction by setting program
goals, measuring performance against goals, and reporting publicly on
progress. Furthermore, GPRA was to improve congressional decision
making by providing better information on achieving objectives, and on
the relative effectiveness and efficiency of federal programs and
spending. Ten years after enactment, GPRA's requirements have laid a
foundation of results-oriented agency planning, measurement, and
reporting that have begun to address these purposes. Focus group
participants, high-level political appointees, and OMB officials we
interviewed cited positive effects of GPRA that they generally
attributed to GPRA's statutory requirements for planning and reporting.
Our survey results indicate that since GPRA went into effect
governmentwide in 1997, federal managers reported having significantly
more of the types of performance measures called for by GPRA--
particularly outcome-oriented performance measures. GPRA has also begun
to facilitate the linking of resources to results, although much
remains to be done in this area.
GPRA Statutory Requirements Laid a Foundation for Agencywide Results-
Oriented Management:
Prior to enactment of GPRA, our 1992 review of the collection and use
of performance data by federal agencies revealed that, although many
agencies collected performance information at the program level, few
agencies had results-oriented performance information to manage or make
strategic policy decisions for the agency as a whole.[Footnote 16]
Federal agencies surveyed indicated that many had a single, long-term
plan that contained goals, standards, or objectives for the entire
agency or program. Many of these agencies also reported they collected
a wide variety of performance measures. However, in validating the
survey responses with a sample of agencies, we found that measures were
typically generated and used by program-level units within an agency
and focused on measuring work activity levels and outputs or compliance
with statutes. Little of this performance information was transparent
to Congress, OMB, or the public and few of the agencies we visited used
performance measures to manage toward long-term objectives. Few of the
agencies surveyed had the infrastructure in place, such as a unified
strategic plan with measurable goals, an office that collected
performance measures, and regular consolidated reports, to tie plans
and measures.
GPRA addressed these shortcomings by creating a comprehensive and
consistent statutory foundation of required agencywide strategic plans,
annual performance plans, and annual performance reports. In contrast
to prior federal government efforts to measure performance, GPRA
explicitly emphasized that, in addition to performance indicators that
agencies may need to manage programs on a day-to-day basis, such as
quantity, quality, timeliness, and cost, agencies also needed outcome-
oriented goals and measures that assess the actual results, effects, or
impact of a program or activity compared to its intended purpose.
Expert and agency focus group participants cited the creation of this
statutory foundation as one of the key accomplishments of GPRA.
Participants agreed that GPRA created a framework in statute for
federal agencies to plan their activities in order to become more
results oriented and provided a managerial tool for program
accountability. Using this framework, agencies could develop and focus
on strategies to carry out the programs they administer; set goals and
identify performance indicators that will inform them whether or not
they achieved the performance they expected; and determine what impact,
if any, their programs have had on the American public. According to
the experts in one of our focus groups, comparing federal agencies'
current mission statements contained in their strategic plans to what
they were in the past demonstrates that the agencies have done some
"soul searching" to get a better sense of what their role is (or should
be) and how they can achieve it. Given that GPRA is in statute,
participants indicated that the use of this planning framework is
likely to be sustained within agencies.
One of the premises of GPRA is that both congressional and executive
branch oversight of federal agency performance were seriously hampered
by a lack of adequate results-oriented goals and performance
information. As noted above, prior to the enactment of GPRA few
agencies reported their performance information externally. OMB
officials we interviewed as part of our current review suggested that
OMB has been a key consumer of the performance information produced
under GPRA and that it has provided a foundation for their efforts to
oversee agency performance.
For example, during the development of the fiscal year 2004 budget, OMB
used PART to review and rate 234 federal programs. We recently reported
that one of PART's major impacts was its ability to highlight OMB's
recommended changes in program management and design.[Footnote 17] PART
reviews look at four elements--program purpose and design, strategic
planning, program management, and program results/accountability--and
rate the program on how well each of these elements is executed.
However, without the foundation of missions, goals, strategies,
performance measures, and performance information generated under GPRA,
such oversight would be difficult to carry out.
Participants in most of our focus groups also agreed that GPRA has been
a driving force behind many cultural changes that have occurred within
federal agencies. Highlighting the focus on results, participants
stated that GPRA had stimulated a problem-solving approach within
federal agencies and encouraged agency managers to think creatively
when developing performance indicators for their programs. GPRA has
also changed the dialogue within federal agencies; front-line managers
and staff at lower levels of the organization now discuss budget issues
in connection with performance. Similarly, experts noted that
information about performance management and resource investments are
more frequently communicated between agency officials and Congress than
in the past. Within agencies, GPRA documents can provide a context of
missions, goals, and strategies that political appointees can use to
articulate agencies' priorities.
Views on GPRA's Effect on the Federal Government's Ability to Deliver
Results to the American Public Were Mixed:
A key purpose of GPRA was "to improve the confidence of the American
people in the capability of the Federal Government, by systematically
holding Federal agencies accountable for achieving program results."
When asked about the direct effects of GPRA on the public in our 2003
survey, an estimated 23 percent of federal managers agreed to a
moderate or greater extent that GPRA improved their agency's ability to
deliver results to the American public; a larger percentage--38
percent--chose a "no basis to judge/not applicable" category.
When a similar question was posed in our focus groups with experts and
federal managers, participants' views were generally mixed. Some
federal managers in our focus groups agreed that GPRA has had a
positive effect on raising awareness on many performance issues, and
that in and of itself is a way of delivering results. The information
gathered and reported for GPRA allows agencies to make better-informed
decisions, which improves their ability to achieve results. Other
participants stated that while certain aspects of GPRA-related work
have been positive, agencies' ability to deliver results and public
awareness of their activities cannot always be exclusively attributed
to GPRA. For example, some participants stated that many agencies rely
on grant recipients to carry out their work, and delivering results to
the American public depends, to a large extent, on the diligence of
these organizations to implement their programs; such results would not
change dramatically if GPRA were no longer a requirement.
A number of the political appointees we interviewed cited examples of
outcomes they believe would not have occurred without the structure of
GPRA. For example, a former deputy secretary of the Department of
Veterans Affairs (VA) stated that "the Results Act brought about a
fundamental rethinking of how we managed our programs and processes. .
. . We developed a strategic plan that was veteran-focused. . . . We
made every effort to define program successes from the veteran's
perspective." A former Chief Financial Officer (CFO) cited Customs
Service goals to reduce the quantity of illegal drugs flowing into the
United States and the Food and Drug Administration's focus on speeding
up the approval of new drugs as examples of outcomes that can make a
big difference in people's lives.
Another major accomplishment of GPRA cited by our focus group
participants is that GPRA improved the transparency of government
results to the American public. As noted above, prior to GPRA, few
agencies reported performance results outside of their agencies. Focus
group participants indicated a key accomplishment of GPRA was its value
as a communication tool by increasing the transparency to the public of
what their agencies did in terms the public could understand. For
example, information on agencies' strategic plans, performance goals,
measures, and results are easily obtainable from agency Web sites. One
focus group participant commented that GPRA helps bureaucrats explain
to nonbureaucrats what the federal government does in terms they can
better understand. Other comments indicated that because of GPRA
agencies could now tell Congress and the American public what they are
getting for their money.
More Managers Reported Having Performance Measures:
A fundamental element in an organization's efforts to manage for
results is its ability to set meaningful goals for performance and to
measure performance against those goals. From our 2003 survey we
estimate that 89 percent of federal managers overall said there were
performance measures for the programs they were involved with. This is
a statistically significantly higher percentage than the 76 percent of
managers who answered yes to this item on our 1997 survey. (See fig.
2.):
Figure 2: Percentage of Federal Managers Who Reported That There Were
Performance Measures for the Programs with Which They Were Involved:
[See PDF for image]
[A] There was a statistically significant difference between 1997 and
2003 surveys.
[End of figure]
Moreover, when we asked managers who said they had performance measures
which of five types of measures they had to a great or very great
extent, they reported increases in all five types of measures between
1997 and 2003,[Footnote 18] all of which were statistically
significant. (See fig. 3.) Notably, managers indicated the existence of
outcome measures, defined as "performance measures that demonstrate to
someone outside the organization whether or not intended results are
being achieved," grew from a low of 32 percent in 1997 to the current
estimate of 55 percent, a level that is on par with output measures for
the first time since we began our survey.
Figure 3: Percentage of Federal Managers Who Reported Having Specific
Types of Performance Measures to a Great or Very Great Extent:
[See PDF for image]
[A] There was a statistically significant difference between 1997 and
2003 surveys.
[End of figure]
Similarly, focus group participants commented on certain cultural
changes that had taken place within their agencies since the passage of
GPRA in which the "vocabulary" of performance planning and measurement-
-e.g., a greater focus on performance management; orientation toward
outcomes over inputs and outputs; and an increased focus on program
evaluation--had become more pervasive. This perception is partly born
out by our survey results. Since 1997 those reporting a moderate to
extensive knowledge of GPRA and its requirements shifted significantly
from 26 percent to 41 percent in 2003, while those reporting no
knowledge of GPRA declined significantly from 27 percent to 20 percent.
(See fig. 4.):
Figure 4: Percentage of Federal Managers Who Reported Their Awareness
of GPRA:
[See PDF for image]
[A] There was a statistically significant difference between 1997 and
2003 surveys.
[End of figure]
Consistent with our survey results indicating increases in results-
oriented performance measures and increasing GPRA knowledge, we also
observed a significant decline in the percentage of federal managers
who agreed that certain factors hindered measuring performance or using
the performance information. For example, as shown in figure 5, of
those who expressed an opinion, the percentage of managers who noted
that determining meaningful measures was a hindrance to a great or very
great extent was down significantly from 47 percent in 1997 to 36
percent in 2003. Likewise, the percentage that agreed to a great or
very great extent that different parties are using different
definitions to measure performance was a hindrance also declined
significantly from 49 percent in 1997 to 36 percent in 2003.
Figure 5: Percentage of Federal Managers Who Reported Hindrances to
Measuring Performance or Using the Performance Information to a Great
or Very Great Extent:
[See PDF for image]
Note: Percentages are based on those respondents answering on the
extent scale.
[A] There was a statistically significant difference between 1997 and
2003.
[End of figure]
Finally, our survey data suggested that more federal managers,
especially at the Senior Executive Service (SES) level, believed that
OMB was paying attention to their agencies' efforts under GPRA.
Moreover, there was no corresponding increase in their concern that OMB
would micromanage the programs in their agencies. In our survey, we
asked respondents to assess the extent to which OMB pays attention to
their agencies' efforts under GPRA. As seen in figure 6, in 2003, the
percentage of respondents who responded "Great" or "Very Great" to this
question (31 percent) was significantly higher than in 2000 (22
percent). Of those, SES respondents showed an even more dramatic
increase, from 33 to 51 percent. We also asked respondents the extent
to which their concern that OMB would micromanage programs in their
agencies was a hindrance to measuring performance or using performance
information. The percentage among those expressing an opinion that it
was a hindrance to a great or very great extent was low--around 24
percent in 2003--with no significant difference between 2000 and 2003.
Figure 6: Percentage of Federal Managers and SES Managers Who Reported
That OMB Paid Attention to Their Agency's Efforts under GPRA to a Great
or Very Great Extent:
[See PDF for image]
[A] There was a statistically significant difference between 2000 and
2003 surveys.
[End of figure]
GPRA Has Begun to Establish a Link between Resources and Results:
Among its major purposes, GPRA aims for a closer and clearer linkage
between requested resources and expected results. The general concept
of linking performance information with budget requests is commonly
known as performance budgeting. Budgeting is and will remain an
exercise in political choice, in which performance can be one, but not
necessarily the only, factor underlying decisions. However, efforts to
infuse performance information into resource allocation decisions can
more explicitly inform budget discussions and focus them--both in
Congress and in agencies--on expected results, rather than on inputs.
GPRA established a basic foundation for performance budgeting by
requiring that an agency's annual performance plan cover each program
activity in the President's budget request for that agency. GPRA does
not specify any level of detail or required components needed to
achieve this coverage. Further, GPRA recognizes that agencies' program
activity structures are often inconsistent across budget accounts and
thus gives agencies the flexibility to consolidate, aggregate, or
disaggregate program activities, so long as no major function or
operation of the agency is omitted or minimized. In addition, OMB
guidance has traditionally required agencies to display, by budget
program activity, the funding level being applied to achieve
performance goals. OMB's guidance on developing fiscal year 2005
performance budgets also encourages a greater link between performance
and funding levels, however, it places greater emphasis on linking
agencies' long-term and annual performance goals to individual
programs. At a minimum, agencies are to align resources at the program
level, but they are encouraged to align resources at the performance
goal level. Resources requested for each program are to be the amounts
needed to achieve program performance goal targets.
Our 1998 assessment of fiscal year 1999 performance plans found that
agencies generally covered the program activities in their budgets, but
most plans did not identify how the funding for those program
activities would be allocated to performance goals.[Footnote 19]
However, our subsequent reviews of performance plans indicate that
agencies have made progress in demonstrating how their performance
goals and objectives relate to program activities in the budget.
Over the first 4 years of agency efforts to implement GPRA, we observed
that agencies continued to tighten the required link between their
performance plans and budget requests.[Footnote 20] Of the agencies we
reviewed over this period, all but three met the basic requirement of
GPRA to define a link between their performance plans and the program
activities in their budget requests, and most of the agencies in our
review had moved beyond this basic requirement to indicate some level
of funding associated with expected performance described in the plan.
Most importantly, more of the agencies we reviewed each year--almost 75
percent in fiscal year 2002 compared to 40 percent in fiscal year 1999-
-were able to show a direct link between expected performance and
requested program activity funding levels--the first step in defining
the performance consequences of budgetary decisions. However, we have
also observed that the nature of these linkages varied considerably.
Most of the agencies in our review of fiscal year 2002 performance
plans associated funding requests with higher, more general levels of
expected performance, rather than the more detailed "performance goals
or sets of performance goals" suggested in OMB guidance.
Although not cited by our group of experts, participants at six of our
seven focus groups with federal managers cited progress in this area as
a key accomplishment of GPRA. However, the participants also commented
that much remains to be done in this area. The comments ranged from the
general--GPRA provides a framework for planning and budgeting, to the
more specific--GPRA created a definition of programs and how they will
help the agency achieve its goals/objectives and the amount of money
that will be required to achieve said goals/objectives. One of the
comments implied that GPRA has helped to prioritize agency efforts by
helping agencies align their efforts with programs or activities that
make a difference. A political appointee we interviewed echoed this
comment, stating that GPRA was pushing the department to think about
what it gets out of the budget, not just what it puts into it--12 to 15
years ago the "so what" was missing from the budget process. Another
political appointee we interviewed stated that the department was in
the process of tying its goals to its budget formulation and execution
processes and linking program costs to departmental goals. A former
political appointee discussed how his department used program
performance information to inform a major information systems
investment decision.
Furthermore, GAO case studies on the integration of performance
information in budget decision making found that performance
information has been used to inform the allocation of resources and for
other management purposes at selected agencies. For example, the
Veterans Health Administration provides its health care networks with
performance information on patterns of patient care and patient health
outcomes, which can be used to analyze resource allocation and costs
and reallocate resources as appropriate.[Footnote 21] Officials at the
Administration for Children and Families said that training and
technical assistance and salaries and expense funds are often allocated
based on program and performance needs.[Footnote 22] The Nuclear
Regulatory Commission monitors performance against targets and makes
resource adjustments, if needed, to achieve those targets.[Footnote 23]
Although there has been progress in formally establishing the linkages
between budgets and plans, our survey results are somewhat conflicting
and have not reflected any notable changes either in managers'
perceptions governmentwide as to their personal use of plans or
performance information when allocating resources, or in their
perceptions about the use of performance information when funding
decisions are made about their programs. Our 2003 survey data show that
a large majority of federal managers reported that they consider their
agency's strategic goals when they are allocating resources. As shown
in figure 7, on our 2003 survey, an estimated 70 percent of all federal
managers agreed to a great or very great extent that they considered
their agency's strategic goals when allocating resources. However,
using our 1997 survey responses as a baseline, it was not a
statistically significant increase over 64 percent of the managers who
responded comparably then. As shown in figure 8, a similar, but
somewhat smaller, majority (60 percent) of managers who expressed an
opinion on our 2003 survey agreed to a great or very great extent that
they used information from performance measurement when they were
involved in allocating resources. In 1997, the comparable response was
about the same at 62 percent. When we asked managers on another item,
however, about the extent to which they perceived funding decisions for
their programs being based on results or outcome-oriented performance
information, only 25 percent of federal managers in 2003 endorsed this
view to a great or very great extent. In 1997, 20 percent of managers
expressed a comparable view, again not a significant increase. (See
fig. 9.):
Figure 7: Percentage of Federal Managers Who Reported They Considered
Strategic Goals to a Great or Very Great Extent When Allocating
Resources:
[See PDF for image]
[End of figure]
Figure 8: Percentage of Federal Managers Who Reported They Considered
Performance Information to a Great or Very Great Extent When Allocating
Resources:
[See PDF for image]
Note: Percentages are based on those respondents answering on the
extent scale.
[End of figure]
Figure 9: Percentage of Federal Managers Who Reported That Funding
Decisions Were Based on Results or Outcome-Oriented Performance
Information to a Great or Very Great Extent:
[See PDF for image]
[End of figure]
[End of section]
Chapter 3: Agencies Have Addressed Many Critical Performance Planning
and Reporting Challenges, but Weaknesses Persist:
[End of section]
Beginning with federal agencies' initial efforts to develop effective
strategic plans in 1997 and annual performance plans and reports for
fiscal year 1999, Congress, GAO, and others have commented on the
quality of those efforts and provided constructive feedback on how
agency plans and reports could be improved. On the basis of our current
review of the strategic plans, annual performance plans, and annual
performance reports of six selected agencies--Education, DOE, HUD, DOT,
SBA, and SSA--we found that these documents reflect much of the
feedback that was provided. For example, goals were more quantifiable
and results oriented, and agencies were providing more information
about goals and strategies to address performance and accountability
challenges and the limitations to their performance data. However,
certain weaknesses, such as lack of detail on how annual performance
goals relate to strategic goals and how agencies are coordinating with
other entities to achieve common objectives, persist. A detailed
discussion of our scope and methodology and the results of our reviews
of the six agencies' most recent strategic plans, annual performance
plans, and annual performance reports compared to initial efforts are
contained in appendixes III, IV, and V, respectively.
Quality of Selected Strategic Plans Reflects Improvements over Initial
Drafts:
Under GPRA, strategic plans are the starting point and basic
underpinning for results-oriented management. GPRA requires that an
agency's strategic plan contain six key elements: (1) a comprehensive
agency mission statement; (2) agencywide long-term goals and objectives
for all major functions and operations; (3) approaches (or strategies)
and the various resources needed to achieve the goals and objectives;
(4) a description of the relationship between the long-term goals and
objectives and the annual performance goals; (5) an identification of
key factors, external to the agency and beyond its control, that could
significantly affect the achievement of the strategic goals; and (6) a
description of how program evaluations were used to establish or revise
strategic goals and a schedule for future program evaluations.
Our 1997 review of agencies' draft strategic plans found that a
significant amount of work remained to be done by executive branch
agencies if their strategic plans were to fulfill the requirements of
GPRA, serve as a basis for guiding agencies, and help congressional and
other policymakers make decisions about activities and
programs.[Footnote 24] Our assessment of 27 agencies' initial draft
strategic plans revealed several critical strategic planning issues
that needed to be addressed. These planning issues were as follows:
* Most of the draft plans did not adequately link required elements in
the plans, such as strategic goals to annual performance goals.
* Long-term strategic goals often tended to have weaknesses.
* Many agencies did not fully develop strategies explaining how their
long-term strategic goals would be achieved.
* Most agencies did not reflect in their draft plans the identification
and planned coordination of activities and programs that cut across
multiple agencies.
* The draft strategic plans did not adequately address program
evaluations.
We noted that Congress anticipated that it may take several planning
cycles to perfect the process and that strategic plans would be
continually refined as various planning cycles occur. We also
recognized that developing a strategic plan is a dynamic process and
that agencies, with input from OMB and Congress, were continuing to
improve their plans.
Agencies have now had 6 years to refine their strategic planning
processes. Although the six strategic plans we looked at for this
review reflected many new and continuing strengths as well as
improvements over the 1997 initial drafts, we continued to find certain
persistent weaknesses. As depicted in table 1, of the six elements
required by GPRA, the plans generally discussed all but one--program
evaluation, an area in which we have found capacity is often lacking in
federal agencies.[Footnote 25] Although the strategic plans generally
listed the program evaluations agencies planned to complete over the
planning period, they generally did not address how the agencies
planned to use their evaluations to establish new or revise existing
strategic goals, as envisioned by GPRA. Finally, although not required
by GPRA, the strategic plans would have benefited from more complete
discussions of how agencies planned to coordinate with other entities
to address common challenges or achieve common or complementary goals.
Appendix III provides a more detailed discussion of (1) the required
and other useful elements we reviewed to assess strategic plan
strengths and weaknesses and (2) changes in the quality of the six
agencies' strategic plans we reviewed.
Table 1: Agencies' Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Agency strategic plans: Department of Education;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Education;
Plan year: 2002;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Energy;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Energy;
Plan year: 2003[A];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Housing and Urban Development;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: No;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: No;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Housing and Urban Development;
Plan year: 2003;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship between long-
term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Small Business Administration;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Small Business Administration;
Plan year: 2001[B];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship between long-
term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Social Security Administration;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Social Security Administration;
Plan year: 2003;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Transportation;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: No;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Transportation;
Plan year: Plan year: 2003[A];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Sources: GAO analysis of agencies' strategic plans in effect at the time of our review. See also,
U.S. General Accounting Office, The Results Act: Observations on the
Department of Education's June 1997 Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R]
(Washington, D.C.: July 18, 1997);
Results Act: Observations on DOE's Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R]
(Washington, D.C.: July 11, 1997);
The Results Act: Observations on the Department of Transportation's
Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R]
(Washington, D.C.: July 30, 1997);
The Results Act: Observations on the Social Security Administration's
June 1997 Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R]
(Washington, D.C.: July 22, 1997);
The Results Act: Observations on the Small Business Administration's
Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R]
(Washington, D.C.: July 11, 1997);
The Results Act: Observations on the Department of Housing and Urban
Development's Draft Strategic Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R]
(Washington, D.C.: Aug. 8, 1997).
[A] The 2003 plans for DOE and DOT were in draft form during the time
of our review.
[B] At the time of our review, the most recent SBA strategic plan was
for fiscal years 2001-2008. SBA released a new strategic plan for
fiscal years 2003-2008 in October 2003.
[End of table]
Strategic Planning Strengths and Improvements from Initial Draft Plans:
Consistent with our review of agencies' 1997 strategic plans, the
recent strategic plans we reviewed generally contained mission
statements that were results oriented, distinct from other agencies,
and covered the agencies' major activities. DOT's mission statement had
improved by reflecting additional language from its enabling
legislation that we recommended adding during our 1997 review. Still
improvement could be made in this area as is shown by DOE's mission
statement. DOE's mission was results oriented but did not address the
department's activities related to energy supply and conservation.
Our review of the current strategic plans also revealed improvements in
the development of agencies' long-term, strategic goals--essential for
results-oriented management. Although GPRA does not require that all of
an agency's long-term, strategic goals be results oriented, the intent
of GPRA is to have agencies focus their strategic goals on results to
the extent feasible. In addition, as required by GPRA, the goals should
be expressed in a manner that could be used to gauge success in the
future and should cover an agency's major functions or activities. All
of the strategic plans we reviewed contained long-term, strategic goals
that demonstrated improvements in the quality of their 1997 goals.
Agencies' long-term strategic goals generally covered their missions,
were results oriented, and were expressed in a manner that could be
used to gauge future success. For example, SBA improved the quality of
its long-term goals by focusing more on key outcomes to be achieved and
less on process improvements, as was the case in its 1997 plan. In some
cases, we observed strategic goals that addressed the agency's
organizational capacity to achieve results, such as SSA's long-term
goal to strategically manage and align staff to support its mission.
We also found improvements in how agencies' current plans addressed
performance and accountability challenges we had identified, a key
weakness we identified in our earlier review. Each of the agency plans
we reviewed discussed the long-term goals and strategies to address the
challenges that we had identified. For example, Education's strategic
plan contained a long-term strategic goal to modernize the Federal
Student Assistance programs and address identified problems in this
area, which we have designated as high risk since 1990.[Footnote 26]
SSA noted that it considered GAO-identified performance and
accountability challenges when it determined its strategic goals and
objectives, however not all of the challenges are clearly addressed in
the plan.
A third area of improvement we observed was in the description of the
strategies agencies planned to use to achieve their long-term strategic
goals. In our review of agencies' 1997 draft strategic plans, we found
that many agencies did not fully develop strategies explaining how
their long-term strategic goals would be achieved. In contrast, all six
of the current strategic plans we reviewed contained strategies that
appeared logically linked to achieving the agencies' long-term goals.
Other strengths and improvements we observed in meeting GPRA's basic
requirements involved the reporting of external factors that could
affect the achievement of the long-term goals and the identification of
crosscutting activities, although as indicated below these discussions
could be improved. The six agencies reviewed for this report each
reported on external factors in current strategic plans. For example,
for each of the strategic objectives in DOT's strategic plan, DOT lists
factors external to its control and how those factors could affect the
achievement of its objectives. Although not a requirement, some of the
better plans we reviewed discussed strategies to ameliorate the effect
of external factors. For example, for an external factor on teacher
certification under a goal on reading, Education's plan states that the
agency "will work with the states and national accreditation bodies to
encourage the incorporation of research-based reading instruction into
teacher certification requirements.":
We have frequently reported that a focus on results, as envisioned by
GPRA, implies that federal programs contributing to the same or similar
results should be closely coordinated to ensure that goals are
consistent and, as appropriate, program efforts are mutually
reinforcing. This means that federal agencies are to look beyond their
organizational boundaries and coordinate with other agencies to ensure
that their efforts are aligned. During our 1997 review, we found that
most agencies did not reflect in their draft plans the identification
and planned coordination of activities and programs that cut across
multiple agencies. In contrast, each of the six current agency
strategic plans that we reviewed identified at least some activities
and programs that cut across multiple agencies. For example, SBA's 1997
plan contained no evidence of how the agency coordinated with other
agencies, but the current plan contained a separate section describing
crosscutting issues in the areas of innovation and research assistance,
international trade assistance, business development assistance,
veterans affairs, and disaster assistance.
Critical Strategic Planning Issues Needing Further Improvement:
First, consistent with our 1997 review, the strategic plans we reviewed
did not adequately link required elements in the plans. Although all of
the agencies we reviewed provided some information on the relationship
between their long-term and annual goals, the extent of information
provided on how annual goals would be used to measure progress in
achieving the long-term goals varied greatly. In the case of DOE, the
plan provides a very brief description of the overall relationship
between its long-term and annual goals with examples, but does not
demonstrate how it will assess progress for each of its long-term goals
and objectives. Another plan, DOT's, refers the reader to the annual
performance plan for information about annual goals. We have reported
that this linkage is critical for determining whether an agency has a
clear sense of how it will assess progress toward achieving its
intended results.
Second, although the agencies' descriptions of their strategies had
improved since our initial reviews, with few exceptions, their
strategies generally did not include information on how the agencies
plan to align their activities, core processes, human capital, and
other resources to support their mission-critical outcomes and whether
they have the right mix of activities, skills, and resources to achieve
their goals. Such information is critical to understanding the
viability of the strategies. Furthermore, none of the agencies
discussed alternative strategies they had considered in developing
their plans. Without such discussions, it is unclear whether agency
planning processes were truly strategic or simply a recasting of
existing activities, processes, etc.
HUD was the only agency that provided any details of how it intended to
coordinate with other agencies to achieve common or complementary goals
for its crosscutting programs or activities. For example, to support
its goal of "Equal Opportunity in Housing," HUD's plan states that HUD
and the Department of Justice continue to coordinate their fair housing
enforcement activities, especially with respect to responding quickly
and effectively to Fair Housing Act complaints that involve criminal
activity (e.g., hate crimes), a pattern and practice of housing
discrimination, or the legality of state and local zoning or other land
use laws or ordinances. We have reported that mission fragmentation and
program overlap are widespread throughout the federal
government.[Footnote 27] As such, interagency coordination is important
for ensuring that crosscutting programs are mutually reinforcing and
efficiently implemented.
Finally, the draft strategic plans did not adequately address program
evaluations. In combination with an agency's performance measurement
system, program evaluations can provide feedback to the agency on how
well its activities and programs contributed to achieving strategic
goals. For example, evaluations can be a potentially critical source of
information for Congress and others in assessing (1) the
appropriateness and reasonableness of goals; (2) the effectiveness of
strategies by supplementing performance measurement data with impact
evaluation studies; and (3) the implementation of programs, such as
identifying the need for corrective action. Evaluations are important
because they potentially can be critical sources of information for
ensuring that goals are reasonable, strategies for achieving goals are
effective, and that corrective actions are taken in program
implementation. Five out of the six current plans that we reviewed
included a discussion of program evaluations, however for most of these
plans the discussions lacked critical information required by GPRA,
such as a discussion of how evaluations were used to establish
strategic goals or a schedule of future evaluations. For example, DOE's
plan stated that internal, GAO, and Inspector General (IG) evaluations
were used as resources to develop its draft strategic plan, but
specific program evaluations were not identified.
Fiscal Year 2004 Annual Performance Plans Addressed Some Weaknesses of
Earlier Plans, but Still Have Room for Significant Improvement:
According to our review of agencies' first annual performance plans,
which presented agencies' annual performance goals for fiscal year
1999,[Footnote 28] we found that substantial further development was
needed for these plans to be useful in a significant way to
congressional and other decision makers. Most of the fiscal year 1999
plans that we reviewed contained major weaknesses that undermined their
usefulness in that they (1) did not consistently provide clear pictures
of agencies' intended performance, (2) generally did not relate
strategies and resources to performance, and (3) provided limited
confidence that agencies' performance data will be sufficiently
credible. Although all of the fiscal year 1999 plans contained valuable
information for decision makers, their weaknesses caused their
usefulness to vary considerably within and among plans.
As shown in table 2, our current review of agencies' fiscal year 2004
performance plans found that five agencies--Education, HUD, SBA, SSA,
and DOT--improved their efforts to provide a clear picture of intended
performance, with SSA and DOT being the clearest. Furthermore, the same
five agencies improved the specificity of the strategies and resources
they intended to use to achieve their performance goals, with DOT being
the most specific. Finally, the same five agencies--Education, HUD,
SBA, SSA, and DOT--made improvements in the area of greatest weakness-
-reporting on how they will ensure performance data will be credible.
However, only DOT's plan provided a full level of confidence that the
performance data the agency intended to collect would be credible.
Appendix IV provides a more detailed discussion of (1) the required and
other useful elements we reviewed to assess the clarity of the picture
of intended performance, the specificity of the strategies and
resources, and the level of confidence in the performance data and (2)
changes in the quality of the six agencies' annual performance plans we
reviewed.
Table 2: Characterizations of Agencies' Fiscal Year 1999 and 2004
Annual Performance Plans:
Agency: Department of Education;
Picture of intended performance: (unclear, limited, general, clear):
1999: Limited;
Picture of intended performance: (unclear, limited, general, clear):
2004: General;
Strategies and resources: (no, limited, general, specific): 1999:
Limited;
Strategies and resources: (no, limited, general, specific): 2004:
General;
Data credible: (no, limited, general, full): 1999: Limited;
Data credible: (no, limited, general, full): 2004: General.
Agency: Department of Energy;
Picture of intended performance: (unclear, limited, general, clear):
1999: Limited;
Picture of intended performance: (unclear, limited, general, clear):
2004: Limited;
Strategies and resources: (no, limited, general, specific): 1999:
General;
Strategies and resources: (no, limited, general, specific): 2004:
General;
Data credible: (no, limited, general, full): 1999: Limited;
Data credible: (no, limited, general, full): 2004: Limited.
Agency: Department of Housing and Urban Development;
Picture of intended performance: (unclear, limited, general, clear):
1999: Limited;
Picture of intended performance: (unclear, limited, general, clear):
2004: General;
Strategies and resources: (no, limited, general, specific): 1999:
Limited;
Strategies and resources: (no, limited, general, specific): 2004:
General;
Data credible: (no, limited, general, full): 1999: Limited;
Data credible: (no, limited, general, full): 2004: General.
Agency: Small Business Administration;
Picture of intended performance: (unclear, limited, general, clear):
1999: Limited;
Picture of intended performance: (unclear, limited, general, clear):
2004: General;
Strategies and resources: (no, limited, general, specific): 1999:
Limited;
Strategies and resources: (no, limited, general, specific): 2004:
General;
Data credible: (no, limited, general, full): 1999: Limited;
Data credible: (no, limited, general, full): 2004: General.
Agency: Social Security Administration;
Picture of intended performance: (unclear, limited, general, clear):
1999: Limited;
Picture of intended performance: (unclear, limited, general, clear):
2004: Clear;
Strategies and resources: (no, limited, general, specific): 1999:
Limited;
Strategies and resources: (no, limited, general, specific): 2004:
General;
Data credible: (no, limited, general, full): 1999: No;
Data credible: (no, limited, general, full): 2004: General.
Agency: Department of Transportation;
Picture of intended performance: (unclear, limited, general, clear):
1999: General;
Picture of intended performance: (unclear, limited, general, clear):
2004: Clear;
Strategies and resources: (no, limited, general, specific): 1999:
General;
Strategies and resources: (no, limited, general, specific): 2004:
Specific;
Data credible: (no, limited, general, full): 1999: Limited;
Data credible: (no, limited, general, full): 2004: Full.
Sources: GAO analysis of agencies' fiscal year 2004 annual performance plans and
U.S. General Accounting Office, Results Act: Observations on the
Department of Education's Fiscal Year 1999 Annual Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-172R]
(Washington, D.C.: June 8, 1998);
Results Act: Observations on DOE's Annual Performance Plan for Fiscal
Year 1999,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-194R]
(Washington, D.C.: May 28, 1998);
Results Act: Observations on the Department of Housing and Urban
Development's Fiscal Year 1999 Annual Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-159R]
(Washington, D.C.: June 5, 1998);
Results Act: Observations on the Small Business Administration's Fiscal
Year 1999 Annual Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98- 200R]
(Washington, D.C.: May 28, 1998);
The Results Act: Observations on the Social Security Administration's
Fiscal Year 1999 Annual Performance Plan,
[Hyperlink, http://www.gao.gov/cgi- bin/getrpt?GAO/HEHS-98-178R]
(Washington, D.C.: June 9, 1998);
Results Act: Observations on the Department of Transportation's Annual
Performance Plan for Fiscal Year 1999,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-180R]
(Washington, D.C.: May 12, 1998).
[End of table]
Plans Generally Provided a Clearer Picture of Intended Performance,
Except for Crosscutting Areas:
At the most basic level, an annual performance plan is to provide a
clear picture of intended performance across the agency. Such
information is important to Congress, agency managers, and others for
understanding what the agency is trying to achieve, identifying
subsequent opportunities for improvement, and assigning
accountability. Our current review of agencies' fiscal year 2004
performance plans found that five of the six agencies provided a
clearer picture of intended performance than their fiscal year 1999
plans did, although only two of the 2004 plans--DOT's and SSA's--
received the highest rating possible. As shown in table 2, except for
DOT, the six agencies we reviewed for this report initially provided a
limited picture of intended performance. Most of the fiscal year 1999
performance plans we previously reviewed had at least some objective,
quantifiable, and measurable goals, but few plans consistently included
a comprehensive set of goals that focused on the results that programs
were intended to achieve. Moreover, agencies did not consistently
follow OMB's guidance that goals for performance and accountability
challenges be included in the plans. Agencies' plans generally showed
how their missions and strategic goals were related to their annual
performance goals and covered all of the program activities in the
agencies' budget requests.[Footnote 29] In addition, many agencies took
the needed first step of identifying their crosscutting efforts, with
some including helpful lists of other agencies with which they shared a
responsibility for addressing similar national issues. However, the
plans generally did not go further to describe how agencies expected to
coordinate their efforts with other agencies.
The fiscal year 2004 plans improved the picture of performance by
making annual goals and performance measures more results oriented,
objective, and quantifiable. For example, Education's plan included a
measure for the number of states meeting their eighth grade mathematics
achievement targets under the long-term goal to improve mathematics and
science achievement for all students. We previously criticized
Education's 1999 plan for lacking such outcome-oriented measures.
Another overall improvement we observed was that all of the plans
described intended efforts to address performance and accountability
challenges we and others had previously identified. For instance, to
address the governmentwide high-risk area of strategic human capital
management, HUD states that to develop its staff capacity, it will
complete a comprehensive workforce analysis in 2004 to serve as the
basis to fill mission critical skill gaps through succession planning,
hiring, and training initiatives in a 5-year human capital management
strategy. The clarity of DOE's plan remained limited because its annual
goals were not clearly linked to its mission, the long-term goals in
its strategic plan, or the program activities in its budget request.
Although five of the six agencies improved the clarity of the picture
of intended performance, improvement is still needed in reporting on
crosscutting efforts. In both the 1999 and 2004 plans, many agencies
identified their crosscutting efforts, with some including helpful
lists of other agencies with which they shared a responsibility for
addressing similar national issues. Our review of fiscal year 2004
plans shows that the six agencies we reviewed still did not discuss how
they expected to coordinate with other agencies to address common
challenges or to achieve common or complementary performance goals. As
we have reported previously, improved reporting on crosscutting efforts
can help Congress use the annual performance plan to evaluate whether
the annual goals will put the agency on a path toward achieving its
mission and long-term strategic goals. In addition, the plans can aid
in determining efforts to reduce significant program overlap and
fragmentation that can waste scarce resources, confuse and frustrate
program customers, and limit overall program effectiveness.
None of the six agencies' plans indicated an intention to request
waivers of specific administrative procedural requirements and controls
that may be impeding an agencies' ability to achieve results. This
provision of GPRA allows agencies greater managerial flexibility in
exchange for accountability for results. We previously reported on the
results of the pilot project to implement this provision of GPRA and
found that the pilot did not work as intended.[Footnote 30] OMB did not
designate any of the seven departments and one independent agency that
submitted a total of 61 waiver proposals as GPRA managerial
accountability and flexibility pilots. For about three-quarters of the
waiver proposals, OMB or other central management agencies determined
that the waivers were not allowable for statutory or other reasons or
that the requirement for which the waivers were proposed no longer
existed. For the remaining proposals, OMB or other central management
agencies approved waivers or developed compromises by using authorities
that were already available independent of GPRA.
Plans More Specifically Related Strategies and Resources to Performance
Goals:
To judge the reasonableness of an agency's proposed strategies and
resources, congressional and other decision makers need complete
information on how the proposed strategies and resources will
contribute to the achievement of agency goals. Agencies generally
improved their plans by better relating strategies and resources to
performance. Education's, HUD's, SBA's, and SSA's 1999 plans had a
limited discussion, while DOE's and DOT's 1999 plans had a general
discussion. In 2004, five of the six plans--Education's, DOE's, HUD's,
SBA's, and SSA's--provided general discussions of how their strategies
and resources would contribute to achieving their performance goals.
DOT's 2004 plan improved to include a specific discussion.
Our review of the 1999 plans found that most agencies' performance
plans did not provide clear strategies that described how performance
goals would be achieved. In contrast, the 2004 performance plans we
reviewed generally provided lists of the agencies' current array of
programs and initiatives. Several plans provided a perspective on how
these programs and initiatives were necessary or helpful for achieving
results. For example, DOE and HUD included in their plans a "means and
strategies" section for each of their goals that described how the goal
would be achieved. One strategy DOE identified to meet its goal of
contributing unique, vital facilities to the biological environmental
sciences was to conduct peer reviews of the facilities to assess the
scientific output, user satisfaction, the overall cost-effectiveness of
each facility's operations, and their ability to deliver the most
advanced scientific capability.
In addition, each of the agencies' plans identified the external
factors that could influence the degree to which goals are achieved.
Some of the better plans, such as DOT's and SBA's, provided strategies
to mitigate the negative factors or take advantage of positive factors,
as appropriate. For example, for its transportation accessibility
goals, DOT's plan states that as the population ages, more people will
require accessible public transit, for which states and local agencies
decide how best to allocate federally provided resources. One of the
strategies DOT intends to employ to address this external factor is the
"Special Needs of Elderly Individuals and Individuals with
Disabilities" grant program. The plan states the grant program will
help meet transportation needs of the elderly and persons with
disabilities when regular transportation services are unavailable,
insufficient, or inappropriate to meet their needs.
Agencies' 2004 plans did not consistently describe all the resources
needed and how they would be used to achieve agency goals. Our review
of agencies' fiscal year 1999 plans found that most did not adequately
describe--or reference other appropriate documents that describe--the
capital, human, information, and financial resources needed to achieve
their agencies' performance goals. The 2004 plans we reviewed generally
described the funding levels needed to achieve their performance goals
overall and in some cases broke out funding needs by high-level
performance goal. For example, SSA's plan provides a general
perspective on the financial resources needed to achieve its
performance goals because it provides budget information by account and
program activity. However, the plan is neither structured by budget
program activity or account, nor does it provide a crosswalk between
the strategic goals and budget program accounts. In contrast, HUD's
plan presented its requested funding and staffing levels at the
strategic goal level, but did not present budget information at the
level of its annual goals. In addition, although the plans make brief
mention of nonfinancial resources, such as human capital, information
technology, or other capital investments, little information is
provided on how such resources would be used to achieve performance
goals.
Plans Continue to Provide Less Than Full Confidence That Performance
Data Will Be Credible:
Credible performance information is essential for accurately assessing
agencies' progress towards the achievement of their goals and, in cases
where goals are not met, identifying opportunities for improvement or
whether goals need to be adjusted. Under GPRA, agencies' annual
performance plans are to describe the means that will be used to verify
and validate performance data. To help improve the quality of agencies'
performance data, Congress amended GPRA through the Reports
Consolidation Act of 2000 to require that agencies assess the
completeness and reliability of the performance data in their
performance reports. Agencies were also required to discuss in their
report any material inadequacies in the completeness and reliability of
their performance data and discuss actions to address these
inadequacies. Meeting these new requirements suggests the need for
careful planning to ensure that agencies can comment accurately on the
quality of the performance data they report to the public.
As shown in table 2, although five of the six agencies we reviewed
improved in reporting how they plan to ensure that performance data
will be credible, only one agency--DOT--improved enough over its 1999
plan to provide a full level of confidence in the credibility of its
performance data. Four agencies--Education, HUD, SBA, and SSA--improved
enough to provide a general level of confidence. However, DOE provided
the same limited level of confidence in the credibility of the
performance data as in its 1999 plan. Regarding all 24 of the fiscal
year 1999 performance plans we reviewed, we found most provided only
superficial descriptions of procedures that agencies intended to use to
verify and validate performance data. Moreover, in general, agencies'
performance plans did not include discussions of documented limitations
in financial and other information systems that may undermine efforts
to produce high-quality data. As we have previously noted, without such
information, and strategies to address those limitations, Congress and
other decision makers cannot assess the validity and reliability of
performance information.
We found that each of the 2004 plans we reviewed contained some
discussion of the procedures the agencies would use to verify and
validate performance information, although in some cases the discussion
was inconsistent or limited. For example, the discussions of SBA's
verification and validation processes for its indicators in the 2004
plan were generally one-or two-sentence statements. SBA also noted that
it does not independently verify some of the external data it gathers
or that it does not have access to the data for this purpose. In
contrast, the DOT plan referred to a separate compendium available on-
line that provides source and accuracy statements, which give more
detail on the methods used to collect performance data, sources of
variation and bias in the data, and methods used to verify and validate
the data.
In addition, all of the agencies except DOE discussed known limitations
to performance data in their plans. These agencies' plans generally
provided information about the quality of each performance measure,
including any limitations. According to DOE officials, DOE's plan
generally does not discuss data limitations because the department
selected goals for which data are expected to be available and
therefore did not anticipate finding any limitations. However, in our
2003 Performance and Accountability Series report on DOE, we identified
several performance and accountability challenges where data were a
concern, such as the need for additional information on the results of
contractors' performance to keep projects on schedule and within
budget.[Footnote 31] DOE's contract management continues to be a
significant challenge for the department and remains at high risk.
Finally, the remaining five agencies also discussed plans to address
limitations to the performance data. For example, DOT's plan provided a
general discussion of the limitations to the internal and external
sources of data used to measure performance. Detailed discussions were
contained in an appendix to the plan and separate source and accuracy
statements. This information had been lacking in its 1999 plan.
Education, HUD, SBA, and SSA also provided information on limitations
to their performance data and plans for improvement.
Strengths and Weaknesses of Selected Agencies' Fiscal Year 2002 Annual
Performance Reports:
Key to improving accountability for results as Congress intended under
GPRA, annual performance reports are to document the results agencies
have achieved compared to the goals they established. To be useful for
oversight and accountability purposes, the reports should clearly
communicate performance results, provide explanations for any unmet
goals as well as actions needed to address them, and discuss known data
limitations as well as how the limitations are to be addressed in the
future. Compared to our reviews of the six agencies' fiscal year 1999
performance reports, we identified a number of strengths and
improvements as well as areas that continued to need improvement.
Because the scope of our review of the fiscal year 2002 reports was
broader than that for the fiscal year 1999 reports we previously
reviewed, we were unable to make specific comparisons for the three
characteristics we used to assess the fiscal year 2002 reports.
However, we discuss comparative information on aspects of the reports
where available. Table 3 shows the results of our assessment of the six
agencies' annual performance reports for fiscal year 2002. Appendix V
provides a more detailed discussion of (1) the required and other
useful elements we reviewed to assess the clarity of the picture of
performance, the clarity of the linkage between costs and performance,
and the level of confidence in the performance data and (2) changes in
the quality of the six agencies' annual performance plans we reviewed.
Table 3: Characterizations of Agencies' 2002 Annual Performance
Reports:
Agency: Department of Education;
Picture of performance: (unclear, limited, general, clear): Limited;
Resources linked to results: (no, limited, general, clear): Clear;
Data credible: (no, limited, general, full): General.
Agency: Department of Energy;
Picture of performance: (unclear, limited, general, clear): General;
Resources linked to results: (no, limited, general, clear): Limited;
Data credible: (no, limited, general, full): Limited.
Agency: Department of Housing and Urban Development;
Picture of performance: (unclear, limited, general, clear): General;
Resources linked to results: (no, limited, general, clear): No;
Data credible: (no, limited, general, full): General.
Agency: Small Business Administration;
Picture of performance: (unclear, limited, general, clear): Limited;
Resources linked to results: (no, limited, general, clear): General;
Data credible: (no, limited, general, full): General.
Agency: Social Security Administration;
Picture of performance: (unclear, limited, general, clear): General;
Resources linked to results: (no, limited, general, clear): Limited;
Data credible: (no, limited, general, full): General.
Agency: Department of Transportation;
Picture of performance: (unclear, limited, general, clear): General;
Resources linked to results: (no, limited, general, clear): No;
Data credible: (no, limited, general, full): Full.
Sources: GAO analysis of agencies' fiscal year 2002 annual performance
reports and
U.S. General Accounting Office, Observations on the Department of
Education's Fiscal Year 1999 Performance Report and Fiscal Year 2001
Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-00-128R];
(Washington, D.C. June 30, 2000);
Observations on the Department of Energy's Fiscal Year 1999
Accountability Report and Fiscal Years 2000 and 2001 Performance Plans,
GAO/RCED-00-209R; (Washington, D.C. June 30, 2000);
Observations on the Department of Housing and Urban Development's
Fiscal Year 1999 Performance Report and Fiscal Year 2001 Performance
Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00- 211R];
(Washington, D.C. June 30, 2000);
Observations on the Small Business Administration's Fiscal Year 1999
Performance Report and Fiscal Year 2001 Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00-207R];
(Washington, D.C. June 30, 2000);
Observations on the Social Security Administration's Fiscal Year 1999
Performance Report and Fiscal Year 2001 Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-00-126R];
(Washington, D.C. June 30, 2000);
and Observations on the Department of Transportation's Fiscal Year 1999
Performance Report and Fiscal Year 2001 Performance Plan,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-00-201R];
(Washington, D.C. June 30, 2000).
[End of table]
Progress in Providing a Clear Picture of Performance:
The six agency reports that we reviewed contained a number of
strengths, some of which we can describe as improvements over the
reports on fiscal year 1999 performance. A key strength of four of the
2002 reports (Education, HUD, DOT, SSA) was a discussion of the
relationship between the strategic plan, performance plan, and
performance report. For example, SSA's report identified relevant
results that were linked to its strategic objective to deliver
"citizen-centered, world-class service," such as maintaining the
accuracy, timeliness, and efficiency of service to people applying for
its benefit programs. The clarity of the DOE and SBA reports was
limited by not clearly relating agency performance results to strategic
and annual performance goals. For example, the structure of SBA's
report reflected the objectives in its draft 2003 to 2008 strategic
plan rather than those in its 2002 performance plan, making it
difficult to assess progress against the original 2002 objectives.
Furthermore, although there is no "right" number of performance
measures to be used to assess progress, a number of the plans allowed
for an easier review of results by limiting the overall number of
measures presented or by highlighting key performance measures of
greatest significance to their programs. For example, SBA discussed a
total of 19 performance goals and DOT discussed a total of 40. Although
SSA discussed a total of 69 performance goals, the report highlighted
its progress in achieving 14 key goals. In contrast, Education, HUD,
and DOE presented a total of 120, 184, and 260 measures, respectively.
Furthermore, while Education and SSA each provided a table showing
progress across all its measures, the other agencies did not provide
such summary information.
As we found in our earlier reviews, the six agencies' fiscal year 2002
reports generally allowed for an assessment of progress made in
achieving agency goals. Some of the reports made this assessment easier
than others by providing easy-to-read summary information. For example,
SSA provided a table at the beginning of the report that summarized the
results for each of its 69 indicators with the following dispositions:
met, not met, almost met, and data not yet available. Other reports,
such as HUD's, required an extensive review to make this assessment. In
addition, to place current performance in context, each of the
agencies' reports contained trend information, as required by GPRA,
which allowed for comparisons between current year and prior year
performance.
In addition, the majority of agencies maintained, or demonstrated
improvements over, the quality of their 1999 reports in discussing the
progress achieved in addressing performance and accountability
challenges identified by agency IGs and GAO. For example, SBA's report
contained two broad overviews and an appendix describing the status of
GAO audits and recommendations, as well as a description of the most
serious management challenges SBA faces as identified by the agency's
IG.
Unfortunately, many of the weaknesses we identified in the agencies'
fiscal year 2002 reports were similar to those we found in their fiscal
year 1999 reports related to the significant number of performance
goals (1) which were not achieved and lacked explanations or plans for
achieving the goal in the future and (2) for which performance data
were unavailable. Three of the six agencies we reviewed--HUD, SSA, and
Transportation--did not consistently report the reasons for not meeting
their goals. For example, Transportation provided explanations for only
5 of the 14 goals it did not meet. In addition, similar to our 1999
report findings, three of the six agencies we reviewed--HUD, SBA, and
DOT--did not discuss their plans or strategies to achieve unmet goals
in the future. For example, HUD reported "substantially meeting" only
47 percent of the performance targets in fiscal year 2002. However,
although HUD provides various reasons for not meeting all its targets,
it offers no information on plans or time frames to achieve the goals
in the future. Finally, we continued to observe a significant number of
goals for which performance data were unavailable. For example,
performance data for 10 of SBA's 19 performance goals were unavailable.
In addition, the majority of the reports we reviewed did not include
other GPRA requirements. The reports generally did not evaluate the
performance plan for the current year relative to the performance
achieved toward the performance goals in the fiscal year covered by the
report. The reports also did not discuss the use or effectiveness of
any waivers in achieving performance goals. In addition, for two of the
agencies--DOE and SBA--program evaluation findings completed during the
fiscal year were not summarized. As we have previously noted, such
evaluations could help agencies understand the relationship between
their activities and the results they hope to achieve.
Progress in Linking Resources to Results:
Although linking costs to performance goals is not a requirement of
GPRA, both GPRA and the CFO Act emphasized the importance of linking
program performance information with financial information as a key
feature of sound management and an important element in presenting to
the public a useful and informative perspective on federal spending.
The committee report for GPRA suggested that developing the capacity to
relate the level of program activity with program costs, such as cost
per unit of result, cost per unit of service, or cost per unit of
output, should be a high priority. In our survey of federal managers,
this year we asked for the first time the extent to which federal
managers had measures of cost-effectiveness for the programs they were
involved with. Only 31 percent of federal managers we surveyed reported
having such measures to a great or very great extent, lower than any of
the other types of measures associated with GPRA we asked about by at
least 12 percent (see fig. 3 in ch. 2). Under the PMA, the current
administration has set an ambitious agenda for performance budgeting,
calling for agencies to better align budgets with performance goals and
focus on capturing full budgetary costs and matching those costs with
output and outcome goals. All this suggests that agencies will need to
develop integrated financial and performance management systems that
will enable the reporting of the actual costs associated with
performance goals and objectives along with presentations designed to
meet other budgetary or financial purposes, such as the accounts and
program activities found in the President's Budget and responsibility
segments found in financial statements.[Footnote 32]
Of the six agencies we reviewed, only Education's report clearly linked
its budgetary information to the achievement of its performance goals
or objectives. Education's report laid out, using both graphics and
text, the estimated appropriations associated with achieving each of
its 24 objectives. In addition the report provided the staffing in
full-time equivalent employment (FTEs) and an estimate of the funds
from salaries and expenses contributing to the support of each of these
objectives. SBA's report contained crosswalks that showed the
relationship between SBA's strategic goals, outcome goals, performance
goals, and programs. Because SBA shows the resources for each program,
a reader can infer a relationship between SBA's resources and
performance goals. However, the linkage between resources and results
would be clearer if results and resources were presented by performance
goal as well. SSA provided a limited view of the costs of achieving its
performance goals by providing the costs associated with four out of
five of its strategic goals.[Footnote 33] However, as reported by the
IG, SSA needs to further develop its cost accounting system, which
would help link costs to performance.[Footnote 34] DOE also provided a
limited view of the costs of achieving its performance goals by
organizing its performance information by budget program activity and
associated net costs. According to DOE officials, the department plans
to link its individual performance measures to the costs of program
activities in future reports. Neither HUD nor DOT provided information
on the cost of achieving individual performance goals or objectives.
Progress in Providing Confidence in the Credibility of Performance
Data:
To assess the degree to which an agency's report provided full
confidence that the agency's performance information was credible, we
examined the extent to which the reports discussed the quality of the
data presented. As shown in table 3, only DOT's report provided a full
level of confidence in the quality of the data. The other agencies
provided general or limited confidence in their data.
All six agencies in our current review complied with the Reports
Consolidation Act of 2000 by including assessments of the completeness
and reliability of their performance data in their transmittal letters.
In contrast, we found that only 5 of the 24 CFO Act agencies included
this information in their fiscal year 2000 performance
reports.[Footnote 35] Of the six agencies in our current review, only
DOE provided this assessment in its fiscal year 2000 report. For
example, the Secretary of DOT stated in the transmittal letter that the
2002 report "contains performance and financial data that are
substantially complete and reliable." However, only two of the six
agencies also disclosed material inadequacies in the completeness and
reliability of their performance data and discussed actions to address
the inadequacies in their transmittal letters. For example, SBA stated
in its transmittal letter that it is "working to improve the
completeness and reliability of the performance data for the advice
provided to small business through SBA's resource partners." SBA
explained that data for this aspect of its performance are collected
through surveys, which are inconsistent and not comparable, and for
which client responses are difficult to obtain. SBA stated that it is
working to improve the survey instruments it uses to obtain performance
data.
In addition to the requirements of the Reports Consolidation Act, we
have previously reported on other practices that enhance the
credibility of performance data that are not specifically required by
GPRA. For instance, discussions of standards and methods used by
agencies to assess the quality of their performance data in their
performance reports provide decision makers greater insight into the
quality and value of the performance data. None of the reports
explicitly referred to a specific standard they used, however, DOE
described its method for assuring data quality. The report states that
the heads of DOE's organizational elements certified the accuracy of
their performance data. DOE subsequently reviewed the data for quality
and completeness.
Other useful practices that help foster transparency to the public and
assist decision makers in understanding the quality of an agency's data
include: (1) discussion of data quality, including known data
limitations and actions to address the limitations, and (2) discussion
of data verification and validation procedures, including proposals to
review data collection and verification and validation procedures.
All six agencies' reports described data limitations, although
discussions were mostly brief and very high level. One exception was
DOT, which directed readers to the DOT Web site to obtain an assessment
of the completeness and reliability of its performance data and
detailed information on the source, scope, and limitations of the
performance data. HUD and SBA also discussed plans for addressing the
limitations. For example, HUD stated that to address problems with its
indicator on the number of homeowners who have been assisted with the
Home Investment Partnership Program (HOME), HUD has established a team
of managers, technical staff, and contractors to make a series of
improvements to the Integrated Disbursement and Information System
beginning in fiscal year 2003 that should reduce the need to clean up
the data.
Each of the six agencies' reports also discussed the procedures they
used to verify and validate their performance data. However, these
discussions ranged from the very general description of the DOE method
(noted previously), to the very detailed discussions provided by DOT.
DOT provides an on-line compendium that discusses the source and
accuracy of its data. Furthermore, DOT's 2002 report also describes
strategies being undertaken to address the quality of its data. The
report states that a DOT intermodal working group addressed data
quality issues by developing departmental statistical standards and by
updating source and accuracy statements for all of DOT's data programs.
The working group also worked to improve quality assurance procedures,
evaluate sampling and nonsampling errors, and develop common
definitions for data across modes.
[End of section]
Chapter 4: Challenges to GPRA Implementation Persist:
While a great deal of progress has been made in making federal agencies
more results oriented, numerous challenges still exist to effective
implementation of GPRA. The success of GPRA depends on the commitment
of top leadership within agencies, OMB, and Congress. However,
according to federal managers surveyed, top leadership commitment to
achieving results has not grown significantly since our 1997 survey.
Furthermore, although OMB has recently shown an increased commitment to
management issues, it significantly reduced its guidance to agencies on
GPRA implementation compared to prior years, and it is not clear how
the program goals developed through its PART initiative will complement
and integrate with the long-term, strategic focus of GPRA. Obtaining
leadership commitment to implement a strategic plan depends in part on
the usefulness and relevance of agency goals and strategies to agency
leaders, Congress, and OMB. However, GPRA's requirement to update
agency strategic plans every 3 years is out of sync with presidential
and congressional terms and can result in updated plans that do not
have the support of top administration leadership and key congressional
stakeholders.
As noted in chapter 2, more federal managers surveyed reported having
results-oriented performance measures for their programs and we would
expect to have seen similar increases in the use of this information
for program management. However, we did not observe any growth in their
reported use of this information for key management activities, such as
adopting new program approaches or changing work processes.
Additionally, managers noted human capital-related challenges that
impede results-oriented management, including a lack of authority and
training to carry out GPRA requirements, as well as a lack of
recognition for the results achieved.
Consistent with our previous work, federal managers in our focus groups
reported that significant challenges persist in setting outcome-
oriented goals, measuring performance, and collecting useful data.
However, our survey data suggested that federal managers do not
perceive issues, such as "difficulty distinguishing between the results
produced by the program and results caused by other factors" and
"difficulty obtaining data in time to be useful," to be substantial
hindrances to measuring performance or using performance information.
Additionally, mission fragmentation and overlap contribute to
difficulties in addressing crosscutting issues, particularly when those
issues require a national focus, such as homeland security, drug
control, and the environment. GAO has previously reported on a variety
of barriers to interagency cooperation, such as conflicting agency
missions, jurisdiction issues, and incompatible procedures, data, and
processes. We have also reported that OMB could use the provision of
GPRA that calls for OMB to develop a governmentwide performance plan to
integrate expected agency-level performance. Unfortunately, this
provision has not been fully implemented and the federal government
lacks a tool, such as a strategic plan, that could provide a framework
for a governmentwide reexamination of existing programs, as well as
proposals for new programs. Finally, federal managers in our focus
groups and political appointees we interviewed believed that Congress
does not use performance information to the fullest extent to conduct
oversight and to inform appropriations decisions. While there is
concern regarding Congress' use of performance information, it is
important to make sure that this information is initially useful. As a
key user of performance information, Congress needs to be considered a
partner in shaping agency goals at the outset. GPRA provides Congress
opportunities to influence agency performance goals through the
consultation requirement for strategic plans and through Congress'
traditional oversight role.
Top Leadership Does Not Consistently Show Commitment to Achieving
Results:
We have previously testified that perhaps the single most important
element of successful management improvement initiatives is the
demonstrated commitment of top leaders to change.[Footnote 36] This
commitment is most prominently shown through the personal involvement
of top leaders in developing and directing reform efforts.
Organizations that successfully address their long-standing management
weaknesses do not "staff out" responsibility for leading change. Top
leadership involvement and clear lines of accountability for making
management improvements are critical to overcoming organizations'
natural resistance to change, marshalling the resources needed in many
cases to improve management, and building and maintaining the
organizationwide commitment to new ways of doing business.
Results from our surveys show that while the majority of managers
continue to indicate top leadership demonstrates a strong commitment to
achieving results, we have not seen a noteworthy improvement in the
percentage of managers expressing this view. From our 1997 survey, we
estimated about 57 percent of managers overall reported such commitment
to a great or very great extent. On our 2003 survey, 62 percent of
managers expressed a comparable view--a higher but not statistically
significant increase. (See fig. 10.):
Figure 10: Percentage of Federal Managers Who Reported to a Great or
Very Great Extent Their Top Leadership Has a Strong Commitment to
Achieving Results:
[See PDF for image]
[End of figure]
As shown in figure 11, however, we continued to see a significant
difference between the perceptions of SES and non-SES managers on this
issue. That is, the percentage of SES managers reporting that top
leadership demonstrated strong commitment to a great or very great
extent in 2003 was 22 percent higher than for non-SES managers.
Figure 11: Percentage of SES and Non-SES Managers Who Reported to a
Great or Very Great Extent Their Agency Top Leadership Demonstrated
Strong Commitment to Achieving Results:
[See PDF for image]
[A] There was a statistically significant difference between SES and
non-SES.
[End of figure]
We observed in our 1997 and 2000 reports on governmentwide
implementation of GPRA that we would expect to see managers' positive
perceptions on items, such as the extent to which top leadership is
committed to achieving results, become more prevalent and the gap
between SES and non-SES managers begin to narrow as GPRA and related
reforms are implemented; however, these changes do not appear to be
happening as expected.[Footnote 37]
Demonstrating the willingness and ability to make decisions and manage
programs based on results and the ability to inspire others to embrace
such a model are important indicators of leadership commitment to
results-oriented management. However, in both our 1997 and 2000
surveys, only about 16 percent of managers reported that changes by
management above their levels to the programs for which they were
responsible were based on results or outcome-oriented performance
information to a great or very great extent. In our 2003 survey, this
indicator increased to 23 percent, a statistically significant increase
from prior surveys. Twenty-eight percent of federal managers surveyed
who expressed an opinion reported that the lack of ongoing top
executive commitment or support for using performance information to
make program/funding decisions hindered measuring performance or using
performance information to a great or very great extent.
Our interviews with 10 top political appointees from the Clinton and
current Bush administrations indicated a high level of support and
enthusiasm for effectively implementing the principles embodied in
GPRA. For example, one appointee noted that GPRA focused senior
management on a set of goals and objectives to allow the organization
to understand what is important and how to deal with accomplishment at
a macro-level, as well as provided a structure for problem solving.
Another political appointee noted that GPRA has made it important to
look at what you get out of the budget, not just what you put into it,
while another concluded that GPRA brought about a fundamental
rethinking of how they managed their programs and processes. Such
indications of support for GPRA are promising. However, to support the
transition to more results-oriented agency cultures, top agency
management will need to make a more concerted effort to translate their
enthusiasm for GPRA into actions that communicate to employees that top
management cares about performance results and uses the information in
its decision making.
The need for strong, committed leadership extends to OMB as well. OMB
has shown a commitment to improving the management of federal programs,
both through its leadership in reviewing agency program performance
using the PART tool as well as through the PMA, which calls for
improved financial performance, strategic human capital management,
competitive sourcing, expanded electronic government, and performance
budget integration. Using the foundation of information generated by
agencies in their strategic plans, annual performance plans, and
program performance reports, OMB has used the PART tool to exercise
oversight of selected federal programs by assessing program purpose and
design, the quality of strategic planning, the quality of program
management, and the extent to which programs can demonstrate results.
PART provides OMB a lens through which to view performance information
for use in the budget formulation process. PART, and OMB's use of
performance data in the budget formulation process, potentially can
complement GPRA's focus on increasing the supply of credible
performance information by promoting the demand for this information in
the budget formulation process. As we reported in chapter 2, more
federal managers noted that OMB was paying attention to their agencies'
efforts under GPRA. (See fig. 6.) Additionally, OMB convened a
performance measurement workshop in April 2003 to identify practical
strategies for addressing common performance measurement challenges. As
a result of this workshop, it produced a paper in June 2003 that
included basic performance measurement definitions and concepts and
common performance measurement problems that were discussed at the
workshop. This was part of OMB's continuing efforts to improve PART as
an evaluation tool.
However, there are areas where OMB could further enhance its
leadership. OMB has stated that the PART exercise presents an
opportunity to inform and improve on agency GPRA plans and reports and
establish a meaningful, systematic link between GPRA and the budget
process. OMB has instructed agencies that, in lieu of a performance
plan, they are to submit a performance budget that includes information
from the PART assessments, including all performance goals used in the
assessment of program performance done under the PART process. The
result is that program-specific performance measures developed through
the PART review are to substitute for other measures developed by the
agency through its strategic planning process. GPRA is a broad
legislative framework that was designed to be consultative with
Congress and other stakeholders and address the needs of many users of
performance information--Congress to provide oversight and inform
funding decisions, agency managers to manage programs and make internal
resource decisions, and the public to provide greater accountability.
Changing agency plans and reports for use in the budget formulation
process may not satisfy the needs of these other users. Users other
than OMB are not likely to find the information useful unless it is
credible and valid for their purposes. PART's program-specific focus
may fit with OMB's agency-by-agency budget reviews, but it is not well
suited to achieving one of the key purposes of strategic plans--to
convey agencywide, long-term goals and objectives for all major
functions and operations. PART's focus on program-specific measures
does not substitute for the strategic, long-term focus of GPRA on
thematic goals and department-and governmentwide crosscutting
comparisons.
To reach the full potential of performance management, agency planning
and reporting documents need to reflect the full array of uses of
performance information, which may extend beyond those needed for
formulating the President's Budget. However, it is not yet clear
whether the results of those reviews, such as changes to agencies'
program performance measures, will complement and be integrated with
the long-term, strategic goals and objectives agencies have established
in consultation with Congress and other stakeholders under GPRA. OMB
has not yet clearly articulated how PART is to complement GPRA. Focus
group participants suggested that the administration and OMB needed to
reinforce GPRA's usefulness as a management tool for agencies. They
also emphasized the need for OMB to help agencies understand how to
integrate GPRA with other management initiatives, such as PART.
As we noted in chapter 3, agencies' plans and reports still suffer from
persistent weaknesses and could improve in a number of areas, such as
attention to issues that cut across agency lines, and better
information about the quality of the data that underlie agency
performance goals. However, OMB's July 2003 guidance for the
preparation and submission of strategic plans, annual performance
plans, and annual performance reports is significantly shorter and less
detailed than its 2002 guidance. For example, OMB no longer provides
detailed guidance to agencies for the development of performance plan
components. OMB's 2002 guidance on the preparation and submission of
annual performance plans is approximately 39 pages long; in its 2003
guidance, that discussion spans only 2 pages. The 2003 guidance in this
area does not include entire sections found in the 2002 guidance, such
as principles for choosing performance goals and indicators for
inclusion in the annual plan, types of performance goals, crosscutting
programs, and requirements for verifying and validating data.
OMB needs to maintain and strengthen its leadership role in working
with agencies to help them produce the highest quality GPRA documents
through its formal guidance and reviews of strategic plan and report
submissions. Focus group participants discussed the need for consistent
guidance on how to implement GPRA. Furthermore, there is no evidence
that agencies have institutional knowledge of GPRA requirements that
would obviate the need for OMB's guidance. New managers will need a
consistent resource that provides practical guidance on what agencies
need to include in their planning and reporting documents to comply
with GPRA and reflect best practices. Consistent, explicit OMB guidance
on preparing GPRA documents can help ensure that gains in the quality
of GPRA documents are maintained and provide a resource for agencies to
make further improvements in those documents. For example, guidance on
how to discuss coordination of crosscutting programs or improvements to
the credibility of performance data in agency performance plans goes
hand-in-hand with OMB's enhanced oversight of agency performance
through the PART exercise.
The success of GPRA depends on the commitment of top leadership within
agencies, OMB, and Congress. Obtaining such leadership commitment
depends in part on the usefulness and relevance of agency goals and
strategies to these parties. GPRA requires an agency to develop a
strategic plan at least every 3 years to cover the following 5-year
period. Thus, there have been two required updates of strategic plans
since the initial strategic plans were submitted for fiscal year 1997-
-fiscal year 2000 and fiscal year 2003. The fiscal year 2000 update
occurred the year before a new presidential term began. According to
our focus group participants--both the experts and federal managers--it
makes little sense to require an update of a strategic plan shortly
before a new administration is scheduled to take office. For example,
changes in political leadership generally result in a new agenda with
new objectives. Such changes force agencies to revise their plans,
management initiatives, and strategies, which translates into
additional GPRA-related work. A strategic plan that does not reflect
the participation and buy-in of top administration leadership and key
congressional stakeholders is unlikely to be successfully implemented.
Therefore, GPRA's requirement to update agency strategic plans
according to a schedule that is out of sync with presidential and
congressional terms means that effort may be wasted on plans that lack
the support of top leadership.
Managers Report Mixed Results in Use of Performance Information:
GPRA's usefulness to agency leaders and managers as a tool for
management and accountability was cited as a key accomplishment
numerous times by focus group participants. However, a number of
alternative views indicated use of performance information for key
management decisions has been mixed. For example, one participant said
they did not believe GPRA has been used as a tool yet, while another
participant commented that only better managers take advantage of GPRA
as a management tool. According to focus group participants, although
many federal managers understand and use results-oriented management
concepts in their day-to-day activities, such as strategic planning and
performance measurement, they do not always connect these concepts to
the requirements of GPRA.
This view was strongly supported by our survey results. Prior to
mentioning GPRA in our survey, we asked federal managers the extent to
which they consider their agency's strategic goals when engaging in key
management tasks such as setting program activities, allocating
resources, or considering changes in their programs. A relatively high
percentage of managers--ranging from 66 to 79 percent--responded to a
great or very great extent. However, when we asked similar questions
about the extent to which they considered their agency's annual
performance goals as set forth in the agency's GPRA annual performance
plan for the same activities, the comparable responses were
considerably lower, ranging from 22 to 27 percent.
Because the benefit of collecting performance information is only fully
realized when this information is actually used by managers, we asked
them about the extent to which they used the information obtained from
measuring performance for various program management activities. As
shown in figure 12, for seven of the nine activities we asked about,
the majority of managers who expressed an opinion reported using
performance information to a great or very great extent in 2003. Across
all nine activities, the percentage of managers saying they used
performance information to a great or very great extent ranged from 41
percent for developing and managing contracts to 60 percent for
allocating resources, setting individual job expectations, and
rewarding staff. While we had observed a decline in the reported use of
performance information to this extent for many of these activities
between 1997 and 2000, our 2003 results increased to levels not
significantly different from 1997 for all but one category--adopting
new program approaches or changing work processes. This category of use
continued to be significantly lower at 56 percent in 2003 than it was
in 1997 at 66 percent. Although another category, coordinating program
efforts with other internal or external organizations, shows a similar
pattern of limited recovery, the difference between the 1997 and 2003
results is not statistically significant.
Figure 12: Percentage of Federal Managers Who Reported Using
Information Obtained from Performance Measurement to a Great or Very
Great Extent for Various Management Activities:
[See PDF for image]
Note: Percentages are based on those respondents answering on the
extent scale.
[A] There was a statistically significant difference between the 1997
and 2003 surveys.
[B] This question was not asked in 1997.
[End of figure]
We have reported that involving program managers in the development of
performance goals and measures is critical to increasing the relevance
and usefulness of this information to their day-to-day
activities.[Footnote 38] Yet, our survey data indicate that
participation in activities related to the development and use of
performance information has also been mixed. In 2003, only 14 percent
of managers believed to a great or very great extent that their
agencies considered their contributions to or comments on their
agency's GPRA plans or reports. However, significantly more SES
managers (43 percent) than non-SES managers (12 percent) expressed this
view. Also, when compared to our 2000 survey when we first asked this
question, the percentage of SES managers expressing this view in 2003
was significantly higher than in 2000 (32 percent). The percentage of
non-SES managers was essentially unchanged from 2000 (10 percent).
Furthermore, as shown in figure 13, overall around half or fewer of
managers responded "yes" on our 2003 survey to questions about being
involved in developing ways to measure whether program performance
goals are being achieved (46 percent), gathering and analyzing data to
measure whether programs were meeting their specific performance goals
(51 percent), or using measures for program performance goals to
determine if the agency's strategic goals were being achieved (43
percent). None of these overall results were significantly different
from our 1997 results. We did find, however, that significantly more
SES managers responded "yes" on the 2003 survey (72 percent) than the
1997 survey (55 percent) with regard to being involved in using
performance measurement information to determine if the agency's
strategic goals were being achieved when compared to our 1997 results.
Figure 13: Percentage of Federal Managers Responding "Yes" about Being
Involved in the Following Activities:
[See PDF for image]
[End of figure]
Managers Continue to Confront a Range of Human Capital Management
Challenges:
Managing people strategically and maintaining a highly skilled and
energized workforce that is empowered to focus on results are
critically important. Such human capital management practices are
essential to the success of the federal government in the 21st century
and to maximizing the value of its greatest asset--its people. Our
survey results showed continuing challenges related to the adequacy of
managerial decision making authority, training, and incentives.
Federal Managers Report That They Are Held Accountable for Program
Results but Do Not Have the Decision-Making Authority They Need to
Accomplish Agency Goals:
High-performing organizations seek to shift the focus of management and
accountability from activities and processes to contributions and
achieving results. In each of our three surveys, we asked managers
about the amount of decision-making authority they had and the degree
to which they were held accountable for results.
As shown in figure 14, for 2003, an estimated 40 percent of federal
managers overall reported that they had the decision-making authority
they needed to help the agency accomplish its strategic goals to a
great or very great extent. This was a statistically significant
increase over our 1997 estimate of 31 percent. While there were more
SES and non-SES managers expressing this view on our 2003 survey than
the 1997 survey, it was the non-SES managers that showed the
significant increase. Despite this promising trend, however, there
continued to be substantial differences in 2003, as well as on the two
previous surveys, between the responses of SES and lower-level managers
on this question. Compared to the 57 percent of SES managers who
reported having such authority to a great or very great extent in 2003,
only 38 percent of non-SES managers reported having such authority to a
great or very great extent.
Figure 14: Percentage of Federal Managers Reporting to a Great or Very
Great Extent That Managers/Supervisors at Their Levels Had the
Decision-Making Authority They Needed to Help the Agency Accomplish Its
Strategic Goals:
[See PDF for image]
[A] There was a statistically significant difference between the 1997
and 2003 surveys.
[B] There was a statistically significant difference between SES
compared to non-SES for each survey.
[End of figure]
However, when asked the extent to which managers or supervisors at
their levels were held accountable for the accomplishment of agency
strategic goals, 57 percent responded to a great or very great extent
in 2003. Unlike in other areas, where SES managers had significantly
different views from non-SES managers, there was little difference in
the area of accountability. (See fig. 15.):
Figure 15: Percentage of Federal Managers, SES, and Non-SES in 2003
Reporting to a Great or Very Great Extent That They Were Held
Accountable for the Accomplishment of Agency Strategic Goals:
[See PDF for image]
[End of figure]
This 57 percent is significantly higher than the 40 percent of managers
overall who indicated that they had comparable decision-making
authority. However, in contrast to the question on authority, as shown
in figure 14, where more SES managers than non-SES managers expressed
the view that they had the authority, there was little difference, as
shown in figure 15, between the two groups in their views about being
held accountable for achieving agency strategic goals to a great or
very great extent. As figures 14 and 15 further illustrate, roughly the
same percentage of SES managers perceived to a great or very great
extent that managers at their level had decision-making authority and
accountability for achieving agency strategic goals. This result
suggests that their authority was perceived to be on par with their
accountability. In contrast, only 38 percent of non-SES managers
perceived that managers at their levels had the decision-making
authority they needed to a great or very great extent, while 57 percent
perceived that they were held accountable to a comparable extent.
Managers are hard-pressed to achieve results when they do not have
sufficient authority to act. In our report containing the results of
our 1997 survey, we noted that agencies needed to concentrate their
efforts on areas where managers were not perceiving or experiencing
progress, such as that concerning devolving decision-making authority
to managers throughout their organizations. While authority for
achieving results appears to be in a modestly upward trend, the balance
between authority and accountability that fosters decision making to
achieve results could be further improved, particularly among non-SES
managers.
Fewer Than Half of Managers Reported Training on Key Tasks:
We previously reported on the need for agencies to expend resources on
effective training and professional development to equip federal
employees to work effectively.[Footnote 39] Among the resources focus
group participants cited as lacking included federal managers and staff
with competencies and skills needed to plan strategically, develop
robust measures of performance, and analyze what the performance data
mean. Our 2003 Guide calls for training and development efforts to be
strategically focused on improving performance toward the agency's
goals and put forward with the agency's organizational culture firmly
in mind.[Footnote 40] Throughout this process it is important that top
leaders in the agencies communicate that investments in training and
development are expected to produce clearly identified results. By
incorporating valid measures of effectiveness into the training and
development programs they offer, agencies can better ensure that they
will adequately address training objectives and thereby increase the
likelihood that desired changes will occur in the target population's
skills, knowledge, abilities, attitudes, or behaviors. Furthermore, if
managers understand and support the objectives of training and
development efforts, they can provide opportunities to successfully use
the new skills and competencies on the job and model the behavior they
expect to see in their employees.
In response to our 2003 survey, fewer than half of managers answered
"yes" when we asked them whether, during the past 3 years, their
agencies had provided, arranged, or paid for training that would help
them accomplish any of seven critical results-oriented management-
related tasks. However, progress is indicated in our survey results. As
shown in figure 16, more managers answered "yes" in 2003 on all seven
training areas than in previous surveys. These increases were
statistically significant for five of the tasks--setting program
performance goals, using program performance information to make
management decisions, linking program performance to the achievement of
agency strategic goals, and implementing the requirements of GPRA.
Figure 16: Percentage of Federal Managers in Each Survey Year Who
Reported That during the Past 3 Years Their Agencies Provided,
Arranged, or Paid for Training That Would Help Them Accomplish Specific
Tasks:
[See PDF for image]
[A] This question was not asked in the 1997 survey.
[B] There was a statistically significant difference between the 2000
and 2003 surveys.
[C] There was a statistically significant difference between the 1997
and 2003 surveys.
[End of figure]
As with our 2000 survey results, the 2003 survey results continued to
demonstrate that there is a positive relationship between agencies
providing training and development on setting program performance goals
and the use of performance information when setting or revising
performance goals. For those managers who responded "yes" to training
on setting performance goals, 60 percent also reported that they used
information obtained from performance measurement when setting new or
revising existing performance goals to a great or very great extent. In
contrast, for those managers who responded "no" to training on setting
performance goals, only 38 percent reported that they used information
obtained from performance measurement for setting new or revising
existing performance goals to a great or very great extent. The
difference between these percentages is statistically significant.
Effective training and development programs are an integral part of a
learning environment that can enhance the federal government's ability
to attract and retain employees with the skills and competencies needed
to achieve results. Training and developing new and current staff to
fill new roles and work in different ways will be a crucial part of the
federal government's endeavors to meet its transformation challenges.
Ways that employees learn and achieve results will also continue to
transform how agencies do business and engage employees in further
innovation and improvements.
Managers Perceive a Lack of Positive Recognition for Helping Agencies
Achieve Results:
Another fundamental aspect of the human capital management challenge
agencies face is providing the incentives to their employees to
encourage results-oriented management. Monetary and nonmonetary
incentives can be used as a method for federal agencies to reward
employees and to motivate them to focus on results.
Overall, an increasing but still small percentage of managers reported
in 1997, 2000, and 2003 that employees in their agencies received
positive recognition to a great or very great extent for helping
agencies accomplish their strategic goals. In 1997, 26 percent of
federal managers reported such an extent of positive recognition as
compared to 37 percent in 2003, a statistically significant increase.
Interestingly, this improvement is seen in the responses of non-SES
managers. As shown in figure 17, the percentage of SES managers
expressing this view stayed at about the same level over the three
surveys, while the percentage of non-SES managers holding this view was
significantly higher in 2003 than in 1997. Even with this improvement
on the part of the responses from non-SES managers, significantly more
SES managers (47 percent) than non-SES managers (36 percent) expressed
this perception to a comparable extent in 2003.
Figure 17: Percentage of Federal Managers Who Reported to a Great or
Very Great Extent That Employees in Their Agencies Received Positive
Recognition for Helping Their Agencies Accomplish Their Strategic
Goals:
[See PDF for image]
[A] There was a statistically significant difference between the 1997
and 2003 surveys.
[End of figure]
Unfortunately, most existing federal performance appraisal systems are
not designed to support a meaningful performance-based pay system in
that they fail to link institutional, program, unit, and individual
performance measurement and reward systems. In our view, one key need
is to modernize performance management systems in executive agencies so
that they link to the agency's strategic plan, related goals, and
desired outcomes and are therefore capable of adequately supporting
more performance-based pay and other personnel decisions.
We have reported federal agencies can develop effective performance
management systems by implementing a selected, generally consistent set
of key practices. These key practices helped public sector
organizations both in the United States and abroad create a clear
linkage--"line of sight"--between individual performance and
organizational success and, thus, transform their cultures to be more
results oriented, customer-focused, and collaborative in nature.
Examples of such practices include:
* aligning individual performance expectations with organizational
goals,
* connecting performance expectations to crosscutting goals,
* linking pay to individual and organizational performance, and:
* making meaningful distinctions in performance.[Footnote 41]
Beyond implementing these key practices, high-performing organizations
understand that their employees are assets whose value to the
organization must be recognized, understood, and enhanced. They view an
effective performance management system as an investment to maximize
the effectiveness of people by developing individual potential to
contribute to organizational goals. To maximize this investment, an
organization's performance management system is designed, implemented,
and continuously assessed by the standard of how well it helps the
employees help the organization achieve results and pursue its mission.
Persistent Challenges in Setting Outcome-Oriented Goals, Measuring
Performance, and Collecting Useful Data:
In prior reports, we have described difficulties faced by federal
managers in developing useful, outcome-oriented measures of performance
and collecting data indicating progress achieved.[Footnote 42] One of
the most persistent challenges has been the development of outcome-
oriented performance measures. Additionally, it is difficult to
distinguish the impact of a particular federal program from the impact
of other programs and factors, thus making it difficult to attribute
specific program performance to results. The lack of timely and useful
performance information can also hinder GPRA implementation.
Meaningful, Outcome-Oriented Performance Measures Are Sometimes Hard to
Develop:
In the past, we have noted that federal managers found meaningful
performance measures difficult to develop. Focus group participants and
survey respondents noted that outcome-oriented performance measures
were especially difficult to establish when the program or line of
effort was not easily quantifiable. The challenge of the "complexity of
establishing outcome-oriented goals and measuring performance" was
cited by six of the eight focus groups as one of the key challenges
that managers face in implementing GPRA. Focus group participants
agreed that they often felt as if they were trying to measure the
immeasurable, not having a clear understanding of which performance
indicators could accurately inform the agency how it is carrying out a
specific activity. Managers from agencies engaged in basic science
research and development and grant-making functions noted that this
effort was particularly difficult for them because federal programs,
especially those that are research-based, often take years to achieve
the full scope of their goals. On our most recent survey, we estimated
that 36 percent of federal managers who had an opinion indicated that
the determination of meaningful measures hindered the use of
performance information or performance measurement to a great or very
great extent. While this number was significantly lower than the
percentage of managers expressing the comparable view on the 1997 or
2000 survey and may reflect some lessening of this as a hindrance to
some managers, it nonetheless continues to be among those items having
the largest percentage of managers citing it as a substantial
hindrance.
Impact of Federal Programs Difficult to Discern:
In our June 1997 report on GPRA, we noted that "the often limited or
indirect influence that the federal government has in determining
whether a desired result is achieved complicates the effort to identify
and measure the discrete contribution of the federal initiative to a
specific program result."[Footnote 43] This occurs primarily because
many federal programs' objectives are the result of complex systems or
phenomena outside the program's control. In such cases, it is
particularly challenging for agencies to confidently attribute changes
in outcomes to their program--the central task of program impact
evaluation. This is particularly challenging for regulatory programs,
scientific research programs, and programs that deliver services to
taxpayers through third parties, such as state and local governments.
We have reported that determining the specific outcomes resulting from
federal research and development has been a challenge that will not be
easily resolved.[Footnote 44] Due to the difficulties in identifying
outcomes, research and development agencies typically have chosen to
measure a variety of proxies for outcomes, such as the number of
patents resulting from federally funded research, expert review and
judgments of the quality and importance of research findings, the
number of project-related publications or citations, and contributions
to expanding the number of research scientists.
We have also reported that implementing GPRA in a regulatory
environment is particularly challenging.[Footnote 45] Although federal
agencies are generally required to assess the potential benefits and
costs of proposed major regulatory actions, they generally do not
monitor the benefits and costs of how these and other federal programs
have actually performed. For example, in the case of the Environmental
Protection Agency (EPA), to determine if existing environmental
regulations need to be retained or improved, we previously recommended
that EPA study the actual costs and benefits of such
regulations.[Footnote 46]
In the past, regulatory agencies have cited numerous barriers to their
efforts to establish results-oriented goals and measures. These
barriers included problems in obtaining data to demonstrate results,
accounting for factors outside of the agency's control that affect
results, and dealing with the long time periods often needed to see
results. Our prior work discussed best practices for addressing
challenges to measuring the results of regulatory programs. In
particular, to address the challenge of discerning the impact of a
federal program, when other factors also affect results, we suggested
agencies "establish a rationale of how the program delivers results."
Establishing such a rationale involves three related practices: (1)
taking a holistic or "systems" approach to the problem being addressed,
(2) building a program logic model that described how activities
translated to outcomes, and (3) expanding program assessments and
evaluations to validate the model linkages and rationale.
We have also reported on the difficulties encountered in meeting GPRA
reporting requirements for intergovernmental grant programs.[Footnote
47] Programs that do not deliver a readily measurable product or
service are likely to have difficulty meeting GPRA performance
measurement and reporting requirements. Intergovernmental grant
programs, particularly those with the flexibility inherent in classic
block grant design, may be more likely to have difficulty producing
performance measures at the national level and raise delicate issues of
accountability. Although most flexible grant programs we reviewed
reported simple activity or client counts, relatively few of them
collected uniform data on the outcomes of state or local service
activities. Collecting such data requires conditions (such as
uniformity of activities, objectives, and measures) that do not exist
under many flexible program designs, and even where overall performance
of a state or local program can be measured, the amount attributable to
federal funding often cannot be separated out.
Focus group participants also suggested that they faced challenges in
obtaining timely performance data from relevant partner organizations
and in identifying what the federal government's contribution has been
to a specific outcome. Furthermore, survey respondents provided some
corroboration for these views. Across all three of our surveys, we
estimate that roughly a quarter of all federal managers reported this
difficulty--distinguishing between the results produced by the program
they were involved with and results caused by other factors--as a
substantial hindrance. In response to a survey question about what the
federal government could do to improve its overall focus on managing
for results, one respondent noted: "Defining meaningful measures for
the work we do is extremely difficult; and even if they could be
defined, performance and accomplishment is (sic) dependent on so many
factors outside our control that it is difficult, if not impossible, to
make valid conclusions.":
Timely, Useful Performance Information Not Always Available:
In February 2000, we reported that intergovernmental programs pose
potential difficulties in collecting timely and consistent national
data.[Footnote 48] We also noted that agencies had limited program
evaluation capabilities and weaknesses in agencies' financial
management capabilities make it difficult for decision makers to
effectively assess and improve many agencies' financial performance. On
the basis of our current findings, these issues still exist. Federal
managers who participated in our focus groups cited difficulties in
gathering data from state or local entities, as well as statutory
limitations regarding the nature and breadth of data that they were
permitted to collect. However, in our 2003 survey, only 27 percent of
federal managers indicated that obtaining data in time to be useful was
a substantial hindrance; 31 percent expressed a comparable view with
regard to obtaining valid or reliable data.
Focus group participants also noted that OMB's accelerated time frames
for reporting performance information will contribute to the challenge
of producing complete, timely information in their agencies'
performance reports. Over the past 2 fiscal years, OMB has moved the
deadline for submission of agencies' performance reports (now
performance and accountability reports) back from the statutory
requirement of March 31; for fiscal year 2003 data, the deadline is
January 30, 2004. In fiscal year 2004, these reports will be due on
November 15, 2004. According to the managers, individual agencies may
work on different time frames based partially on the population they
serve or the stakeholders they must work with, such as state or local
agencies. This "one size fits all" approach does not take such
differences into account.
Additionally, OMB requires agencies to report on their performance data
quarterly; managers noted that this was particularly difficult for
outcomes that may be achieved over extended periods of time, such as
outcomes associated with basic science. As we have previously reported,
measuring the performance of science-related projects can be difficult
because a wide range of factors determine if and how a particular
research and development project will result in a commercial
application or have other benefits. Efforts to cure diseases or pursue
space exploration are difficult to quantify and break down into
meaningful quarterly performance measures.
Crosscutting Issues Hinder Successful GPRA Implementation:
Crosscutting issues continue to be a challenge to GPRA implementation.
Mission fragmentation and program overlap are widespread across the
federal government. Moreover, addressing this challenge is essential to
the success of national strategies in areas such as homeland security,
drug control, and the environment.
We have reported that agencies could use the annual performance
planning cycle and subsequent annual performance reports to highlight
crosscutting program efforts and to provide evidence of the
coordination of those efforts. Our review of six agencies' strategic
and annual performance plans showed some improvement in addressing
their crosscutting program efforts, but a great deal of improvement is
still necessary. Few of the plans we reviewed attempted the more
challenging task of discussing planned strategies for coordination and
establishing complementary performance goals and complementary or
common performance measures. For example, SSA's 2004 performance plan
makes some mention of the agency's efforts to coordinate with other
agencies to preserve the integrity of the Social Security number as a
personal identifier, but there are very few details about this
important component of its mission.
Previous GAO reports and agency managers identified several barriers to
interagency coordination. First, missions may not be mutually
reinforcing or may even conflict, making reaching a consensus on
strategies and priorities difficult. In 1998 and 1999, we found that
mission fragmentation and program overlap existed in 12 federal mission
areas, ranging from agriculture to natural resources and the
environment. Implementation of federal crosscutting programs is often
characterized by numerous individual agency efforts that are
implemented with little apparent regard for the presence of related
activities. Second, we reported on agencies' interest in protecting
jurisdiction over missions and control over resources.[Footnote 49]
Focus group participants echoed this concern, noting that there can be
"turf battles" between agencies, where jurisdictional boundaries, as
well as control over resources, are hotly contested. Finally,
incompatible procedures, processes, data, and computer systems pose
difficulties for agencies to work across agency boundaries. For
example, we reported how the lack of consistent data on federal
wetlands programs implemented by different agencies prevented the
government from measuring progress toward achieving the governmentwide
goal of no net loss of the nation's wetlands.[Footnote 50]
We have previously reported and testified that GPRA could provide OMB,
agencies, and Congress with a structured framework for addressing
crosscutting program efforts.[Footnote 51] OMB, for example, could use
the provision of GPRA that calls for OMB to develop a governmentwide
performance plan to integrate expected agency-level performance.
Unfortunately, this provision has not been fully implemented. OMB
issued the first and only such plan in February 1998 for fiscal year
1999. In our review of the plan,[Footnote 52] we found that it included
a broad range of governmentwide management objectives and a mission-
based presentation of key performance goals based on agency performance
plans and the plan's framework should ultimately allow for a cohesive
presentation of governmentwide performance. However, the specific
contents of this initial plan did not always deliver an integrated,
consistent, and results-oriented picture of fiscal year 1999 federal
government performance goals.
OMB officials we interviewed at the time stressed that developing the
governmentwide plan was viewed as an essential and integral component
of the President's budget and planning process. From OMB's perspective,
both the plan and the budget submission were intended to serve as
communication tools for a range of possible users. In their opinion,
the plan added value by reflecting a governmentwide perspective on
policy choices made throughout the budget formulation process. OMB
acknowledged that the plan itself did not serve to change the process
through which decisions on government priorities were made, but
enhanced it by placing a greater emphasis on results. As one official
described it, the governmentwide performance plan was a derivative
document, reflecting the budget and management decisions made
throughout the process of formulating the President's budget
submission. However, we found that focusing broadly on governmentwide
outcomes should be a central and distinguishing feature of the federal
government performance plan. To be most effective and supportive of the
purposes of GPRA, the governmentwide plan must be more than a
compilation of agency-level plans; integration, rather than repetition,
must be its guiding principle.
OMB has not issued a distinct governmentwide performance plan since
fiscal year 1999. Most recently, the President's fiscal year 2004
budget focused on describing agencies' progress in addressing the PMA
and the results of PART reviews of agency programs. Although such
information is important and useful, it does not provide a broader and
more integrated perspective of planned performance on governmentwide
outcomes. Additionally, the fiscal year 2004 budget identified budget
requests and performance objectives by agency, such as the U.S.
Department of Defense, as opposed to crosscutting governmentwide
themes. From this presentation, one could assume that the only
activities the U.S. government planned to carry out in support of
national defense were those listed under the chapter "Department of
Defense." However, the chapter of the fiscal year 2004 budget
discussing "the Department of State and International Assistance
Programs," contains a heading titled, "Countering the Threat from
Weapons of Mass Destruction." And while OMB may have a technical reason
for not classifying this task as being related to national defense or
homeland security, it is unclear that a lay reader could make that
distinction. The fiscal year 2005 budget also identified budget
requests by agency, not by crosscutting theme. Without such a
governmentwide focus, OMB is missing an opportunity to assess and
communicate the relationship between individual agency goals and
outcomes that cut across federal agencies and more clearly relate and
address the contributions of alternative federal strategies. The
governmentwide performance plan also could help Congress and the
executive branch address critical federal performance and management
issues, including redundancy and other inefficiencies in how we do
business. It could also provide a framework for any restructuring
efforts.
A strategic plan for the federal government, supported by key national
indicators to assess the government's performance, position, and
progress, could provide an additional tool for governmentwide
reexamination of existing programs, as well as proposals for new
programs. If fully developed, a governmentwide strategic plan could
potentially provide a cohesive perspective on the long-term goals of
the federal government and provide a much needed basis for fully
integrating, rather than merely coordinating, a wide array of federal
activities. Successful strategic planning requires the involvement of
key stakeholders. Thus, it could serve as a mechanism for building
consensus. Further, it could provide a vehicle for the President to
articulate long-term goals and a road map for achieving them. In
addition, a strategic plan could provide a more comprehensive framework
for considering organizational changes and making resource decisions.
Developing a strategic plan for the federal government would be an
important first step in articulating the role, goals, and objectives of
the federal government. It could help provide critical horizontal and
vertical linkages. Horizontally, it could integrate and foster
synergies among components of the federal government as well as help to
clarify the role of the federal government vis-à-vis other sectors of
our society. Vertically, it could provide a framework of federal
missions and goals within which individual federal agencies could align
their own missions and goals that would cascade down to individual
employees. The development of a set of key national indicators could be
used as a basis to inform the development of the governmentwide
strategic and annual performance plans. The indicators could also link
to and provide information to support outcome-oriented goals and
objectives in agency-level strategic and annual performance plans.
Managers View Congress' Use of Performance Information as Limited:
Focus group members believed that one of the main challenges to GPRA
implementation was the reluctance of Congress to use that information
when making decisions, especially appropriations decisions. This
concern was cited as a significant challenge in each of the focus
groups, and was one of the top three "challenges" in five of the eight
focus groups. In some cases, managers in our focus groups noted that
this lack of usage was a significant disincentive to doing a good job
in preparing GPRA plans and reports. Agency managers made the following
criticisms regarding the perceived lack of congressional use of
performance information:
* appropriators have not bought into GPRA, so there is no incentive to
do this well,
* failure of congressional leadership in developing and using
performance measures,
* appropriators do not use performance data or tools to make decisions,
and:
* GPRA does not drive public policy decisions.
Results from our survey provide some further information in support of
this view. On our 2003 survey, when we asked federal managers about the
extent to which they thought congressional committees paid attention to
agency efforts under GPRA, only 22 percent of federal managers
responded in the great to very great categories. This result was not
significantly different from the results we observed on our 2000 survey
when we asked this question about three specific types of congressional
committees--authorization, appropriation, and oversight. On the 2000
survey, only 18 percent of federal managers held a similar view
concerning authorizing committees, 19 percent for appropriations
committees, and 20 percent for oversight committees. As we noted
earlier, when this item was asked in relation to OMB, there was a
significant increase in the percentage of managers responding to a
great or very great extent from 2000 to 2003. The 31 percent of
managers who viewed OMB as paying attention to a great or very great
extent in 2003 was significantly higher than the 22 percent holding a
comparable view of congressional committees.
Although managers expressed these concerns about the use of this
information, a recent review by the CRS suggested that Congress uses
performance information to some extent, as evidenced by citations in
legislation and committee reports.[Footnote 53] For example, in the
106th Congress (1999-2000), 42 public laws contained statutory language
relating to GPRA and performance measures, and 118 legislative
reports[Footnote 54] contained GPRA-associated passages. As shown in
figure 18, across all three of our surveys, only a minority of federal
managers governmentwide viewed the lack of ongoing congressional
commitment for using performance information as a hindrance to a great
or very great extent.
Figure 18: Percentage of Federal Managers Reporting to a Great or Very
Great Extent That a Lack of Ongoing Congressional Commitment or Support
for Using Performance Information in Making Program/Funding Decisions
Is a Hindrance:
[See PDF for image]
Note: Percentages are based on those respondents answering on the
extent scale.
[End of figure]
While there is concern regarding Congress' use of performance
information, it is important to make sure that this information is
initially useful. One of GPRA's purposes is to respond to a need for
accurate, reliable information for congressional decision making. In
2000, we reported that congressional staffs stated that they were
looking for recurring information on spending priorities within
programs; the quality, quantity, and efficiency of program operations;
the populations served or regulated; as well as programs' progress in
meeting their objectives.[Footnote 55] For example, learning who
benefits from a program can help in addressing questions about how well
services are targeted to those most in need. Some of these recurring
needs were met through formal agency documents, such as annual
performance plans. However, some information the agencies provided did
not fully meet the congressional staffs' needs because the presentation
was not clear, directly relevant, or sufficiently detailed. For
example, congressional staffs wanted to see more direct linkages among
the agencies' resources, strategies, and goals. In other cases, the
information was not readily available to the congressional staffs,
either because it had not been requested or reported, or because staff
were not informed that it was available.
As a key user of performance information, Congress also needs to be
considered a partner in shaping agency goals at the outset. For
example, through the strategic planning requirement, GPRA requires
federal agencies to consult with Congress and key stakeholders to
reassess their missions and long-term goals as well as the strategies
and resources they will need to achieve their goals. GPRA also provides
a vehicle for Congress to explicitly state its performance expectations
in outcome-oriented terms when establishing new programs or in
exercising oversight of existing programs that are not achieving
desired results. Congress could use authorizing and appropriations
hearings to determine if agency programs have clear performance goals,
measures, and data with which to track progress and whether the
programs are achieving their goals. If goals and objectives are unclear
or not results oriented, Congress could use legislation to articulate
the program outcomes it expects agencies to achieve. This would provide
important guidance to agencies that could then be incorporated in
agency strategic and annual performance plans.
[End of section]
Chapter 5: Conclusions and Recommendations:
Agenda for Achieving a Sustainable, Governmentwide Focus on Results:
As we have shown in this report, in the 10 years since the enactment of
GPRA, significant progress has been made in instilling a focus on
results in the federal government. First, GPRA statutory requirements
laid a foundation for results-oriented management in federal agencies.
Expert and agency focus group participants cited the creation of this
statutory foundation as one of the key accomplishments of GPRA. Since
GPRA began to be implemented governmentwide in fiscal year 1997, we
have observed significant increases in the percentage of federal
managers who reported having results-oriented performance measures for
their programs. Focus group participants' views on whether GPRA has had
a positive effect on the federal government's ability to deliver
results to the American public were mixed. For example, the information
gathered and reported for GPRA allows agencies to make better-informed
decisions, which improves their ability to achieve results. In
addition, GPRA has made the results of federal programs more
transparent to the public. Other participants stated that while certain
aspects of GPRA-related work have been positive, agencies' ability to
deliver results and public awareness of their activities cannot be
exclusively attributed to GPRA.
Second, GPRA has increased the connection between resources and results
by creating more formal linkages between agency performance goals and
objectives and the program activities in the budget. Over the first 4
years of agency efforts to implement GPRA, we observed that agencies
continued to tighten the required linkage between their performance
plans and budget requests. However, much remains to be done in this
area. For example, we have not observed notable increases in federal
managers' perceptions about their personal use of plans or performance
information when allocating resources, or about the use of performance
information when funding decisions are made about their programs.
However, it should be noted that we estimate a majority have positive
perceptions about the use of performance information to allocate
resources.
Third, GPRA has provided a foundation for examining agency missions,
performance goals and objectives, and the results achieved. We have
seen improvements in the quality of agency strategic plans, annual
performance plans, and performance reports since initial efforts.
However, few of the six agencies we reviewed in this report produced
GPRA planning and reporting documents that met all of our criteria for
the highest level of quality. Most of these agencies continued to miss
opportunities to present clear pictures of their intended and actual
performance results in their GPRA plans and reports and to show how
resources are aligned with actual performance results. Furthermore,
most of the agencies we reviewed did not provide a full level of
confidence in the credibility of their performance data.
Performance-based management, as envisioned by GPRA, requires
transforming organizational cultures to improve decision making,
maximize performance, and assure accountability. This transformation is
not an easy one and requires investments of time and resources as well
as sustained leadership commitment and attention. Challenges to
successful implementation of GPRA include inconsistent top leadership
commitment to creating a focus on results; an approach to setting goals
and developing strategies for achieving critical outcomes that creates
individual agency stovepipes rather than an integrated, holistic
governmentwide approach; getting federal managers to make greater use
of performance information to manage their programs and providing them
authority to act that is commensurate with their accountability for
results; difficulty in establishing meaningful measures of outcomes and
assessing results of federal programs that are carried out by
nonfederal entities; and untimely performance data.
The challenges identified in this report are not new--most have not
changed significantly since we first reported on governmentwide
implementation of GPRA. However, we have frequently reported on
approaches that agencies, OMB, and Congress could use to address the
challenges. These approaches include strengthening the commitment of
top leadership to creating and sustaining a focus on results; taking a
governmentwide approach to achieving outcomes that are crosscutting in
nature; improving the usefulness of performance information to
managers, Congress, and the public; and improving the quality of
performance measures and data. Collectively, these approaches form the
agenda that federal agencies, OMB, and Congress will need to follow to
bring about a more sustainable, governmentwide focus on results.
Strengthening Top Leadership Commitment to Creating and Sustaining
Results-Oriented Cultures:
Successfully addressing the challenges that federal agencies face
requires leaders who are committed to achieving results, who recognize
the importance of using results-oriented goals and quantifiable
measures, and who integrate performance-based management into the
culture and day-to-day activities of their organizations. Top
leadership must play a critical role in creating and sustaining high-
performing organizations. Without the clear and demonstrated commitment
of agency top leadership--both political and career--organizational
cultures will not be transformed, and new visions and ways of doing
business will not take root.
To be positioned to address the array of challenges faced by our
national government, federal agencies will need to transform their
organizational cultures so that they are more results oriented,
customer-focused, and collaborative. Leading public organizations here
in the United States and abroad have found that strategic human capital
management must be the centerpiece of any serious change management
initiative and efforts to transform the cultures of government
agencies. Performance management systems are integral to strategic
human capital management. Such systems can be key tools to maximizing
performance by aligning institutional performance measures with
individual performance and creating a "line of sight" between
individual and organizational goals. Leading organizations use their
performance management systems as a key tool for aligning
institutional, unit, and employee performance; achieving results;
accelerating change; managing the organization day to day; and
facilitating communication throughout the year so that discussions
about individual and organizational performance are integrated and
ongoing.[Footnote 56]
Furthermore, achieving this cultural transformation requires people to
have the knowledge and skills to develop and use performance
information to improve program performance. Our survey data indicated a
significant relationship between those managers who reported they
received training on setting performance goals and those who used
performance information when setting or revising performance goals.
However, federal agencies have not consistently showed a commitment to
investing in needed training and development opportunities to help
ensure that managers and employees have the requisite skills and
competencies to achieve agency goals.
The commitment to focusing on and using performance information needs
to extend to OMB and Congress as well. Through the administration's PMA
and PART initiatives, OMB has clearly placed greater emphasis on
management issues over the past several years. However, the focus of
such oversight needs to extend beyond the emphasis on formulating the
President's Budget to include an examination of the many challenges
agencies face that may be contributing to poor performance. In spite of
the persistent weaknesses we found in agencies' strategic plans and
annual performance plans and reports, OMB significantly reduced the
scope of its guidance to agencies on how to prepare these documents. By
emphasizing a focus on resource allocation through its PART exercise
and providing less information on how to comply with GPRA, OMB may be
sending a message to agencies that compliance with GPRA is not
important. Without strong leadership from OMB, the foundation of
performance information that has been built could deteriorate.
OMB leadership is critical to addressing the continuing challenges
presented in GPRA implementation and the transformation of the federal
government to an increasingly results-oriented culture. OMB, as the
primary focal point for overall management in the federal government,
can provide the needed impetus by providing guidance, fostering
communication among agencies, and forming intragovernmental councils
and work groups tasked with identifying potential approaches and
solutions to overcoming the persistent challenges to results-oriented
management.
Congress can also play a decisive role in fostering results-oriented
cultures in the federal government by using information on agency goals
and results at confirmation, oversight, authorization, and
appropriation hearings. Consistent congressional interest in the status
of an agency's GPRA efforts, performance measures, and uses of
performance information to make decisions, will send an unmistakable
message to agencies that Congress expects GPRA to be thoroughly
implemented.
We also found that timing issues may affect the development of agency
strategic plans that are meaningful and useful to top leadership. The
commitment of top leadership within agencies, OMB, and Congress is
critical to the success of strategic planning efforts. A strategic plan
should reflect the policy priorities of an organization's leaders and
the input of key stakeholders if it is to be an effective management
tool. However, GPRA specifies time frames for updating strategic plans
that do not correspond to presidential or congressional terms. As a
result, an agency may be required to update its strategic plan a year
before a presidential election and without input from a new Congress.
If a new president is elected, the updated plan is essentially moot and
agencies must spend additional time and effort revising it to reflect
new priorities. Our focus group participants, including GPRA experts,
strongly agreed that this timing issue should be addressed by adjusting
time frames to correspond better with presidential and congressional
terms.
Addressing Governmentwide Needs:
Mission fragmentation and program overlap are widespread throughout the
federal government.[Footnote 57] We have noted that interagency
coordination is important for ensuring that crosscutting program
efforts are mutually reinforcing and efficiently implemented. Our
review of six agencies' strategic and annual performance plans along
with our previous work on crosscutting issues has demonstrated that
agencies' still present their goals and strategies in a mostly
stovepiped manner. They have generally not used their plans to
communicate the nature of their coordination with other agencies, in
terms of the development of common or complementary goals and
objectives or strategies jointly undertaken to achieve those goals.
We have also reported that GPRA could provide a tool to reexamine
federal government roles and structures governmentwide. GPRA requires
the President to include in his annual budget submission a federal
government performance plan. Congress intended that this plan provide a
"single cohesive picture of the annual performance goals for the fiscal
year." The governmentwide performance plan could help Congress and the
executive branch address critical federal performance and management
issues, including redundancy and other inefficiencies in how we do
business. It could also provide a framework for any restructuring
efforts. Unfortunately, this provision has not been fully implemented.
Instead, OMB has used the President's Budget to present high-level
information about agencies and certain program performance issues. The
agency-by-agency focus of the budget does not provide the integrated
perspective of government performance envisioned by GPRA.
If the governmentwide performance plan were fully implemented, it could
provide a framework for such congressional oversight. For example, in
recent years, OMB has begun to develop common measures for similar
programs, such as job training. By focusing on broad goals and
objectives, oversight could more effectively cut across organization,
program, and other traditional boundaries. Such oversight might also
cut across existing committee boundaries, which suggests that Congress
may benefit from using specialized mechanisms to perform oversight
(i.e., joint hearings and special committees).
A strategic plan for the federal government, supported by key national
indicators to assess the government's performance, position, and
progress, could provide an additional tool for governmentwide
reexamination of existing programs, as well as proposals for new
programs. If fully developed, a governmentwide strategic plan can
potentially provide a cohesive perspective on the long-term goals of
the federal government and provide a much needed basis for fully
integrating, rather than merely coordinating, a wide array of federal
activities. Successful strategic planning requires the involvement of
key stakeholders. Thus, it could serve as a mechanism for building
consensus. Further, it could provide a vehicle for the President to
articulate long-term goals and a road map for achieving them. In
addition, a strategic plan can provide a more comprehensive framework
for considering organizational changes and making resource decisions.
Developing a strategic plan for the federal government would be an
important first step in articulating the role, goals, and objectives of
the federal government. It could help provide critical horizontal and
vertical linkages. Horizontally, it could integrate and foster
synergies among components of the federal government as well as help to
clarify the role of the federal government vis-à-vis other sectors of
our society. Vertically, it could provide a framework of federal
missions and goals within which individual federal agencies could align
their own missions and goals that would cascade down to individual
employees. The development of a set of key national indicators could be
used as a basis to inform the development of governmentwide strategic
and annual performance plans. The indicators could also link to and
provide information to support outcome-oriented goals and objectives in
agency-level strategic and annual performance plans.
Improving Usefulness of Performance Information:
We have found that leading organizations that progressed the farthest
to results-oriented management did not stop after strategic planning
and performance measurement. They applied their acquired knowledge and
data to identify gaps in their performance, report on that performance,
and finally use that information to improve their performance to better
support their missions.
Performance data can have real value only if they are used to identify
the gap between an organization's actual performance level and the
performance level it has identified as its goal. Once the performance
gaps are identified for different program areas, managers can determine
where to target their resources to improve overall mission
accomplishment. When managers are forced to reduce their resources, the
same analysis can help them target reductions to keep to a minimum the
impact on their organization's overall mission.[Footnote 58]
Under GPRA, agencies produce a single strategic plan, annual
performance plan, and annual performance report. However, there are
many potential consumers of agencies' performance information--
Congress, the public, and the agency itself. One size need not fit all.
Clearly, an agency will need more detailed information on its programs
for operational purposes than would be suitable for external audiences.
Of the six agencies' performance reports we reviewed, some of them
provided useful summary tables or information that provided overall
snapshots of performance or highlighted progress in achieving key
goals. Other reports that lacked such a summary made it difficult to
assess the progress achieved.
To improve the prospect that agency performance information will be
useful to and used by these different users, agencies need to consider
the different information needs and how to best tailor their
performance information to meet those needs. For example, we have
reported that, although many information needs were met, congressional
staff also identified gaps in meeting their information needs.[Footnote
59] Key to addressing these information gaps was improving
communication between congressional staff and agency officials to help
ensure that congressional information needs are understood, and that,
where feasible, arrangements are made to meet them. Improved two-way
communication might also make clear what information is and is not
available, as well as what is needed and what is not needed. This might
entail the preparation of simplified and streamlined plans and reports
for Congress and other external users.
Another challenge that limits the usefulness of agency performance
reports is the lack of timely data on performance. For the six
performance reports we reviewed we continued to observe a significant
number of goals for which performance data were unavailable. Policy
decisions made when designing federal programs, particularly
intergovernmental programs, may make it difficult to collect timely and
consistent national data. In administering programs that are the joint
responsibility of state and local governments, Congress and the
executive branch continually balance the competing objectives of
collecting uniform program information to assess performance with
giving states and localities the flexibility needed to effectively
implement intergovernmental programs.
Improving Performance Measures and Data Quality:
Another key challenge to achieving a governmentwide focus on results is
that of developing meaningful, outcome-oriented performance goals and
collecting performance data that can be used to assess results.
Performance measurement under GPRA is the ongoing monitoring and
reporting of program accomplishments, particularly progress toward
preestablished goals. It tends to focus on regularly collected data on
the level and type of program activities, the direct products and
services delivered by the programs, and the results of those
activities. For programs that have readily observable results or
outcomes, performance measurement may provide sufficient information to
demonstrate program results. In some programs, however, outcomes are
not quickly achieved or readily observed, or their relationship to the
program is uncertain. In such cases, more in-depth program evaluations
may be needed, in addition to performance measurement, to examine the
extent to which a program is achieving its objectives.
Given the difficult measurement challenges we have identified, it is
all the more important that agency strategic planning efforts include
the identification of the most critical evaluations that need to take
place to address those challenges. However, our previous work has
raised concerns about the capacity of federal agencies to produce
evaluations of program effectiveness.[Footnote 60] Few of the agencies
we reviewed deployed the rigorous research methods required to
attribute changes underlying outcomes to program activities. Yet we
have also seen how some agencies have profitably drawn on systematic
program evaluations to improve their measurement of program performance
or understanding of performance and how it might be improved.[Footnote
61] Our review of six agencies' strategic plans and performance reports
in this report revealed weaknesses in their discussions of program
evaluation. Most of the strategic plans lacked critical information
required by GPRA, such as a discussion of how evaluations were used to
establish strategic goals or a schedule of future evaluations.
Furthermore, two of the six performance reports did not summarize the
results of program evaluations completed that year, as required.
Our work has also identified substantial, long-standing limitations in
agencies' abilities to produce credible data and identify performance
improvement opportunities that will not be quickly or easily
resolved.[Footnote 62] According to our review, five of six agencies'
annual performance plans showed meaningful improvements in how they
discussed the quality of performance data. However, only DOT's
performance plan and report contained information that provided a full
level of confidence in the credibility of its performance data. In
particular, the plans and reports did not always provide detailed
information on how the agencies verified and validated their
performance data.
Recommendations for Executive Action:
To provide a broader perspective and more cohesive picture of the
federal government's goals and strategies to address issues that cut
across executive branch agencies, we recommend that the Director of OMB
fully implement GPRA's requirement to develop a governmentwide
performance plan.
To achieve the greatest benefit from both GPRA and PART, we recommend
that the Director of OMB articulate and implement an integrated and
complementary relationship between the two. GPRA is a broad legislative
framework that was designed to be consultative with Congress and other
stakeholders, and allows for varying uses of performance information.
PART looks through a particular lens for a particular use--the
executive budget formulation process.
To improve the quality of agencies' strategic plans, annual performance
plans, and performance reports and help agencies meet the requirements
of GPRA, we recommend that the Director of OMB provide clearer and more
consistent guidance to executive branch agencies on how to implement
GPRA. Such guidance should include standards for communicating key
performance information in concise as well as longer formats to better
meet the needs of external users who lack the time or expertise to
analyze lengthy, detailed documents.
To help address agencies' performance measurement challenges, we
recommend that the Director of OMB engage in a continuing dialogue with
agencies about their performance measurement practices with a
particular focus on grant-making, research and development, and
regulatory functions to identify and replicate successful approaches
agencies are using to measure and report on their outcomes, including
the use of program evaluation tools. Additionally, we recommend that
the Director of OMB work with executive branch agencies to identify the
barriers to obtaining timely data to show progress against performance
goals and the best ways to report information where there are
unavoidable lags in data availability. Interagency councils, such as
the President's Management Council and the Chief Financial Officers'
Council, may be effective vehicles for working on these issues.
To facilitate the transformation of agencies' management cultures to be
more results-oriented, we recommend that the Director of OMB work with
agencies to ensure they are making adequate investments in training on
performance planning and measurement, with a particular emphasis on how
to use performance information to improve program performance.
Matters for Congressional Consideration:
To ensure that agency strategic plans more closely align with changes
in the federal government leadership, Congress should consider amending
GPRA to require that updates to agency strategic plans be submitted at
least once every 4 years, 12-18 months after a new administration
begins its term. Additionally, consultations with congressional
stakeholders should be held at least once every new Congress and
interim updates made to strategic and performance plans as warranted.
Congress should consider using these consultations along with its
traditional oversight role and legislation as opportunities to clarify
its performance expectations for agencies. This process may provide an
opportunity for Congress to develop a more structured oversight agenda.
To provide a framework to identify long-term goals and strategies to
address issues that cut across federal agencies, Congress also should
consider amending GPRA to require the President to develop a
governmentwide strategic plan.
Agency Comments:
We provided a copy of the draft report to OMB for comment. OMB's
written comments are reprinted in appendix VIII. In general, OMB agreed
with our findings and conclusions. OMB agreed to implement most of our
recommendations, noting that these recommendations will enhance its
efforts to make the government more results oriented. OMB agreed to (1)
work with agencies to ensure they are provided adequate training in
performance management, (2) revise its guidance to clarify the
integrated and complementary relationship between GPRA and PART, and
(3) continue to use PART to improve agency performance measurement
practices and share those practices across government.
In response to our recommendation that OMB fully implement GPRA's
requirement to develop a governmentwide performance plan, OMB stated
that the President's Budget represents the executive branch's
governmentwide performance plan. However, the agency-by-agency focus of
the budget over the past few years does not provide an integrated
perspective of government performance, and thus does not meet GPRA's
requirement to provide a "single cohesive picture of the annual
performance goals for the fiscal year." In response to our matter for
congressional consideration that Congress should consider amending GPRA
to require the President to develop a governmentwide strategic plan,
OMB noted that the budget serves as the governmentwide strategic plan.
However, in our opinion, the President's Budget focuses on establishing
agency budgets for the upcoming fiscal year. Unlike a strategic plan,
it provides neither a long-term nor an integrated perspective on the
federal government's activities. A governmentwide strategic plan should
provide a cohesive perspective on the long-term goals of the federal
government and provide a basis for fully integrating, rather than
primarily coordinating, a wide array of federal activities.
We provided relevant sections of the draft report to Education, DOE,
HUD, SBA, SSA, and DOT. Education and SBA did not provide any comments,
while DOT provided minor technical comments. Written comments from DOE,
HUD, and SSA are reprinted in appendixes IX, X, and XI, respectively,
along with our responses.
DOE disagreed with portions of our analyses of its 2004 Annual
Performance Plan and its 2002 Performance and Accountability Report.
Our analysis of DOE's documents was based on specific criteria (see
appendixes IV and V for details) and was characterized in relation to
our reviews of the other five agencies' documents. We modified or
clarified certain characterizations in response to DOE comments, but
for the most part found that our characterizations were appropriate.
SSA generally agreed with our observations and agreed to incorporate
them in its future planning efforts. SSA made several points of
clarification and disagreed with our observation that its performance
and accountability report does not clearly state how program
evaluations were used to answer questions about program performance and
results and how they can be improved. SSA noted that its evaluations
rely on surveys, and these surveys form the basis for its efforts to
deliver high-quality service. SSA also noted that it lists other
evaluations that are of great importance to its ongoing operations. We
do not discount the usefulness of SSA's surveys in assessing its day-
to-day management of programs. Rather, we believe that it would be
helpful for SSA to clearly identify the range of evaluations conducted
and how each of them contributed to improved program performance.
HUD noted that all of the areas we suggested for further improvement
are under consideration for improvement. However, they disagreed with
us on two observations related to the strategic plan: (1) that the link
between its long-term and intermediate goals is difficult to discern
and (2) that it did not explain how it used the results of program
evaluations to update the current plan and did not include a schedule
for future evaluations. On the basis of OMB guidance for preparing
strategic plans and the criteria we used to evaluate all six agencies'
strategic plans (see app. III for more detail), we maintain that these
two observations are valid and require further attention. HUD also
disagreed with how we presented the performance information in its
summary report cards (see fig. 22). HUD noted that many of the results
were explained in the individual indicator write-ups that followed the
summary information. Our analysis of HUD's information included
qualitative aspects of how the information was presented, such as its
usefulness to inform the average reader with little or no exposure to
the subject matter, and the extent to which HUD presented a complete
summary of performance information in a user-friendly format.
Technical comments from DOE, HUD, and SSA were incorporated, as
appropriate.
[End of section]
Appendixes:
Appendix I: Objectives, Scope, and Methodology:
As agreed with your offices, our objectives for this report were to
determine (1) the effect of the Government Performance and Results Act
(GPRA) over the last 10 years in creating a governmentwide focus on
results and the government's ability to deliver results to the American
public, including an assessment of changes in the overall quality of
agencies' strategic plans, annual performance plans, and annual
performance reports; (2) the challenges that agencies face in measuring
performance and using performance information in management decisions;
and (3) how the federal government can continue to shift toward a more
results-oriented focus. To meet our objectives, we collected
governmentwide data to assess the government's overall focus on
results. We conducted a governmentwide survey of federal managers,
focus groups with federal managers and GPRA experts, and interviews
with top appointed officials. We identified and reviewed previously
published reports on GPRA. Finally, we selected a sample of agencies to
review for changes in the quality of their strategic plans, performance
plans, and performance reports since their initial efforts.
We conducted our work from January through November 2003 in Washington,
D.C. in accordance with generally accepted government auditing
standards. We provided drafts of the relevant sections of this report
to officials from each of the agencies whose GPRA reports we reviewed.
We also provided a draft of this report to OMB.
Methodology for Governmentwide Survey:
A Web-based questionnaire on performance and management issues was
administered to a stratified random probability sample of 800 persons
from a population of approximately 98,000 mid-level and upper-level
civilian managers and supervisors working in the 24 executive branch
agencies covered by the Chief Financial Officers Act of 1990 (CFO). The
sample was drawn from the Office of Personnel Management's (OPM)
Civilian Personnel Data File as of December 31, 2002, using file
designators indicating performance of managerial and supervisory
functions.
The questionnaire was designed to obtain the observations and
perceptions of respondents on various aspects of GPRA as well as such
results-oriented management topics as the presence, use, and usefulness
of performance measures, hindrances to measuring and using performance
information, and agency climate. Most of the items on the questionnaire
were closed-ended, meaning that depending on the particular item,
respondents could choose one or more response categories or rate the
strength of their perception on a 5-point extent scale. Almost all the
items on this questionnaire were asked in two earlier mail-out surveys.
One survey was conducted between November 1996 and January 1997 as part
of the work we did in response to a GPRA requirement that we report on
implementation of the act. The other survey was conducted between
January and August 2000.[Footnote 63]
This survey covered the same CFO Act agencies and was designed to
update the results from the two earlier surveys. Similar to the two
earlier surveys, the sample was stratified by whether the manager or
supervisor was Senior Executive Service (SES) or non-SES. The
management levels covered general schedule (GS), general management
(GM), or equivalent schedules at levels comparable to GS/GM-13 through
career SES or equivalent levels of executive service. The sample also
included the same or equivalent special pay plans that were covered in
our 2000 survey, e.g., Senior Foreign Service executives.
We sent an e-mail to members of the sample that notified them of the
survey's availability on the GAO Web site and included instructions on
how to access and complete the survey. Members of the sample who did
not respond to the initial notice were sent up to two subsequent
reminders asking them to participate in the survey. The survey was
administered from June through August 2003.
During the course of the survey, we deleted 26 persons from our sample
who had either retired, separated, died, or otherwise left the agency
or had some other reason that excluded them from the population of
interest. We received useable questionnaires from 503 sample
respondents, about 65 percent of the eligible sample. The eligible
sample includes 39 persons that we were unable to locate and therefore
unable to request that they participate in the survey.
To assess whether the views of those individuals who chose not to
participate in our survey might be different than those who did, we
made an effort to administer a brief survey over the telephone to those
individuals who still had not responded about a month or more after the
survey had been available to them despite being contacted twice after
the initial notification e-mail had been sent out. This telephone
survey consisted of four items from the full survey. There were 58
persons who agreed to answer these questions over the telephone. This
was 41 percent of those individuals who had not responded at the time
we attempted to contact them for the purpose of asking these four
questions.
We analyzed the responses of this group on the four selected items
compared to the responses received from all other respondents. Our
analyses of the items showed very few differences between nonresponders
and responders. There was no sufficient or consistent pattern of
responding that would warrant a conclusion that the views of
nonresponders were notably different than responders. The responses of
each eligible sample member who provided a useable questionnaire were
subsequently weighted in the analysis to account statistically for all
the members of the population.
The overall survey results are generalizable to the population of
managers as described above at the CFO Act agencies. All results are
subject to some uncertainty or sampling error as well as nonsampling
error. As part of our effort to reduce nonsampling sources of error in
survey results, we checked and edited (1) the survey data for responses
that failed to follow instructions and (2) the programs used in our
analyses. In general, percentage estimates in this report for the
entire sample have confidence intervals ranging from about ± 4 to ±11
percentage points at the 95 percent confidence interval. In other
words, if all CFO Act agency managers and supervisors in our population
had been surveyed, the chances are 95 out of 100 that the result
obtained would not differ from our sample estimate in the more extreme
cases by more than ±11 percent. Appendix VI shows the questions asked
with the weighted percentage of managers responding to each item.
Because a complex sample design was used in the current survey as well
as the two previous surveys and different types of statistical analyses
are being done, the magnitude of sampling error will vary across the
particular surveys, groups, or items being compared due to differences
in the underlying sample sizes and associated variances. The number of
participants in the current survey is only about one fifth of the
number in the 2000 survey (2,510) and slightly more than half of those
in the first survey (905). The 2000 survey was designed with a larger
sample than the other two surveys in order to provide estimates for
each individual agency as well as all the CFO Act agencies
collectively. Consequently, in some instances, a difference of a
certain magnitude may be statistically significant. In other instances,
depending on the nature of the comparison being made, a difference of
equal or even greater magnitude may not achieve statistical
significance. For example, when comparing a result from the current
survey to the larger 2000 survey with its relatively smaller confidence
interval, a difference of a certain magnitude may be significant.
However, when comparing the current survey with the first survey, that
difference may not be significant given the greater imprecision in the
estimates due to both surveys' smaller sample sizes. We note throughout
the report when differences are significant at the .05 probability
level.
Methodology for Focus Groups:
We held a series of focus groups as one of our methods for obtaining
information about the accomplishments and challenges agencies face in
implementing and overseeing GPRA-related activities. Focus groups are a
form of qualitative research in which a specially trained leader, a
moderator, meets with a small group of people (usually 8 to 10) who are
knowledgeable about the topics to be discussed.
In all, we conducted eight focus groups--one with experts on GPRA and
performance management and seven with federal managers. For our focus
group with experts, we invited individuals from the private sector,
academia, the National Academy of Public Administration, and OMB. These
individuals were involved either in drafting GPRA, overseeing its
implementation, or studying and critiquing implementation, from outside
government. Out of the 14 experts we invited, a total of 11 attended
the focus group.
For our focus groups with agency managers, we obtained a list of
potential participants for our focus groups by contacting all 24 CFO
Act agencies and requesting that they submit a list of candidates and
their profiles based on the following criteria: federal managers (1) in
the GS-13 pay grade and above, including members of the SES; (2) having
at least 3 years of managerial experience; (3) currently located in the
Washington, D.C., metropolitan area; (4) having hands-on experience
with GPRA or performance management;[Footnote 64] and (5) representing
both departments and their component bureaus. We received profiles of
candidates from all agencies; however, no managers from the OPM chose
to participate in the focus groups.
To select focus group participants, we reviewed the profiles submitted
by agencies and selected candidates with diverse experience who held a
variety of different positions within the agency in order to capture a
broad range of perspectives. For example, we invited a comptroller; a
deputy director of management, administration, and planning; budget
directors; budget officers; management analysts; and program managers;
among others. We contacted the candidates and provided them with the
list of questions to be discussed at the focus group in advance so they
would be aware of our interests and be better able to provide us, where
possible, with examples to illustrate their responses to our questions.
Out of 104 agency officials we invited, 70 participated in the focus
groups.[Footnote 65]
During each session, the moderator explained the scope of our work and
elaborated on how the focus groups were one of several methods we were
using to collect information relevant to our objectives. As part of the
focus group process, the moderator asked participants at each session
to identify the main accomplishments and challenges that, in their
view, could be attributed to GPRA and to mention possible solutions to
these challenges. During the sessions, we created lists of the
accomplishments, challenges, and solutions identified by group
participants and posted these lists around the meeting room.
Participants were then asked to review the lists and vote on the three
most important accomplishments and the top three challenges.[Footnote
66]
To organize the information collected during the focus groups, we
reviewed the statements made by participants in response to our
questions. We identified related sets of statements and summarized them
as a general theme. We noted how often a theme was expressed both
within and across each focus group. We also examined the number of
votes each posted statement obtained. Our analysis focused on those
themes that were supported by statements that obtained a high number of
votes and were mentioned frequently within and across the majority of
focus groups.
The focus group results discussed in this report are summary
descriptions reflecting the range of views and perceptions held by
employees, supervisors, or project managers who participated in the
focus groups. Although we cannot assume all federal managers share
these views, the extent to which certain opinions or perceptions were
repeatedly expressed or endorsed by many participants from multiple
focus groups provides a rough gauge of the significance of these views.
Methodology for Interviews with Political Appointees:
To obtain an additional perspective from top political managers of
federal agencies on GPRA, we held telephone or in-person interviews
with 10 high-level officials serving under political appointments with
CFO Act agencies. Five former officials from the Clinton administration
and five serving under the current Bush administration were
interviewed. For example, we interviewed deputy secretaries, chief
financial officers, and deputy assistant secretaries for management. We
asked them to provide their perspective on the main accomplishments or
other effects of GPRA, the key challenges to implementation, and
possible improvements to GPRA. We summarized the interviewees' answers
and identified recurring themes or observations for our analysis.
Methodology for Selecting Agencies to Review for Changes in the Quality
of Their Strategic Plans, Annual Performance Plans, and Annual
Performance Reports:
To address how the quality of agency strategic plans, performance
plans, and performance reports have changed since their initial
efforts, we reviewed a sample of six agencies' current strategic plans,
annual performance plans, and annual performance reports and compared
the results to our findings from prior reviews of the agencies' initial
efforts in producing these documents. We did not independently verify
or assess the information we obtained from agency plans and reports. If
an agency chose not to discuss its efforts concerning elements in the
plans and reports, it does not necessarily mean that the agency is not
implementing those elements.
We selected the departments and agencies to review based on the extent
to which they collectively represented the full range of
characteristics in the following four areas: (1) agency size (small,
medium, large); (2) primary program types (direct service, research,
regulatory, transfer payments, and contracts or grants); (3) quality of
fiscal year 2000 performance plan based on our previous review (low,
medium, high);[Footnote 67] and (4) type of agency (cabinet department
and independent agency).
Based on these characteristics, we selected the following departments
and agencies:
* Department of Education (Education),
* Department of Energy (DOE),
* Department of Housing and Urban Development (HUD),
* Small Business Administration (SBA),
* Social Security Administration (SSA), and:
* Department of Transportation (DOT).
Table 4 shows the characteristics represented by each of these
agencies.
Table 4: Summary of Characteristics of Agencies Selected for Review of
Strategic Plans, Annual Performance Plans, and Annual Performance
Reports:
Departments: Education;
Size[A]: Full time equivalent positions: 4,756;
Size[A]: Small;
Functions: Research: Yes;
Functions: Direct service: Yes;
Functions: Regulatory: Yes;
Functions: Transfer payments: Yes;
Functions: Grants/contracts: Yes;
Quality of fiscal year 2000 performance plans: Medium.
Departments: DOE;
Size[A]: Full time equivalent positions: 16,067;
Size[A]: Medium;
Functions: Research: Yes;
Functions: Direct service: No;
Functions: Regulatory: Yes;
Functions: Transfer payments: No;
Functions: Grants/contracts: Yes;
Quality of fiscal year 2000 performance plans: Low.
Departments: HUD;
Size[A]: Full time equivalent positions: 10,752;
Size[A]: Medium;
Functions: Research: No;
Functions: Direct service: Yes;
Functions: Regulatory: Yes;
Functions: Transfer payments: Yes;
Functions: Grants/contracts: Yes;
Quality of fiscal year 2000 performance plans: Medium.
Departments: DOT;
Size[A]: Full time equivalent positions: 135,022[B];
Size[A]: Large;
Functions: Research: Yes;
Functions: Direct service: Yes;
Functions: Regulatory: Yes;
Functions: Transfer payments: No;
Functions: Grants/contracts: Yes;
Quality of fiscal year 2000 performance plans: H.
Agencies: SBA;
Size[A]: Full time equivalent positions: 4,005;
Size[A]: Small;
Functions: Research: No;
Functions: Direct service: Yes;
Functions: Regulatory: Yes;
Functions: Transfer payments: No;
Functions: Grants/contracts: Yes;
Quality of fiscal year 2000 performance plans: Low.
Agencies: SSA;
Size[A]: Full time equivalent positions: 64,418;
Size[A]: Large;
Functions: Research: No;
Functions: Direct service: Yes;
Functions: Regulatory: No;
Functions: Transfer payments: Yes;
Functions: Grants/contracts: No;
Quality of fiscal year 2000 performance plans: High.
Source: GAO.
[A] The size of the agencies is based on data from December 2002.
[B] In March 2003, the U.S. Coast Guard and the Transportation Security
Administration (TSA) were transferred from DOT to the Department of
Homeland Security. According to the fiscal year 2005 President's
Budget, in fiscal year 2003, the Coast Guard and TSA had 43,702 and
57,324 full-time equivalent positions, respectively.
[End of table]
A more detailed discussion of the criteria we used to assess the
quality of the agencies' planning and reporting documents and the
results of our review is contained in appendixes III, IV, and V.
[End of section]
Appendix II: Focus Group Participants Agreed GPRA Provides a Framework
for Federal Agencies to Become More Results Oriented:
While GPRA's goal is to make the federal government more results
oriented, work carried out in support of this effort, such as planning
activities, implementing programs, reporting on outcomes, and
evaluating performance, generally lies in the hands of federal
managers. To get a better appreciation for the main accomplishments and
challenges agencies face in implementing and overseeing GPRA-related
activities, we conducted a total of seven focus groups comprised of
federal managers from 23 CFO Act agencies, and an eighth focus group
comprised of 11 experts on GPRA and performance management and
budgeting.
For our focus groups, we asked participants to discuss (1) the key
accomplishments of GPRA to date, (2) whether GPRA has created a focus
on achieving results across the federal government, (3) the effect of
GPRA on the government's ability to deliver results to the American
public, (4) the persistent and prevalent challenges agencies face in
implementing and overseeing GPRA-related activities, and (5)
suggestions to address these challenges.
We recorded the views expressed by participants and categorized them
into themes that were most commonly expressed or endorsed both within
and across the groups. The focus group results discussed in this report
are organized according to the themes we identified and are summary
descriptions reflecting the range of views and perceptions expressed by
the experts, supervisors, and project managers. A more complete
discussion of our scope and methodology can be found in appendix I.
Focus group participants indicated that GPRA has helped to make the
federal government more results oriented. However, they noted that a
number of obstacles have made GPRA implementation challenging, such as
difficulty in establishing results-oriented goals and measuring
performance, addressing frequently changing priorities resulting from
changes in administration, and lack of top leadership support for GPRA.
In all, participants generally perceive the information contained in
GPRA reports to be important and useful; however, they do not believe
that lawmakers use this information to make resource decisions or
conduct oversight. To address these problems and concerns, focus group
participants stated that, among other things, Congress should provide
guidance to agencies on how to make GPRA reports more useful, OMB
should reinforce its value as a management tool, and agencies need to
commit the resources necessary to carry out GPRA-related activities.
GPRA Accomplishments:
Overall, focus group participants stated that GPRA has had a positive
effect on federal agencies' efforts to become more results oriented.
Based in statute, GPRA has created a framework for agencies to focus on
achieving results by requiring them to establish program goals and
objectives, develop performance indicators, and measure the extent to
which they have made progress towards achieving program goals. As a
result, federal managers noted that they have been increasingly able to
view their programs in terms of outcomes, not outputs, and have been
generally learning how to use this framework as a management tool.
Participants also attributed a series of cultural changes within
federal agencies to GPRA, where problem solving, creative thinking, and
agencywide discussions on budget and performance have become more
common. The strategic and annual performance plan and performance
reports that federal agencies are required to submit to OMB and
Congress under GPRA have increased the transparency of government
activities. These documents have also helped agencies justify their
budget requests based on their performance.
Creating a Framework in Statute and a Management Tool for Agency
Leadership:
Participants agreed that GPRA created a framework in statute for
federal agencies to plan their activities in order to become more
results oriented and provided a managerial tool for program
accountability. Using this framework, agencies can develop and focus on
strategies to carry out the programs they administer; set goals and
identify performance indicators that will inform them whether or not
they achieved the performance they expected; and determine what impact,
if any, their programs have had on the American public. According to
the experts in one of our focus groups, comparing federal agencies'
current mission statements contained in their strategic plans to what
they were in the past demonstrates that agencies have done some "soul
searching" to get a better sense of what their role is (or should be)
and how they can achieve it. Given that GPRA is in statute, the use of
this planning framework is likely to be sustained within agencies.
Participants also mentioned that GPRA has encouraged federal managers
to view their programs in terms of results, not just inputs and
outputs. Such a change is important, as it has encouraged federal
managers to reflect on the statutory intent of their programs and use
this framework as a management tool for establishing accountability
within their programs.
Cultural Change within Federal Agencies:
Participants in the majority of focus groups agreed that GPRA has been
a driving force behind many cultural changes that have occurred within
federal agencies. Highlighting the focus on results, participants
stated that GPRA has stimulated a problem-solving approach within
federal agencies and encouraged them to think creatively when
developing performance indicators for their programs. GPRA has also
changed the dialogue within federal agencies; front-line managers and
staff at lower levels of the organization now discuss budget issues in
connection with performance. Similarly, experts noted that performance
management and resource investments are more frequently communicated
between agency officials and Congress than in the past. Within
agencies, GPRA documents can provide a context of missions, goals, and
strategies that political appointees can use to articulate agencies'
priorities.
Increased Transparency of Government Results:
Some participants agreed that GPRA has increased federal agencies'
ability to present their results to the American public, benefiting
both stakeholders and agency staff. On the one hand, GPRA reports
enable federal agencies to communicate the results of government
programs and activities to a broad public. For example, GPRA reports
are available on agencies' Web sites and provide information to OMB,
Congress, and the American public on what agencies plan to do, how they
plan to do it, and, as summarized in the performance and accountability
reports, what agencies have accomplished and how much money it cost
them to do it.
Similarly, some participants agreed that GPRA allows federal employees
to see exactly how their work can produce a positive outcome,
increasing employee morale. Using information contained in GPRA
reports, agency employees can compare department goals to the results
of their activities, and see how their work contributes to these goals.
For example, a focus group participant from the Indian Health Service
in California stated that he was pleased to learn that health care
indicators of some Native American tribes had already exceeded levels
originally projected by his agency to be reached by the year 2010.
Link between Budget and Performance:
Participants also agreed that the GPRA framework has had a positive
effect on agencies' ability to link their performance to their budget.
By focusing on their mission and outcomes, agencies are learning to
prioritize activities and align their resources to ensure that they
will be able to achieve results. For example, a participant stated that
the National Wild Horse and Burro Program, managed by the Bureau of
Land Management in the U.S. Department of the Interior, recently
developed a model for its GPRA-related work which divided their program
into discrete components, identifying the results it could accomplish
in the short-and long-term and specifying what additional resources
were needed. According to the participant, the program received
additional funding based on this needs assessment.
In addition, a few managers noted that the link between budget and
performance has given agencies an incentive to commit the resources
necessary to modernize information systems. Having the right
information on time enables agencies to integrate budget requests and
performance information in ways that are more meaningful. This
information also increases Congress's ability to make informed budget
decisions based on agency performance.
Views on Delivering Results to the American Public Were Mixed:
Participants' views on whether GPRA has helped agencies deliver results
to the American public were generally mixed. Some federal managers in
our focus groups agreed that GPRA has had a positive effect on raising
awareness on many issues, and that in and of itself is a way of
delivering results. The information gathered and reported for GPRA
allows agencies to make better-informed decisions, which improves their
ability to achieve results. Of note, GPRA has allowed agencies to move
towards performance-based budgeting, which helps agencies identify
resources available to use in programs where outcomes can be achieved.
For example, programs and expenses that do not add value to the
agency's mission could be eliminated. Having performance data readily
available is another key area where GPRA has enabled agencies to
deliver results to the American public.
Other participants stated that while certain aspects of GPRA-related
work have been positive, agencies' ability to deliver results and
public awareness of their activities cannot always be exclusively
attributed to GPRA. For example, while measuring performance is a move
in the right direction, GPRA reports provide too many indicators and it
is unclear how this has led to better performance among federal
agencies. A few participants also stated that agencies deliver results
in ways that the American public does not fully recognize. For example,
a participant stated that agencies working in the area of international
relations generally lack recognition for the results of their
activities, as the American public is generally unfamiliar with the
nuances of foreign policy and all the programs the U.S. government
finances overseas. And while GPRA has helped these agencies prioritize
their work to achieve results, it is unclear that GPRA has improved the
visibility and understanding in the public eye of what these agencies
do because the American public does not see the results of their work.
A participant stated that research-based agencies have also used GPRA
to plan activities that benefit the American public in ways they are
not fully aware of. For example, National Aeronautics and Space
Administration's (NASA) space program includes, among other things,
predicting whether or not an asteroid will strike the earth, although
on a daily basis the American public is probably not worried about such
a rarity.
For other participants, the link between GPRA and agencies' service
delivery was not clear. Participants characterized GPRA-related
activities as time consuming, and there is no real evidence that this
work has improved their ability to deliver results to the American
public. Other participants stated that many agencies rely on grant
recipients to carry out their work, and delivering results to the
American public depends, to a large extent, on the diligence of these
organizations to implement their programs. Overall, they held the view
that performance would not change dramatically if GPRA were no longer a
requirement for federal agencies.
Alternate Views on GPRA's Effect:
Participants in one of our focus groups stated that GPRA, per se, had
not led federal agencies to achieve specific accomplishments. These
participants believed that managers' initiative, not the framework
established by GPRA, has been key to improving agencies' planning
efforts and focus on results. A few participants also mentioned that
the results framework established by GPRA is somewhat intangible unless
managers can use it effectively; without the adequate infrastructure to
implement GPRA, an agency's compliance with the law is simply
paperwork, as GPRA does not allow for a systematic and thorough
approach to performance management. For example, while agencies can
develop performance indicators to gauge progress towards a program
goal, they often encounter problems in collecting relevant data and
measuring the agencies' contribution to a specific outcome. In
addition, agencies are not able to make changes to the programs they
manage, limiting their ability to deliver results.
Challenges in Implementing and Overseeing GPRA Activities:
Participants stated that agencies face significant challenges in
complying with GPRA-related activities. Focus group participants also
agreed Congress does not appear to use agencies' performance
information when making budget decisions. In carrying out GPRA-related
activities, managers find it difficult to identify performance
indicators and establish program goals that are results oriented, as
required by GPRA. In addition, a few participants stated that they lack
the support from senior management to work on GPRA activities. Changes
in administration also tend to disrupt progress made by agencies in
support of GPRA.
Performance Information Not Used:
Participants strongly believed that Congress does not take into account
agencies' performance information when overseeing agency activities and
making budget decisions. As a result, there is a negative effect on how
agencies view their efforts in producing GPRA plans and reports--many
believe that they are less worthwhile than the effort and resources
invested. On the other hand, participants perceive the budget process
strictly as a political exercise, and it is unclear how useful
performance information can be in this context.
Complexity of Establishing Results-Oriented Goals and Measuring
Performance:
Participants stated that establishing results-oriented goals and
identifying performance indicators are generally complex undertakings.
Participants agreed that they often feel as if they were trying to
measure the immeasurable, not having a clear understanding of which
performance indicators could accurately inform the agency how it is
carrying out a specific activity. And while agencies generally try to
improve the indicators from one year to another, in doing so, they
generally lose their ability to develop trend information to track
progress made over time.
Participants also mentioned that there appears to be a disconnect
between some federal programs that generally produce results over a
longer time period and GPRA's requirement that agencies report annually
on their progress towards achieving their goals. Participants stated
that federal programs, especially those that are research-based, often
take years to achieve the full scope of their work. Consequently, and
for reasons that range from lack of performance data to an absence of
short-term outcomes, it could appear as though resources invested in
carrying out their activities led to no results in the short run.
Focus group participants generally agreed that in cases where third
parties, such as states or localities, implement federal programs, some
agencies face challenges in obtaining timely performance data from
relevant partner organizations. In some instances, federal managers
have trouble identifying what the government's contribution has been to
a specific outcome. While the experts generally attributed this to
agencies not doing a good job at setting appropriate goals and the
corresponding measures and objectives not being clear enough, managers
stated that their lack of control over grantees and the strict
reporting deadlines imposed by GPRA were factors that worked against
their efforts to deliver results.
Managing Strategically with Frequently Changing Priorities:
Some participants stated that it is difficult for them to manage
strategically, given the frequently changing priorities that come with
changes in administrations. While GPRA requires an agency to develop a
strategic plan at least every 3 years to cover the following 5-year
period, participants agreed that it makes little sense to update it
shortly before a new administration is scheduled to take office. In
addition, changes in political leadership generally result in a new
agenda with new objectives. These changes force agencies to revise
their plans, management initiatives, and strategies, which translate
into additional GPRA-related work, generally characterized by focus
group participants as a burden agency staff must add to their normal
work load.
Lack of Top Leadership Support:
Some participants stated that they often encounter resistance from
agency leadership to endorse GPRA-related activities. In some
instances, senior-level support for GPRA falters or is nonexistent.
Participants attributed this to the reluctance within some federal
agencies to think about outcomes and performance. Some focus group
participants stated that in some instances high-level managers are
somewhat averse to being held accountable for the results of programs
they run.
Suggestions to Address GPRA Challenges:
To address these challenges, focus group participants made the
following suggestions.
Congress Should Clarify How Performance Information Could Be More
Useful:
Congressional staff should provide guidance on agencies' GPRA
submissions--specifically, how information could be presented in the
reports to make it more useful in the decision-making process. They
could also communicate to agencies how the performance information is
being used, so that agencies do not perceive their data gathering
efforts as inconsequential. Additionally, in using performance
information to make budget decisions, Congress should consider the
results achieved by agencies in addition to results not achieved.
The Administration and OMB Need to Reinforce the Value of GPRA as a
Management Tool:
Agencies should embrace GPRA as a management tool, not just an external
requirement that is separate from their day-to-day activities. To this
end, the administration and OMB need make sure agency officials
understand how GPRA can be further integrated with other management
initiatives.
Agency Guidance on GPRA Should Recognize Diversity of Federal Agencies:
OMB's guidance to agencies on how to implement GPRA should recognize
that one-size-fits-all approaches are unlikely to be useful. OMB should
also afford agencies some flexibility by simplifying the reporting
process. For example, some participants believed that not everything
needed to be measured. OMB should make exceptions for unique situations
and programs, e.g., science programs, and it could consider providing
multiyear funding for them.
OMB Should Publish a Governmentwide Performance Report:
OMB should commit to regularly publishing a governmentwide performance
report that would articulate the government's accomplishments to the
public. It would also be useful if it singled out higher-performing
programs so agencies could use them as models to guide their planning
efforts.
Agencies Need to Obtain and Commit Resources to Carry Out GPRA-Related
Activities:
Agency leadership needs to ensure that staff members have the necessary
resources to work on GPRA-related activities. In addition, they need to
invest resources to train staff, including political appointees, on
GPRA and the benefits of linking budget to performance.
Timing of GPRA Reports Should Take into Account Changes of
Administration:
Under GPRA, agencies are to update their strategic plans every 3 years.
However, this effort can be wasted. Given that federal administrations
generally span at least 4 years, participants suggested that the
required update be changed to every 4 years to correspond with new
presidential terms.
Agencies Should Share Experiences on How to Address Common Problems:
Agencies should collaborate more to develop strategies to address
difficult issues, such as how to identify performance indicators and
measure agency contributions to specific outcomes. It would be useful
if more agencies created structured forums for managers to share
experiences, talk about effective practices, and share solutions.
[End of section]
Appendix III: Observations on Agencies' Strategic Plans:
Under GPRA, strategic plans are the starting point and basic
underpinning for results-oriented management. One of our objectives was
to assess the changes in the overall quality of agencies' goals,
strategies, and data articulated in their strategic plans. To meet this
objective, we judgmentally selected six agencies--Education, DOE, HUD,
SBA, SSA, and DOT--using criteria such as agency size, primary program
types, and previous GAO reviews. To assess the overall quality and
improvements made to the agencies' strategic plans, we relied on
requirements contained in GPRA and accompanying committee report
language,[Footnote 68] guidance to agencies from OMB for developing
strategic plans,[Footnote 69] previous GAO reports and
evaluations,[Footnote 70] and our knowledge of agencies' operations and
programs. In conducting our reviews, we compared our assessments of
agencies' current strategic plans to our assessments of draft plans
from fiscal year 1997.[Footnote 71] A more detailed discussion of our
scope and methodology can be found in appendix I.
Required Elements of Agency Strategic Plans:
GPRA requires an agency's strategic plan to contain six key elements:
1. A comprehensive agency mission statement. The agency mission
statement should concisely summarize what the agency does, as required
by law, presenting the main purposes for all its major functions and
operations. According to OMB guidance issued in 2002, a mission
statement is brief, defining the basic purpose of the agency, and
corresponds directly with an agency's core programs and activities. The
program goals should flow from the mission statement as well. The
federal government's adaptive responses over time to new needs and
problems have contributed to fragmentation and overlap in a host of
program areas, such as food safety, employment training, early
childhood development, and rural development, which could limit the
overall effectiveness of the federal effort. The mission statement
helps to distinguish agencies' roles from one another and reduce the
overlap and identify areas needing coordination and collaboration.
2. Agencywide long-term goals and objectives. General goals and
objectives--or strategic goals--explain what results are expected from
the agency's major functions and when to expect those results. Thus,
such goals are an outgrowth of the mission and are very often results
oriented. OMB guidance states that the goals should be defined in a
manner that allows a future assessment to be made on whether the goal
was or is being achieved. General goals should predominately be
outcomes, and are long-term in nature.
3. Approaches or strategies to achieve goals and objectives. Strategies
help in aligning an agency's activities, core processes, and resources
to support achievement of the agency's strategic goals and mission.
Under GPRA, strategies are to briefly describe the operational
processes, staff skills, and technologies, as well as the human,
capital, information, and other resources needed. According to OMB
guidance, descriptions should be brief, but more detailed data should
be provided if a significant change in a particular means or strategy
would be essential for goal achievement. In addition, the plan should
summarize agencies' efforts to provide high-quality and efficient
training and skill improvement opportunities for employees. As we have
reported previously, agencies' planning processes should support making
intelligent resource allocation decisions that minimize, to the extent
possible, the effect of funding reductions on mission accomplishment.
4. A description of the relationship between long-term and annual
goals. Under GPRA, agencies' long-term strategic goals and objectives
are to be linked to their annual performance plans and the day-to-day
activities of their managers and staff. Without this linkage, Congress
may not be able to judge whether an agency is making progress toward
achieving its long-term goals. OMB guidance states that an updated and
revised strategic plan should briefly outline (1) the type, nature, and
scope of the performance goals being included in annual performance
plans and (2) how these annual performance goals relate to the long-
term, general goals and their use in helping determine the achievement
of the general goals.
5. An identification of key external factors. Identification of key
factors, external to the agency and beyond its control that could
significantly affect the achievement of the strategic goals, are
important for Congress or the agencies to judge the likelihood of
achieving the strategic goals and actions needed to better meet those
goals. Such external factors could include economic, demographic,
social, technological, or environmental factors. Information on these
factors can be useful for goal setting and also for explaining results
in the agency's annual performance reports, including, when applicable,
the reasons annual performance goals were not met. According to OMB
guidance, if key factors cannot be identified, a statement of
explanation should be included in the plan.
6. A description of program evaluations. Finally, strategic plans
should include a description of completed program evaluations that were
used in developing the strategic plan, and a schedule for future
program evaluations. Program evaluations can be a potentially critical
source of information for Congress and others in ensuring the validity
and reasonableness of goals and strategies, as well as for identifying
factors likely to affect performance. Such information can also be
useful in explaining results in an agency's annual performance report,
including, when applicable, the reasons annual performance goals were
not met, and identifying appropriate strategies to meet unmet goals.
Program evaluations are defined in the act as objective and formal
assessments of the results, impact, or effects of a program or policy.
The evaluations include assessments of the implementation and results
of programs, operating policies, and practices.
In addition to the six key elements, OMB guidance also states that
agencies participating in crosscutting programs should describe in
their strategic plans how the programs are related and how coordination
will occur to support common efforts. Uncoordinated program efforts can
waste scarce funds, confuse and frustrate program customers, and limit
the overall effectiveness of the federal effort. OMB guidance also
states that the strategic plan should include a brief description of
any steps being taken to resolve mission-critical management problems.
One purpose of GPRA is to improve the management of federal agencies.
Therefore, it is particularly important that agencies develop
strategies to address management challenges that threaten their ability
to meet long-term strategic goals as well as this purpose of GPRA.
As shown in table 5, the majority of agencies have made progress in
addressing the required elements of strategic planning under GPRA.
Table 5: Agencies' Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Agency strategic plans: Department of Education;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Education;
Plan year: 2002;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Energy;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Energy;
Plan year: 2003[A];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Housing and Urban Development;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: No;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: No;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Department of Housing and Urban Development;
Plan year: 2003;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Small Business Administration;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Small Business Administration;
Plan year: 2001[B];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Agency strategic plans: Social Security Administration;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Social Security Administration;
Plan year: 2003;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long- term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Transportation;
Plan year: 1997;
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: No;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: No;
Element included in agency strategic plan? External factors: No;
Element included in agency strategic plan? Evaluations: Yes.
Agency strategic plans: Department of Transportation;
Plan year: 2003[A];
Element included in agency strategic plan? Mission statement: Yes;
Element included in agency strategic plan? Long-term goals: Yes;
Element included in agency strategic plan? Strategies: Yes;
Element included in agency strategic plan? Relationship
between long-term goals and annual goals: Yes;
Element included in agency strategic plan? External factors: Yes;
Element included in agency strategic plan? Evaluations: No.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD- 10.1.16];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R];
and GAO analysis of U.S. Department of Education, Office of the Deputy
Secretary, Planning and Performance Management Service, U.S. Department
of Education Strategic Plan 2002-2007; (Washington, D.C. 2002);
Department of Energy, 2003 Strategic Plan (Draft); (Washington, D.C.
2003);
Department of Transportation, U.S. Department of Transportation Draft
Strategic Plan for Fiscal Years 2003-2008; (Washington, D.C. 2003);
Social Security Administration, Social Security Administration
Strategic Plan 2003-2008; (Washington, D.C. 2003);
U.S. Department of Housing and Urban Development, Strategic Plan FY
2003-FY 2008; (Washington, D.C. 2003);
and Small Business Administration, SBA Strategic Plan, FY 2001-FY 2006;
(Washington, D.C. 2000).
[A] The 2003 plans for DOE and DOT were in draft form during the time
of our review.
[B] At the time of our review, the most recent SBA strategic plan was
for fiscal years 2001-2008. SBA released a new strategic plan for
fiscal years 2003-2008 in October 2003.
[End of table]
The remainder of this appendix discusses our observations on how the
quality of each of the agencies' strategic plans we reviewed has
changed since the agencies submitted their first draft strategic plans
in 1997. We did not independently verify or assess the information we
obtained from agency strategic plans. If an agency chose not to discuss
its efforts concerning elements in the strategic plan, it does not
necessarily mean that the agency is not implementing those elements.
Observations on Changes in the Quality of Education's Strategic Plan:
In our review of Education's June 1997 draft strategic plan,[Footnote
72] we found that the plan generally complied with GPRA and included
all but one of the six elements required by GPRA; it did not discuss
how Education's long-term goals and objectives would be related to its
annual performance goals. Also, we observed that the plan presented a
logical, fairly complete description of how Education intended to
achieve its mission, but a few areas could have been improved. In
comparison, Education's 2002-2007 strategic plan has improved on
several areas we identified in our 1997 review. However, we still found
areas where Education could improve. Table 6 summarizes these findings.
Table 6: Education's Progress in Addressing Required Elements of
Strategic Planning under GPRA:
Element of strategic planning: Mission statement;
Included in initial draft strategic plan: Yes. Mission statement
clearly and briefly explained why the agency exists, what the agency
does, and how it performs its work;
Included in current strategic plan: Yes. Mission statement is outcome
oriented, comprehensive, covers all of the agency's functions and
activities, and is the same as in the 1997 draft plan.
Element of strategic planning: Long-term goals;
Included in initial draft strategic plan: Yes. These goals were related
to the mission and were results oriented, but did not appear to reflect
civil rights enforcement and monitoring responsibilities;
Included in current strategic plan: Yes. The goals have changed, but
are related to each other and the mission and are results oriented.
They now additionally reflect civil rights enforcement and monitoring
responsibilities.
Element of strategic planning: Strategies;
Included in initial draft strategic plan: Yes. The plan outlined
strategies to achieve goals overall and to hold managers accountable
for achieving objectives, and generally described some of its resource
requirements throughout the plan;
Included in current strategic plan: Yes. The plan includes strategies
that are linked to the goals. Several strategies relate to resource
alignment to achieve outcomes, but the actual resources required are
not always specified.
Element of strategic planning: Relationship between long-term goals and
annual goals;
Included in initial draft strategic plan: No. Education did not discuss
the relationship between its strategic plan goals and those to be
included in its annual plan, but indicated this would be done once its
annual plan was prepared;
Included in current strategic plan: Yes. Plan contains the annual
performance measures and targets, which represent the annual
performance goals, and aligns them with long-term goals, with which
they have a logical relationship.
Element of strategic planning: External factors;
Included in initial draft strategic plan: Yes. The plan generally
described factors outside program scope and responsibilities that could
negatively affect Education's ability to achieve goals, but factors
were not directly linked to particular goals;
Included in current strategic plan: Yes. The plan adequately describes
external factors that could affect achieving its goals and they are
linked to the particular goals. The plan also briefly discusses
strategies to ameliorate the effects of a number of these factors.
Element of strategic planning: Evaluations;
Included in initial draft strategic plan: Yes. Education said it would
provide detailed descriptions of supporting evaluations once it
consulted with Congress, completed the strategic plan, and agreed on
performance indicators. The plan indicated a commitment to using
evaluations, listing evaluations it intended to use to develop sound
measures, but did not describe the evaluations;
Included in current strategic plan: No. Because of the comprehensive
revamping of Education's strategic plan in 2002, its program evaluation
plan was completely restructured and was not released until it was
included in the 2004 annual plan,[A] which contains information on its
new directions for program evaluation.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-176R];
U.S. Department of Education, Office of the Deputy Secretary, Planning
and Performance Management Service, U.S. Department of Education
Strategic Plan 2002-2007; (Washington, D.C. 2002).
[A] According to OMB Circular No. A-11, Part 6, Preparation and
Submission of Strategic Plans, Annual Performance Plans, and Annual
Program Performance Reports, June 2002, general goals, which are
multiyear and long term, are synonymous with general objectives, and
either term can be used. The objectives in Education's strategic plan
are such multiyear, long-term objectives, and are referred to in our
report as the agency's "long-term goals." OMB's Circular A-11 also
indicates that some agencies include strategic goals in their strategic
plan, which represent overarching statements of aim or purpose whose
achievement cannot be determined and which can be used to group outcome
or output goals. Education's strategic goals meet this description.
[End of table]
Current Strategic Plan Strengths and Improvements from Fiscal Year 1997
Draft Strategic Plan:
Education's current mission, "to ensure equal access to education and
to promote educational excellence throughout the nation," is the same
comprehensive, outcome-oriented mission that was included in its 1997
draft plan. All of Education's functions and activities are covered by
it. The plan's long-term goals[Footnote 73] are expressed so as to
allow Education and Congress to assess whether they are being achieved.
Moreover, in contrast to findings in our review of Education's 1997
Draft Strategic Plan, Education's civil rights responsibilities--
including enforcing five civil right statutes that ensure equal
educational opportunity for all students, regardless of race, color,
national origin, sex, disability, or age--appear to be reflected, at
least in part, in the current plan's goals. For example, one long-term
goal is to reduce the gaps in college access and completion among
student populations differing by race/ethnicity, socioeconomic status,
and disability while increasing the educational attainment of all.
Another is to enhance the literacy and employment skills of American
adults. Under the latter, the plan includes a strategy to work with
state vocational rehabilitation agencies to ensure implementation of
standards that will assist individuals with disabilities in obtaining
high-quality employment outcomes. As in the past, some goals are
targeted at results for which Education has limited direct control and
are, instead, greatly influenced by third parties. However, Education
recognizes this situation and shows its intent to work closely with its
partners.
Education's current plan provides some information on linking results
and day-to-day activities within the department. For example, the plan
says employees will be held accountable for implementation and success
from top to bottom and senior officers will have performance contracts
linked to the plan and be recognized for achieving results. In
addition, the strategy to foster a customer service orientation by
ensuring that states, districts, and other partners receive timely
responses to inquiries; to assign senior officers to develop
relationships with individual states; and to create a customer support
team to respond to issues, seems logically linked to the day-to-day
activities of managers and staff. However, the link between results and
day-to-day activities is not apparent for most of the goals and their
strategies.
The current plan includes several annual performance measures that are
at least related to how well information technology is supporting
strategic and program goals, as required by the Clinger-Cohen Act. For
example, for its long-term goal to modernize the Federal Student
Assistance (FSA) Programs and reduce their high-risk status, the plan
contains a measure on the integration of FSA processes and systems that
work together to support FSA program delivery functions. Commendably,
Education includes some goals related to reducing agency program
unintended negative effects, such as a goal and related measure to
reduce the data collection and reporting burden.
Education's current plan shows great improvement on recognizing and
addressing external factors. Beyond adequately describing external
factors that could affect achieving its goals and directly linking
these factors to particular goals, the plan also briefly describes
strategies to ameliorate the effects of a number of factors. For
example, for an external factor on teacher certification under a goal
on reading, the plan says that Education "will work with the states and
national accreditation bodies to encourage the incorporation of
research-based reading instruction into teacher certification
requirements." In addition, the plan includes strategies indicating
that Education monitors some internal factors that could affect
achievement of long term goals. Moreover, the plan recognizes the
overarching critical external factor for Education--that it depends
greatly on third parties who often control the results the department
is trying to achieve--and its strategies related to the goal-specific
external factors often reflect this.
In our 1997 assessment of Education's draft plan, we commented that
although the plan identified several management challenges the
department would face in the coming years, it provided little detail
about them and how they would be addressed. In our January 2001
performance and accountability series, we identified four department-
specific and two governmentwide challenges that we said Education
needed to meet.[Footnote 74] Education's current strategic plan
includes some goals, measures, and strategies that could be used to
address these challenges. For example, Education's goal to "develop and
maintain financial integrity and management and internal controls" and
this goal's strategies and measures are directly related to the
management challenge we identified on improving financial management.
One of the measures under this goal is "the achievement of an
unqualified audit opinion," an important issue in our identification of
financial management weaknesses as a challenge in 2001, and Education
received an unqualified audit opinion in early 2003. Moreover, the
current strategic plan includes goals and strategies meant to improve
Education's strategic human capital management and strengthen
information technology security, two important governmentwide high-
risk areas we identified in 2001.
Critical Strategic Planning Issues Needing Further Improvement:
In our report on Education's 1997 draft strategic plan, we found that,
although the department had numerous crosscutting programs and
activities, the plan had identified key interagency activities for some
programs but not for others. For example, we said that the plan did not
identify or discuss activities for postsecondary programs for which the
Higher Education Act of 1965 required coordination. Education's current
plan includes an appendix, which highlights collaborative initiatives
under each of the department's strategic goals, with some activities
related to postsecondary education, including most of those mentioned
in our 1997 report. However, as the plan states, the appendix presents
a brief overview of the highlights of some of its collaborative
initiatives with other agencies.
In our 1997 review, we stated that some resource requirements were
described throughout the plan. In the current plan, while a number of
strategies under the department's strategic goal to establish
management excellence are related to aligning resources to achieve
outcomes, the actual resources required--such as human, capital, and
information--are not always specifically indicated. The exception is
for information resources, for which a number of strategies discuss the
information resources that will be required to address the related
goal. For example, under the goal to develop and maintain financial
integrity and management and internal controls, the plan says that
Education will "implement a new financial system capable of producing
timely and reliable financial data and reconcile systems to the general
ledger." Moreover, while the plan stresses accountability throughout,
it only refers to providing the authority needed to achieve results
once.[Footnote 75] In addition, consideration of alternate strategies
for achieving goals is not discussed.
In our 1997 review, we reported that Education said it would provide
detailed descriptions of supporting evaluations once it consulted with
Congress, completed the strategic plan, and agreed on performance
indicators. The draft plan indicated Education's commitment to using
evaluations and listed evaluations and strategies it intended to use to
develop sound measures, but did not describe the evaluations. The
current plan does not include descriptions of supporting evaluations or
a future program evaluation schedule. According to Education officials,
this was not done because, with the complete revamping of the strategic
plan based on passage of the No Child Left Behind Act of 2001,[Footnote
76] Education was set on a course to totally restructure its evaluation
program, but could not do so in time to include it in the strategic
plan. Consequently, Education instead included information about its
new directions for program evaluation studies and a schedule of
evaluations in its 2004 annual plan. The schedule, however, lacked a
timetable, except for stating whether the evaluations were new or
continuing. The current strategic plan indicates, in some cases, the
use or planned use of program evaluation findings to develop or revise
components of the plan. For example, for long-term goals on various
types of academic achievement, data from the National Assessment of
Educational Progress were identified as having been used to set related
targets.
Observations on Changes in the Quality of DOE's Strategic Plan:
Overall, DOE's draft 2003 strategic plan meets the required elements of
GPRA and has improved greatly over its 1997 draft strategic plan, as
shown in table 7. DOE made improvements to its plan by establishing
results-oriented and measurable objectives, and addressing elements
that were not included in the department's 1997 plan, such as reporting
on external factors. Although DOE has shown improvement, a few elements
in the plan could still be enhanced. For instance, further improvement
could be made to the mission statement so that it better addresses the
department's major activities, and additional information could be
included in DOE's strategies to achieve its goals, such as management
accountability.
Table 7: DOE's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Element of strategic planning: Mission statement;
Included in initial draft strategic plan: Yes. DOE's mission was
results oriented, met a public need, and covered the agency's major
activities;
Included in current draft strategic plan: Yes. DOE's mission is results
oriented and meets a public need, but it does not address the
department's major activities related to energy supply and
conservation.
Element of strategic planning: Long-term goals;
Included in initial draft strategic plan: Yes. Long-term goals covered
mission and major functions of the agency. Goals and objectives were
results oriented but not measurable;
Included in current draft strategic plan: Yes. Long- term goals cover
the mission and major functions of the agency. Objectives, referred to
as intermediate goals, are results oriented and measurable.
Element of strategic planning: Strategies;
Included in initial draft strategic plan: Yes. The plan included
strategies and measures to evaluate the results of the strategies, but
was missing information on linkages to day-to-day activities, and the
extent to which managers have the knowledge, skills, and abilities to
implement the strategies;
Included in current draft strategic plan: Yes. Strategies to achieve
each goal are included in the plan, along with intermediate goals to
measure success, but information on linkages to day-to-day activities,
and the extent to which managers have the knowledge, skills, and
abilities to implement the strategies, is not included.
Element of strategic planning: Relationship between long-term goals and
annual goals;
Included in initial draft strategic plan: No. Relationship between the
long-term goals and annual performance goals was missing in the draft
plan;
Included in current draft strategic plan: Yes. The draft plan provides
a brief description of the relationship between the long-term strategic
goals and the annual performance goals.
Element of strategic planning: External factors;
Included in initial draft strategic plan: No. Key external factors were
not addressed in the draft plan;
Included in current draft strategic plan: Yes. Key external factors and
how they could affect the ability to achieve each goal are identified
in the draft plan.
Element of strategic planning: Evaluations;
Included in initial draft strategic plan: No. The impact of program
evaluations on the development of strategic goals was not included in
the draft plan;
Included in current draft strategic plan: No. The impact of program
evaluations on the development of strategic goals is not discussed
thoroughly.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-199R];
and Department of Energy, 2003 Strategic Plan; (Draft); (Washington,
D.C. 2003).
[End of table]
Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan:
In its 2003 draft strategic plan, DOE has made significant improvements
in developing measurable objectives, referred to in its plan as
intermediate goals. GAO's review of DOE's 1997 draft strategic plan
found that objectives related to DOE's long-term goals were stated in
ways that would make it difficult to measure whether they were being
achieved. In the 2003 draft plan, the objectives are stated in ways
that will enable DOE to measure its progress in achieving goals. For
example, to meet the goal of enhancing energy security through various
means, one of DOE's affiliated objectives is to ensure that throughout
DOE's 25-year planning period, the Strategic Petroleum Reserve is ready
to supply oil at a sustained rate of 4.3 million barrels per day for 90
days within 15 days notice by the President.
In addition, DOE has improved its draft 2003 strategic plan by
including elements that were not included in its 1997 draft plan. These
elements consisted of (1) identifying external factors and (2)
describing the relationship between long-term and annual goals. For
each of its long-term goals, DOE identified external factors, such as
reduced funding, lack of scientific talent, and unpredictable
technological developments, that could affect its ability to achieve
its goals. The strategic plan also included a description of the
relationship between the long-term strategic goals and the annual
performance goals. The plan included a diagram that showed an example
of a strategic goal, its associated objectives, and how they are
related to the annual performance goals and targets. However, the plan
could be improved if all annual performance goals were discussed in the
plan and linked to each strategic goal so that it would be clear how
annual performance goals would be used to gauge performance. DOE staff
stated that a description of all actual annual performance goals was
not something that they thought should be included in the strategic
plan because the annual goals differ each year.
Finally, our past and current reviews of DOE's 1997 and 2003 draft
strategic plans found that DOE addressed performance and accountability
challenges that we had previously identified. In January 2003, we
identified six areas where DOE's management attention was needed: (1)
addressing security concerns, (2) revitalizing infrastructure, (3)
improving contract management, (4) managing the nuclear weapons
stockpile, (5) cleaning up radioactive and hazardous wastes, and (6)
enhancing leadership in meeting energy needs. All areas were addressed
in the 2003 draft strategic plan, with the exception of improving
contract management. For example, for the challenge of enhancing
leadership in meeting energy needs, one of the intermediate goals
requires DOE to develop and demonstrate technologies that can reduce
emissions by more than 70 metric tons of carbon and equivalent
greenhouse gases by 2012.
Critical Strategic Planning Issues Needing Further Improvement:
There are three elements in DOE's draft 2003 strategic plan that
requires further improvement. To begin with, although DOE's mission is
results oriented, it was revised from the 1997 draft strategic plan and
does not address the department's major activities related to energy
supply and conservation. These activities account for approximately 10
percent of DOE's $23.4 billion fiscal year 2004 budget request. Our
review of the 1997 draft strategic plan found that DOE's mission
addressed all of its major activities.
In addition, the impact of program evaluations on the development of
strategic goals could be discussed in greater detail. The strategic
plan stated that internal, GAO, and the Inspector General (IG)
evaluations were used as resources to develop the draft strategic plan,
but specific program evaluations were not identified. A schedule of
future program evaluations was also not discussed in the strategic
plan. According to DOE, there is no plan to include a table of reports
and evaluations that were used to develop the goals for the strategic
plan because the evaluations were from a prior administration, and when
the administration changes, it usually does not use evaluations from
past administrations.
DOE could also enhance the strategies included in its plan by providing
information on how goals are linked to the department's day-to-day
activities, and the extent to which managers have the knowledge,
skills, and abilities to implement the strategies. None of this
information was included in the 1997 and 2003 draft strategic plans.
According to DOE officials, in drafting the plan, their goal was to
keep the plan at a high level and this additional information would
require more detail than what is needed for a strategic plan. For
example, one official stated that a description of linkages to day-to-
day activities was not discussed in the strategic plan because it would
conflict with the performance plan and budget justification, which is
where the information can be found. As we have stated previously,
without this information, it is difficult to judge DOE's likelihood of
success in achieving the goals or the appropriateness of the
strategies.[Footnote 77]
Finally, DOE's draft strategic plan could be improved by identifying
programs and activities that are crosscutting or similar to those of
other agencies. In the 2003 draft plan, crosscutting activities are
only identified for one goal related to science. According to one DOE
official, as the strategic goals were being developed, DOE staff took
crosscutting activities into consideration. As we stated in 1997,
unless DOE addresses crosscutting issues in its plan, Congress cannot
be assured that federal programs are working effectively.
Observations on Changes in the Quality of HUD's Strategic Plan:
Overall, HUD has made progress in developing its strategic plan for
fiscal years 2003 through 2008 as required under GPRA. In 1997, we
stated that HUD's draft strategic plan did not cover the six components
required under GPRA.[Footnote 78] HUD's fiscal year 2003-2008 strategic
plan addressed several issues we had previously identified, such as
making sure the mission statement is linked to the department's major
operations and functions; including outcome-related goals and
objectives for the department's functions and operations; generally
providing a description of how it will achieve its goals and
objectives; identifying key external factors affecting the achievement
of departmental goals; and explaining how HUD will coordinate with
other agencies to address crosscutting issues. However, HUD could
improve its strategic plan by explaining the relationship between the
long-term and intermediate performance measures listed for each
strategic goal and discussing how it used program evaluations to
develop the plan. Table 8 summarizes these findings.
Table 8: HUD's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Element of strategic planning: Mission statement;
Included in initial draft strategic plan: No. HUD's mission statement
did not cover the major program functions and operations of the agency;
Included in current strategic plan: Yes. HUD's mission statement covers
the agency's major program functions and operations and relevant
statutes.
Element of strategic planning: Long-term goals;
Included in initial draft strategic plan: Yes. The plan included eight
strategic objectives, and they generally covered the department's major
functions and operations;
Included in current strategic plan: Yes. The plan includes long-term
goals that cover the major programs and functions of the agency.
Element of strategic planning: Strategies;
Included in initial draft strategic plan: No. The strategic plan lacked
an adequate description of how its strategic objectives would be
achieved;
Included in current strategic plan: Yes. The strategic plan discusses
the means and strategies to address each strategic goal, including
plans to address human capital issues critical to carrying out its
mission.
Element of strategic planning: Relationship between long-term goals and
annual goals;
Included in initial draft strategic plan: No. HUD's strategic plan
provided limited examples of annual performance goals under each of its
strategic objectives, but it did not describe the relationship between
them;
Included in current strategic plan: Yes. However, some long-term
performance measures do not appear to have corresponding intermediate
measures, and in other instances it is not clear how HUD will measure
progress towards its goals.
Element of strategic planning: External factors;
Included in initial draft strategic plan: No. HUD briefly discussed the
external factors in its draft strategic plan without linking them to
specific strategic objectives;
Included in current strategic plan: Yes. HUD describes the external
factors that could affect achieving its strategic objectives.
Element of strategic planning: Evaluations;
Included in initial draft strategic plan: No. HUD's strategic plan did
not include information on program evaluations;
Included in current strategic plan: No. HUD does not describe how
program evaluations were used to prepare its strategic plan. Although
it mentions that program evaluations will be used to advance key policy
objectives, it states that there are no fixed timetables for when these
evaluations will take place.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-224R];
and U.S. Department of Housing and Urban Development, Strategic Plan FY
2003-FY 2008.; (Washington, D.C. 2003).
[End of table]
Strategic Plan Strengths and Improvements from Fiscal Year 1997 Draft
Plan:
In its most recent strategic plan for fiscal years 2003-2008, HUD has
made progress in crafting a mission statement that generally covers its
major program functions, operations, and relevant statutes that
authorize its programs. In addition, the current plan builds upon the
strength of its first draft strategic plan by including the strategic
objectives that cover the department's major functions and operations.
In contrast to its draft strategic plan of 1997, HUD's most recent plan
provides a general description of how its strategic objectives will be
achieved. For example, HUD lists the means and strategies following
each strategic goal and supporting objectives to describe how it will
carry out its activities. In addition, the strategic plan identifies a
few legislative and regulatory changes HUD will pursue to meet its
objectives, such as a new tax credit for developers of affordable
housing and expanded eligibility for the Assisted Living Conversion
Program. To carry out its work, HUD acknowledges in its plan that it
needs to recruit and retain its current workforce to ensure the proper
skills and abilities needed to carry out its mission.
HUD's most recent strategic plan also describes key factors external to
the department and beyond its control that could significantly affect
the achievement of its objectives. For example, for its goal of
Promoting Decent and Affordable Housing, HUD identifies the impact of
broad economic factors on opportunities for low-income workers as a
factor that will affect the department's ability to assist renters that
depend on HUD's programs to make progress towards self sufficiency.
HUD's strategic plan also provides a general description of current
program evaluations and mentions that the results of these evaluations
will support key policy objectives within the department.
Addressing a shortcoming of its draft strategic plan in 1997, HUD's
current plan generally explains how it will coordinate with other
agencies to address crosscutting problems. For example, the Interagency
Council on the Homeless (ICH), created by the Secretaries of HUD, the
Department of Health and Human Services (HHS), and Veterans Affairs
(VA), will continue to work to identify the obstacles homeless people
face to enroll in the main service programs and recommend specific
changes--legislative, policy, and procedural--that would make federal
supportive service programs more accessible to them.
Lastly, HUD improved its discussion of how it plans to address
performance and accountability challenges we have raised. In January
2003, we reported that programmatic and financial management
information systems and human capital issues were HUD's performance and
accountability challenges.[Footnote 79] While in the past HUD
acknowledged some of these problems and described how they would be
addressed, its plans for management reform were not fully integrated
into its draft strategic plan. In contrast, one of HUD's strategic
goals in its current strategic plan, "Embrace High Standards of Ethics,
Management, and Accountability," lists specific actions the department
will take to address these challenges.
Critical Strategic Planning Issues Needing Further Improvement:
HUD has made progress in linking its strategic objectives to both long-
term goals and intermediate measures in its strategic plan; however,
long-term performance measures are not consistently linked to a
corresponding intermediate measure, making it difficult for the reader
to understand how the department will measure progress towards its
strategic goals. For example, HUD projects that the percentage of
architects and builders indicating awareness of the design and
construction requirements of the Fair Housing Act will increase through
fiscal year 2008. However, HUD does not mention in the plan how many
architects or engineers it will survey to establish a baseline of
awareness or to gauge progress made towards this goal in subsequent
years. HUD officials explained that only those long-term goals that are
critical to being achieved within the next 1-2 years have corresponding
intermediate performance measures. Therefore, it is somewhat unclear
how HUD plans to assess progress made towards its broader goal of
"Ensure Equal Opportunity in Housing.":
Similarly, the strategic plan does not always provide a clear picture
of how HUD will be able to measure progress towards its strategic
goals. For example, the plan states that HUD's Community Development
Block Grants (CDBG) Program will create or retain 400,000 jobs by
fiscal year 2008; HUD's intermediate measure for fiscal year 2004 is
84,000 jobs. These estimates are based on the average jobs created per
grant dollar reported by grantees. However, HUD does not mention how it
will be able to discern between those jobs created by CDBG and those
created by other means.
Similar to our previous finding, HUD's current strategic plan does not
describe how program evaluations were used to develop its strategic
goals or other components of its strategic plan, and does not include a
schedule for future evaluations. While the plan mentions that program
evaluations will be used to advance key policy objectives, it states
that there are no fixed timetables for when these evaluations will take
place.
Observations on Changes in the Quality of SBA's Strategic Plan:
In our review of SBA's March 1997 draft strategic plan,[Footnote 80] we
found that the plan did not meet two of the six requirements set forth
by GPRA: (1) a discussion of the relationship between the long-term
goals and objectives and the annual performance goals and (2) a
description of how program evaluations were used to establish or revise
strategic goals and a schedule for future program evaluations. In
addition, the four elements the plan contained could have better
conformed to GPRA's requirements and OMB guidance. In comparison, SBA's
2001-2006 strategic plan has improved on several areas we identified in
our 1997 review.[Footnote 81] However, we still found areas that SBA
could improve. Table 9 summarizes these findings.
Table 9: SBA's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Element of strategic planning: Mission statement;
Initial draft strategic plan: Yes. The mission statement was results
oriented. However, it did not directly incorporate key aspects of SBA's
legislative mandate and it did not encompass one of SBA's significant
activities;
Current strategic plan: Yes. SBA's mission statement now specifically
mentions its mandate and encompasses SBA's disaster loan program to
families and businesses.
Element of strategic planning: Long-term goals;
Initial draft strategic plan: Yes. Generally the strategic goals
covered the major functions and operations of SBA. Most of SBA's
strategic goals were expressed as processes, not as outcomes;
Current strategic plan: Yes. The strategic goals are outcome oriented
and cover the major functions and operations of SBA.
Element of strategic planning: Strategies;
Initial draft strategic plan: Yes. However, the strategies were too
vague or general to assess whether or not they would help achieve SBA's
goals. Also, the strategies and goals could have been linked more
explicitly;
Current strategic plan: Yes. The strategies will generally help SBA
achieve its goals. The strategies are now listed by strategic goal and
related objective. The plan discusses some of the resources SBA needs
to achieve its goals.
Element of strategic planning: Relationship between long-term goals and
annual goals;
Initial draft strategic plan: No. The linkage between proposed
performance measures, strategies, and objectives was unclear;
Current strategic plan: Yes. The plan lists performance measures by
objectives for each strategic goal. Logic models show the relationship
between measures and outcomes.
Element of strategic planning: External factors;
Initial draft strategic plan: Yes. The plan listed a number of external
factors. A discussion of how the external factors would be taken into
account when assessing progress toward goals was not included in the
plan;
Current strategic plan: Yes. The plan lists several external factors.
Generally, strategies to ameliorate the effects of these factors are
included. However, SBA does not discuss how external factors will be
taken into account when assessing progress toward goals.
Element of strategic planning: Evaluations;
Initial draft strategic plan: No. The plan did not describe how program
evaluations were used to establish or revise strategic goals or include
a schedule for future program evaluations;
Current strategic plan: No. The plan states that lessons learned from
SBA's program evaluations have influenced its strategic direction.
Examples of future program evaluations are given, but a specific
schedule of when these evaluations are to occur is not included.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-205R];
and Small Business Administration, SBA Strategic Plan, FY 2001-FY 2006;
(Washington, D.C. 2000).
[End of table]
Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan:
SBA's current mission statement is an improvement from the one
contained in its 1997 draft strategic plan. In our July 1997 report on
SBA's draft strategic plan, we noted that the mission statement could
be improved by more directly incorporating key aspects of SBA's
legislative mandate to aid, counsel, assist, and protect the interests
of small businesses. In addition, the mission statement did not
encompass one of SBA's significant activities-making loans to
individuals. The mission statement in the 2001-2006 strategic plan now
includes both of these items.
The long-term, or strategic, goals of SBA's 2001-2006 strategic plan
are (1) helping small businesses succeed, (2) helping Americans recover
from disasters, and (3) modernizing SBA. These three strategic goals
are outcome oriented. We stated in our previous report on SBA's draft
strategic plan that only two of SBA's seven strategic goals described
outcomes. The rest of the goals were expressed as processes.
SBA's 2001-2006 strategic plan shows significant improvement in its
strategies. In 1997 we stated that the plan could be improved by making
the linkage between the strategies and goals/objectives more explicit.
Objectives were listed as a group, followed by the strategies, which
were also listed as a group. The 2001-2006 plan describes the
strategies, objective by objective, making the linkage clear. The
strategies contained in SBA's 1997 plan consisted entirely of one-line
statements and most were too vague or general to enable an assessment
of whether or not they would help achieve the goals and objectives in
the plan. While many of the strategies listed in SBA's 2001-2006 plan
are only one-or two-sentence statements, all of the strategies seem to
aid in achieving SBA's goals and objectives. For example, one strategy
for the objective of providing entrepreneurial development assistance
is to train SBA personnel in outreach and business development
approaches. This strategy would presumably aid in making staff more
competent at helping small businesses develop. The current plan also
includes some of the resources needed to achieve SBA's goals. These
resources include a table showing SBA's fiscal year 2001 budget request
and its plans to modernize its information technology systems.
The 2001-2006 strategic plan also makes a clearer connection between
strategic goals and annual performance measures that will be used to
gauge progress in achieving strategic goals. SBA shows this
relationship by linking its performance output measures with the
intended outcomes of SBA's programs for two of the strategic goals. For
the other strategic goal, "Modernizing SBA," SBA lists the performance
measures it will use by objective. For example, the performance measure
"personnel trained or retrained" will gauge progress in achieving SBA's
human capital investments objective. This shows improvement over the
draft strategic plan, which listed the performance measures as a group
without showing how they were related to SBA's strategic goals.
Similar to its 1997 draft plan, SBA's 2001-2006 plan specifies external
factors that could affect the achievement of its strategic goals. The
plan lists eight external factors associated with two of its three
strategic goals; no external factors were associated with the third
goal. The plan includes strategies that seem to be intended to mitigate
the effect of six of these external factors, while two of the external
factors, congressional support and public support, do not seem to be
addressed. An example of an external factor that seems to be addressed
by a strategy is economic conditions, listed for the strategic goal
"helping small businesses succeed." One of the strategies for this
strategic goal is to determine economic trends and conditions. SBA
states that it tracks economic trends that affect small businesses and
the contribution small businesses make to the economy. SBA then brings
these data to the attention of policymakers.
We noted in our 1997 report that SBA did not explicitly address the
relationship of SBA's activities to similar activities in other
agencies and provided no evidence of coordination. In contrast, SBA's
current strategic plan includes a section on crosscutting issues. This
section contains discussions of innovation and research assistance,
international trade assistance, business development assistance,
veterans affairs, and disaster assistance. One example of coordination
is the U.S. Export Assistance centers, which combine the trade-
promotion and export-finance resources of SBA, the Department of
Commerce, the Export-Import Bank, and in some locations, the Agency for
International Development and the Department of Agriculture.
Finally, SBA's 2001-2006 strategic plan generally addresses performance
and accountability challenges that we have previously identified. For
example, we noted in our January 2001 report on major management
challenges and program risks[Footnote 82] that SBA needed to continue
to improve oversight of its lending partners to correct oversight
weaknesses. The plan identifies lender oversight as a management
challenge and states that SBA has developed and implemented a safety
and soundness oversight program for Small Business Lending companies,
institutionalizing the process through the Office of Lender Oversight.
This is an improvement over SBA's draft strategic plan, which we
reported did not clearly address previously identified management
problems.
Critical Strategic Planning Issues Needing Further Improvement:
In 1997, we noted that SBA's draft strategic plan did not include a
discussion of how the external factors would be taken into account when
assessing progress toward goals. This observation holds true for the
current strategic plan. For the external factor mentioned above,
economic conditions, the plan states that if the economy remains
strong, surety bond guaranties will remain constant or decrease, but if
the economy deteriorates, demand will increase. However, the plan does
not state how SBA will assess success or failure in meeting its goals
in relation to this factor.
In its 1997 draft plan, SBA did not describe how program evaluations
were used to establish or revise strategic goals or include a schedule
for future program evaluations. In the 2001-2006 plan, SBA states, "We
have used lessons learned in our performance monitoring and program
evaluations to influence the strategic direction contained in this
plan." The plan includes findings from six completed program
evaluations; however, no further detail is given as to how these
program evaluations were used to establish or revise the strategic
goals. While the current plan gives examples of future program
evaluations, such as conducting a benchmark study on the HUBZone
program to assess the changes in employment and investment in
distressed urban and rural communities, it does not include a schedule
of future evaluations. SBA states that for the next several years, the
agency plans to systematically review programs that offer the most
financial risk to the government and also the programs that can offer
tips on how to improve efforts.
Observations on Changes in the Quality of SSA's Strategic Plan:
SSA's strategic plan for 2003-2008 is well structured and contains all
of the required elements under GPRA. In 1997, we noted that SSA's draft
strategic plan contained all six required components, but suggested a
number of ways it could be strengthened. SSA has addressed some of the
issues we previously identified, such as associating specific programs
with goals and identifying external factors that may affect goal
achievement. SSA could further improve its strategic plan through (1)
ensuring that its strategic objectives will assist SSA in achieving its
strategic goals, (2) explicitly describing the effect of external
factors on goal attainment, (3) providing timetables or schedules for
achieving results, (4) providing details on how each performance and
accountability challenge will be addressed, (5) clearly explaining how
program evaluations were used in formulating the strategic plan, and
(6) discussing the manner in which SSA has coordinated with other
agencies, especially those that serve the same beneficiaries. Table 10
summarizes SSA's progress in addressing the required elements of GPRA.
Table 10: SSA's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Element of strategic planning: Mission statement;
Included in initial draft strategic plan: Yes. SSA's mission statement
was appropriate and reflective of its new status as an independent
agency;
Included in current strategic plan: Yes. SSA's mission statement has
not changed substantially.
Element of strategic planning: Long-term goals;
Included in initial draft strategic plan: Yes. Long-term goals were
established. However, the relationship between long-term goals and
specific programs was unclear and did not identify the results to be
achieved;
Included in current strategic plan: Yes. The goals' relationship to
specific programs is more defined, but SSA's goal for achieving
solvency of the social security system is ambitious, given SSA's
mission and responsibilities. Key outcomes are identified for each
goal. SSA's major management challenges are not all clearly linked to
individual goals or objectives.
Element of strategic planning: Strategies;
Included in initial draft strategic plan: Yes. The strategic plan was
generally complete with regard to processes and technology, but did not
include timetables or schedules. Some of the success was predicated on
changes in processes or technology improvements;
Included in current strategic plan: Yes. SSA added some timetables, but
the required resources are not specified. Some of the success is still
predicated on technological improvements.
Element of strategic planning: Relationship between long-term goals and
annual goals;
Included in initial draft strategic plan: Yes. SSA provided numerous
performance measures relating to strategic goals and objectives, and
plans for developing new measures were discussed. It was sometimes
difficult to link measures with their appropriate objectives. It was
also hard to discern which objectives did not yet have performance
goals. Some data were expressed by program, while other data were
aggregated;
Included in current strategic plan: Yes. SSA provided one or more key
outcomes for each strategic objective. SSA notes that success in
meeting the objectives will be measured in the annual performance plans
by progress in achieving the key outcomes.
Element of strategic planning: External factors;
Included in initial draft strategic plan: Yes. The report mentioned
several key external factors, but did not explicitly link factors to
general goals and state how they could have affected goal attainment.
Also, the report did not discuss needed changes (by Congress) to ensure
solvency;
Included in current strategic plan: Yes. External (environmental)
factors are listed in a separate section. However, the plan does not
explicitly link factors to general goals and state how they could
affect goal attainment. Specific effects are not discussed--most
examples are vague.
Element of strategic planning: Evaluations;
Included in initial draft strategic plan: Yes. The report contained a
broad discussion of program evaluations, but the evaluations were not
clearly described. Also, SSA did not describe how the evaluations were
used to establish or revise specific goals/objectives. Finally, there
was no schedule for completing future evaluations or methodologies;
Included in current strategic plan: Yes. Future evaluations (with brief
descriptions) are listed, but there is no discussion of how current
evaluations are used to establish or revise specific goals/objectives.
The plan states that SSA used internal and external (GAO, IG)
evaluations to determine strategic plans and objectives.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-97-179R];
and Social Security Administration, Social Security Administration
Strategic Plan 2003-2008, (Washington, D.C. 2000).
[End of table]
Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan:
SSA's mission statement changed very little between 1997 and 2003. OMB
Circular A-11 notes that the mission statement should be brief and
define the basic purpose of the agency, with particular focus on its
core programs and activities. SSA's statement conforms to this
guidance--it reads, "To advance the economic security of the nation's
people through compassionate and vigilant leadership in shaping and
managing America's social security programs."[Footnote 83]
In 1997, the relationship between SSA's long-term goals and specific
programs was unclear and did not identify the specific results to be
achieved. Since that time, SSA has improved this linkage and better
articulated intended results, including quantifiable goals. For
example, as part of its strategic goal to ensure superior stewardship
of Social Security programs and resources, SSA notes that one of its
key outcomes is to increase Supplemental Security Income (SSI) payment
accuracy to 96 percent (free of preventable error) by 2008.
SSA improved upon its linkage between long-term goals and annual goals
in its fiscal year 2003-2008 strategic plan. Under each strategic goal
in this plan, SSA provided one or more key outcomes for each strategic
objective; in its 1997 draft strategic plan, we found it difficult to
link measures with the appropriate objectives and discern which
objectives did not yet have performance goals.
Critical Strategic Planning Issues Needing Further Improvement:
Not all of SSA's strategic objectives and associated performance
measures will allow SSA to achieve its related strategic goals.
Specifically, the solvency goal in SSA's current strategic plan reads,
"To achieve sustainable solvency and ensure Social Security programs
meet the needs of current and future generations," but the sole
associated objective--through education and research efforts, support
reforms to ensure sustainable solvency and more responsive retirement
and disability programs--will not allow SSA, on its own, to reach that
goal. While SSA's mission is to advance the economic security of the
nation's people, it is not unilaterally responsible for achieving
solvency in social security programs.
An agency's strategic plan is expected to contain strategies for
achieving the goals articulated. In 1997, SSA's strategic plan was
generally complete with regard to processes and technology, but did not
include timetables or schedules for results. While the current
strategic plan contains processes, anticipated progress in technology,
and some timetables, it does not contain timetables or schedules for
all of the results. For example, as part of its strategic objective to
"efficiently manage Agency finances and assets and effectively link
resources to performance outcomes," SSA's key outcomes include (1)
competing or converting 50 percent of commercial positions and (2) to
"get to green" on all five President's Management Agenda (PMA) items.
SSA has neither identified the required resources to achieve these
goals nor has it identified a time frame for achieving them.
In our review of SSA's 1997 draft strategic plan, we noted that SSA
described several key external factors that may affect its programs,
but did not explicitly link such factors to its general goals and state
how these factors could affect goal attainment. In the current
strategic plan, SSA identifies environmental (external) factors:
demographics, health and disability trends, technological advances, and
workforce trends. However, as we found in our earlier review, the
effects of these factors on specific performance goals are not
specified, even though SSA notes that they drive the development of
such goals.
SSA noted that it considered major management challenges identified by
GAO when it determined its strategic goals and objectives, but not all
of these challenges are clearly addressed in the plan. While these
challenges are not clearly identified, SSA addresses them to some
degree throughout the plan. For example, SSA's strategic goal to
"Strategically manage and align staff to support SSA's mission"
addresses the governmentwide challenge of strategic human capital
management.[Footnote 84]
SSA includes a list of major strategic process and program evaluations
scheduled for the fiscal years 2003-2008 time period, organized by
strategic goal. However, SSA does not list ongoing evaluations or
mention how the results of these evaluations were used to prepare the
current strategic plan. SSA notes that many of the hundreds of process
and program evaluations conducted annually were designed to evaluate
and improve internal processes falling below the strategic level.
However, some of the ongoing evaluations are associated with specific
strategic goals; thus, their outcomes could be discussed in the context
of the strategic goals with which they are affiliated.
SSA's strategic plan contains a very limited discussion of its
interactions with other agencies that have similar goals or serve the
same beneficiaries. In an interview, SSA officials noted that SSA has
extensive interactions with other agencies on such issues as earnings
accuracy and medical information, but the level of interaction varies
by initiative. SSA's strategic plan would benefit from a broader
discussion of these interactions, especially if they were broken down
by initiative. For example, as part of the objective to increase the
accuracy of earnings records, SSA notes that it will collaborate with
the Internal Revenue Service (IRS) to achieve more accurate wage
reporting as part of its means and strategies to reduce the size of the
suspense file.[Footnote 85] It would be helpful if SSA offered more
details as to the nature and extent of its collaboration with IRS.
Observations on Changes in the Quality of DOT's Strategic Plan:
In our review of DOT's July 1997 draft strategic plan, we found that
the plan only met three of the six elements required by GPRA.[Footnote
86] The plan did not meet GPRA's requirements to describe (1)
strategies for achieving the goals, (2) a linkage between DOT's long-
term goals and annual performance goals, and (3) the external factors
that could significantly affect DOT's ability to achieve its goals.
Further, for the three elements that the plan did meet, each had
weaknesses. In comparison, DOT's 2003-2008 draft strategic plan has
improved on several areas we identified in our 1997 review.[Footnote
87] However, we still found areas where DOT could improve. Table 11
summarizes these findings.
Table 11: DOT's Progress in Addressing Required Elements of Strategic
Planning under GPRA:
Element of strategic planning: Mission statement;
Included in initial draft strategic plan: Yes. DOT's mission statement
was comprehensive and covers its major functions and operations;
Included in current draft strategic plan: Yes. The mission statement
continues to cover its major functions and operations and more
explicitly states DOT's statutory authority.
Element of strategic planning: Long-term goals;
Included in initial draft strategic plan: Yes. Five long-term goals
encompassed DOT's major functions and operations. However, it was not
clear as to how DOT would measure success for most of its goals;
Included in current draft strategic plan: Yes. Five strategic
objectives encompass DOT's major functions and operations. Outcome
goals and candidate performance measures for each strategic goal help
show how DOT will measure success.
Element of strategic planning: Strategies;
Included in initial draft strategic plan: No. While the plan listed six
corporate management strategies for achieving its long-term goals, it
did not describe the operational processes, the skills, the technology,
and the resources required to meet them;
Included in current draft strategic plan: Yes. Strategies are listed by
strategic objective and include discussions on leadership, building
expertise, and technology.
Element of strategic planning: Relationship between long term-goals and
annual goals;
Included in initial draft strategic plan: No. The plan did not describe
how performance goals would be related to the long- term goals;
Included in current draft strategic plan: Yes. The plan includes
performance measures and refers to the performance plan for further
information on annual performance goals.
Element of strategic planning: External factors;
Included in initial draft strategic plan: No. Four external factors
were identified, but other key factors were not included in the plan.
Only one external factor was discussed in terms of how it could have
affected DOT's ability to accomplish its goals;
Included in current draft strategic plan: Yes. Several external factors
are listed for each strategic objective. Generally the plan gives
descriptions of how these factors could affect the achievement of
goals.
Element of strategic planning: Evaluations;
Included in initial draft strategic plan: Yes. Program evaluations used
in establishing goals and a schedule of future evaluations were
discussed. However, the plan did not provide enough information to
determine the scope and methodology or the key issues to be addressed
in future evaluations;
Included in current draft strategic plan: No. An extensive table
describes the scope and methodology and the completion date of program
evaluations for fiscal years 2003-2008. However, the plan does not
specifically mention which or how previous evaluations were used in
developing the plan.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-97-208R];
and U.S. Department of Transportation, U.S. Department of
Transportation Draft Strategic Plan for Fiscal Years 2003-2008;
(Washington, D.C. 2003).
[End of table]
Strategic Plan Strengths and Improvements from Fiscal Year 1997 Plan:
The mission statement contained in DOT's 2003-2008 draft strategic plan
is an improvement over the one contained in the 1997 draft plan. DOT's
mission, as stated in the 2003-2008 draft strategic plan, is "to
develop and administer policies and programs that contribute to
providing fast, safe, efficient, and convenient transportation at the
lowest cost consistent with the national objectives of general welfare,
economic growth and stability, the security of the United States, and
the efficient use and conservation of the resources of the United
States." The mission statement covers the major functions and
operations of the department. In our July 1997 report on DOT's 1997
draft strategic plan, we noted that the mission statement could be
improved by including language from the department's enabling
legislation to focus the mission statement more directly on DOT's core
activities. We gave an example of adding the department's purpose to
develop transportation policies and programs that "contribute to
providing fast, safe, efficient, and convenient transportation at the
lowest cost" from DOT's enabling legislation. The mission statement in
the 2003-2008 plan includes such language.
As in its 1997 draft strategic plan, DOT's 2003-2008 draft strategic
plan meets the requirement of GPRA to include long-term goals and
objectives for the major functions and operations of the department.
The 2003-2008 draft strategic plan contains five strategic objectives
(long-term goals) that cover the major functions and activities of the
department and are results oriented. Besides the strategic objectives
of "safety," "mobility," "global connectivity," "environmental
stewardship" and "security," the current draft also contains an
"organizational excellence" objective to "advance the department's
ability to manage for results and achieve the goals of the PMA.":
Each strategic objective section in the 2003-2008 draft plan contains
strategies for attaining DOT's outcomes and objectives. The strategies
for each strategic objective are listed in the categories of
"leadership," "building expertise," and "technology." For example, a
"technology" strategy for the "mobility" strategic objective is to
"examine ways to encourage cargo movements by water through the
development of barge and fast vessel technologies to bring new capacity
to our intermodal transportation system." The plan states that this
strategy supports DOT's outcomes of reduced congestion in all modes of
transportation and increased reliability throughout the system. The
strategies for each strategic objective generally describe the
operational processes, the skills, and the technology required to meet
DOT's goals and objectives. The current draft strategic plan also
states that the resources and programs listed in DOT's annual
performance plans and budgets are necessary to achieve DOT's outcomes
and to execute the strategies. In contrast, the 1997 draft strategic
plan provided insufficient information to describe the operational
processes, the skills, the technology, and the resources required to
meet DOT's long-term goals, as required by GPRA.
Also, each strategic objective section in the current draft plan
contains crosswalks between outcomes in the strategic plan and
performance measures in the annual performance plans and reports. These
crosswalks show the measures that will be used to measure progress in
achieving most of DOT's outcomes and strategic objectives. For example,
the performance measure "number of passengers in international markets
with open skies aviation agreements" is related in a crosswalk to the
outcome "reduced barriers to trade in transportation goods and
services." Together, the measure and outcome will show progress toward
DOT's global connectivity strategic objective. This is an improvement
from DOT's 1997 draft strategic plan when we noted that although
supporting documents showed that DOT had developed information on how
to measure each outcome goal, this information was not included in the
draft.
In contrast to DOT's 1997 draft strategic plan, the 2003-2008 draft
plan lists several external factors for each strategic objective and
generally discusses how these factors could affect the department's
ability to achieve its outcomes and objectives. For example, one of the
external factors for DOT's environmental stewardship strategic
objective is that DOT faces a significant challenge to control and
minimize air, water, and noise pollution. The plan states that if DOT
cannot control and minimize this pollution, the department may
encounter a public backlash that may impede system improvement. For the
external factors relating to the safety and mobility strategic
objectives, the plan lists both positive and negative consequences the
factors could have on achieving goals. One example would be the
possible effects the expansion and integration of the
telecommunications and e-commerce industry sectors could have upon
transportation safety. The plan states that this could affect the
achievement of DOT's safety objective by leading to unsafe practices,
such as the use of cell phones and other personal devices while
driving. On the other hand, these technologies could also contribute to
safety by alerting responders to the location of crashes and vehicles
in distress. The 1997 draft plan identified four external factors and
only discussed how one of those factors could affect DOT's ability to
accomplish its goals.
The current draft strategic plan includes an extensive table listing
program evaluations to be completed during fiscal years 2003-2008. The
table includes the name of the program to be evaluated, which strategic
goal(s) the program supports, the scope and methodology of the
evaluation, and the fiscal year during which the evaluation will be
completed. DOT's 1997 draft strategic plan only listed the titles for
the evaluations scheduled for 1997 and 1998, which was insufficient to
determine the scope and methodology.
DOT's current draft plan lists crosscutting programs by strategic
objective. The discussions of crosscutting programs include the goal of
each program, which of DOT's outcomes each supports, and the agencies
involved. For example, the goal of aviation security is to prevent
explosives, weapons, and other dangerous items from being placed aboard
aircraft. This program supports DOT's security outcome of rapid
recovery of transportation in all modes from intentional harm and
natural disasters. DOT, through the Federal Aviation Administration,
leads this program, which involves the Transportation Security
Administration, Federal Bureau of Investigation, U.S. Customs Service,
and U.S. Postal Service, among others. Previously, the 1997 draft did
not provide evidence that DOT coordinated with other agencies that had
programs and activities that were crosscutting or similar to DOT's.
DOT's major management challenges, which we identified, are generally
discussed in the 2003-2008 draft strategic plan, organized by related
strategic objective. For example, we noted in our January 2003 report
that one major management challenge that DOT faces is building human
capital strategies.[Footnote 88] The current draft strategic plan
includes a discussion of human capital in DOT's organizational
excellence objective. A separate section of this discussion addresses
our concerns regarding human capital strategies and includes several
milestones to address these concerns. These milestones include
conducting workforce planning for mission-critical occupations in
fiscal year 2003, and implementing a departmentwide performance
management system beginning in fiscal year 2003, and a uniform branding
and marketing approach to attract, acquire, and retain diverse high-
quality talent. DOT's 1997 draft strategic plan did not adequately
address the major management challenges we had previously identified.
Critical Strategic Planning Issues Needing Further Improvement:
As stated above, the 2003-2008 plan provides a clear picture of how
success will be measured for most of DOT's strategic objectives and
outcomes. However, this clarity is not provided for a few strategic
objectives and outcomes. We noted the same issue in our review of DOT's
1997 draft strategic plan. For example, in the current plan three of
the outcome goals for the global connectivity strategic objective lack
corresponding performance measures. These outcomes are enhanced
international competitiveness of U.S. transport providers and
manufacturers, harmonized and standardized regulatory and facilitation
requirements, and the most competitive, cost-effective and efficient
environments for passenger travel. The plan states that the measures
are to be determined. However, without these measures it is unclear how
progress will be measured because the outcomes themselves do not lend
themselves to measurement.
While the current strategic plan shows improvement in the schedule for
future program evaluations, it does not sufficiently describe the
evaluations used in establishing or revising DOT's strategic
objectives. DOT states that detailed descriptions of completed program
evaluations are presented in its 2002 performance and accountability
report. Further, the plan states that DOT considered the results of
completed program evaluations, as well as reports from DOT's Inspector
General and GAO, in writing the strategies to achieve its strategic
objectives and outcomes. The plan does not describe which or how
program evaluations were used to write the strategies.
[End of section]
Appendix IV: Observations on Agencies' Annual Performance Plans:
Under GPRA, agencies are to prepare annual performance plans after the
development of their strategic plans. These annual plans are to
establish the connections between the long-term strategic goals
outlined in the strategic plans and the day-to-day activities of
managers and staff. One of our objectives was to assess the changes in
the overall quality of agencies' goals, strategies, and data
articulated in their annual performance plans. To meet this objective,
we judgmentally selected six agencies--Education, DOE, HUD, SBA, SSA,
and DOT--using criteria such as agency size, primary program types, and
previous GAO reviews. To assess the overall quality and improvements
made to the agencies' performance plans, we relied on requirements
contained in GPRA and accompanying committee report language,[Footnote
89] guidance to agencies from the OMB for developing performance
plans,[Footnote 90] best practices identified in our published
work,[Footnote 91] previous GAO evaluations,[Footnote 92] interviews
with agency officials, and our knowledge of agencies' operations and
programs.
Key Elements of Information for Annual Performance Plans:
Although GPRA does not require a specific format for the performance
plan, it does require the plan to (1) identify annual performance goals
and measures for each of an agency's program activities, (2) discuss
the strategies and resources needed to achieve annual performance
goals, and (3) provide an explanation of the procedures the agency will
use to verify and validate its performance data. We categorized each
agency's plan based on the degree to which it collectively addressed
these three characterizations.
To assess the degree to which an agency's plan provides a clear picture
of intended performance across the agency, we examined whether it
included (1) sets of performance goals and measures that address
program results, (2) baseline and trend data for past performance, (3)
performance goals or strategies to resolve mission-critical management
problems, and (4) identification of crosscutting programs (i.e., those
programs that contribute to the same or similar results), common or
complementary performance goals and measures to show how differing
program strategies are mutually reinforcing, and planned coordination
strategies.
To assess the degree to which an agency's plan provides a specific
discussion of strategies and resources the agency will use to achieve
performance goals, we examined whether it included (1) budgetary
resources related to the achievement of performance goals, (2)
strategies and programs linked to specific performance goals and
descriptions of how the strategies and programs will contribute to the
achievement of those goals, (3) a brief description or reference to a
separate document of the human capital, information, and other
resources required to achieve results,[Footnote 93] and (4) strategies
to leverage or mitigate the effects of external factors on the
accomplishment of performance goals.
Finally, to assess the degree to which an agency provides confidence
that its performance information will be credible, we examined how each
report discussed the quality of the data presented. To help improve the
quality of agencies' performance data, Congress included a requirement
in the Reports Consolidation Act of 2000 that agencies assess the
completeness and reliability of their performance data. Under the Act,
agencies were to include this assessment in the transmittal letter with
their fiscal year 2000 performance reports. Agencies were also required
to discuss in their report any material inadequacies in the
completeness and reliability of their performance data and discuss
actions to address these inadequacies.
For each of these elements, we characterized each agency's fiscal year
1999 and fiscal year 2004 plan in one of four ways, based on the degree
to which the plan contained informative practices associated with that
element. Thus, to address the first element concerning the degree to
which the plan provided a clear picture of performance, we
characterized each plan in one of four ways: (1) clear, (2) general,
(3) limited, or (4) unclear. To address the second element, on the
extent to which a plan includes specific discussions of strategies and
resources, we characterized each plan as (1) containing specific
discussions of strategies and resources, (2) general discussions, (3)
limited discussions, or (4) no discussions. Finally, to address the
third element on the extent to which a plan provides confidence that
performance information will be credible, we characterized each plan as
providing (1) full confidence, (2) general confidence, (3) limited
confidence, or (4) no confidence. In conducting our reviews, we
compared our assessments of agencies' fiscal year 2004 plans to our
assessments of plans from fiscal year 1999 using similar
criteria.[Footnote 94] A more detailed discussion of our scope and
methodology and the criteria we used can be found in appendix I.
Table 12 summarizes our characterizations of the six agencies' annual
performance plans based on our current review of fiscal year 2004 plans
and our previously published reviews of 1999 plans. Although the
characterization of agency performance plans did not change
significantly between the 1999 and the 2004 plans, the majority of
agencies' plans showed some improvement.
Table 12: Characterizations of Agencies' Annual Performance Plans:
Agency: Department of Education;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: Limited;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: General;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: Limited;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: General;
Characterizations: Data credible: (no, limited,
general, full): 1999: Limited;
Characterizations: Data credible: (no, limited,
general, full): 2004: General.
Agency: Department of Energy;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: Limited;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: Limited;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: General;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: General;
Characterizations: Data credible: (no, limited,
general, full): 1999: Limited;
Characterizations: Data credible: (no, limited,
general, full): 2004: Limited.
Agency: Department of Housing and Urban Development;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: Limited;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: General;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: Limited;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: General;
Characterizations: Data credible: (no, limited,
general, full): 1999: Limited;
Characterizations: Data credible: (no, limited,
general, full): 2004: General.
Agency: Small Business Administration;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: Limited;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: General;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: Limited;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: General;
Characterizations: Data credible: (no, limited,
general, full): 1999: Limited;
Characterizations: Data credible: (no, limited,
general, full): 2004: General.
Agency: Social Security Administration;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: Limited;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: Clear;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: Limited;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: General;
Characterizations: Data credible: (no, limited,
general, full): 1999: No;
Characterizations: Data credible: (no, limited,
general, full): 2004: General.
Agency: Department of Transportation;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 1999: General;
Characterizations: Picture of intended performance: (unclear, limited,
general, clear): 2004: Clear;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 1999: General;
Characterizations: Strategies and resources: (no discussions, limited,
general, specific): 2004: Specific;
Characterizations: Data credible: (no, limited,
general, full): 1999: Limited;
Characterizations: Data credible: (no, limited,
general, full): 2004: Full.
Sources:
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-172R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-194R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-159R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-180R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/RCED-98-200R];
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS-98-178R];
Department of Education, FY 2004 Annual Performance Plan;
(Washington, D.C. 2003);
Department of Energy, Annual Performance Plan, Fiscal Year 2004;
(Washington, D.C. 2003);
Housing and Urban Development, Annual Performance Plan, Fiscal Year
2004; (Washington, D.C. 2003);
Small Business Administration, Budget Request & Performance Plan: FY
2004 Congressional Submission; (Washington, D.C. 2003);
Social Security Administration, Annual Performance Plan, Fiscal Year
2004; (Washington, D.C. 2003);
Department of Transportation, Fiscal Year 2004 Performance Plan;
(Washington, D.C. 2003).
[End of table]
The remainder of this appendix discusses our observations on how the
quality of each of the agencies' annual performance plans we reviewed
has changed since the agencies submitted their first performance plans
in 1999. We did not independently verify or assess the information we
obtained from agency annual performance plans. If an agency chose not
to discuss its efforts concerning elements in the plan, it does not
necessarily mean that the agency is not implementing those elements.
Observations on Changes in the Quality of Education's Annual
Performance Plan:
Education's fiscal year 2004 annual plan[Footnote 95] provides a
general picture of intended performance across the agency--an
improvement over the 1999 plan--because the measures and indicators
adequately indicate progress toward meeting annual targets; the
measures are objective, measurable, and quantifiable; and baseline and
trend data are included where available. However, the relationship
between the goals and measures in volume 2 and the long-term goals in
volume 1 is not clear and the plan does not make clear whether, and, if
so, how, all of the program activities in the department's budget are
covered by the annual performance goals.[Footnote 96] In another
improvement over the 1999 plan, volume 1 of the 2004 plan provides a
general discussion of Education's strategies and resources to achieve
its goals by presenting strategies and resources and the projected
distribution of fiscal year 2004 funding and staffing for each long-
term goal. The plan also adequately recognizes and discusses external
factors that could affect the department's performance. However, the
resources and many of the strategies are not directly linked to the
achievement of individual annual performance goals and no strategies or
resources are designated for the goals and measures in the program
performance plans in volume 2. Lastly, the 2004 plan provides general
confidence that agency performance information will be credible. The
2004 plan contains information on data sources for most of its
measures, and for some, identifies limitations. The plan also includes
an appendix entitled "Information Quality Guidelines" which recognizes
data quality as a major challenge for the agency and says that its
improvement is a top priority.
Education's Fiscal Year 2004 Performance Plan Provides a General
Picture of Intended Performance:
Education's 2004 annual plan generally defines expected performance.
Measures and indicators are formulated so as to adequately indicate
progress towards meeting annual targets, seem to sufficiently cover key
performance aspects, and adequately capture important program
distinctions. The plan contains about 360 measures between volumes 1
and 2, an improvement over the 860 contained in its 1999 plan, which we
judged to be potentially excessive for an annual performance plan and
possibly interfering with Education's ability to assess its
performance. Unlike in our review of the department's 1999 annual plan,
the measures and indicators in the 2004 plan are objective, measurable,
and quantifiable. For example, most measures and indicators are set up
to measure percentages, cost, counts, or other numerical values with
measurable, quantifiable 2004 targets. In most cases where a measurable
target is not given, the plan provides a reasonable explanation, such
as a new program or the measure being new, or a case where data are not
collected or available each year. In most cases, the plan provides
trend data for measures, which provides a helpful context for assessing
the relevance of the 2004 targets, or an explanation of why such data
were not provided (e.g., the baseline has not yet been established
because the measure and/or program are new).
In our review of Education's 1999 annual plan, we said that greater
outcome measure use would make future annual plans more useful. The
2004 plan frequently includes outcome goals and measures, such as a
measure for the number of states meeting their eighth-grade mathematics
achievement targets under the long-term goal to improve mathematics and
science achievement for all students.
Volume 1 of Education's 2004 annual plan directly aligns strategies,
action steps, measures, and targets with each of Education's long-term
goals and six strategic goals,[Footnote 97] containing the same
strategic goals, long-term goals, and mission as the 2002-2007
strategic plan. In our review of the 1999 plan, we also found that the
plan had performance goals in volume 1 that were directly linked to its
mission, strategic goals, and objectives. However, the relationship
between the goals and measures in volume 2 and the long-term goals in
volume 1 was not made clear in the department's 2004 annual plan, which
was similar to what we found in our review of Education's 1999 plan--
that the department could more directly link the goals and measures in
volume 2 with the strategic objectives (long-term goals). Education's
fiscal year 2002 Performance and Accountability Report includes a table
making it clear that these programs and their goals and measures are
aligned across Education's strategic and long-term goals. By including
such a table in its annual performance plan, Education could clearly
show the link between the goals and measures in volume 2 and the long-
term goals in volume 1.
Although volume 2 of the 2004 annual plan states that it contains
individual program performance plans for all major programs and many
smaller programs, the annual plan does not make clear whether, and, if
so, how all of the program activities in the department's budget are
covered by performance goals. In contrast, we found that the 1999 plan
provided sufficient information to determine which performance goals
and measures in volume 2 covered which program activities in the budget
and whether all were covered. For example, the 1999 plan contained
tables indicating the funding levels for the program activities in the
department's budget and how those activities related to the programs in
volume 2.
In our review of Education's 1999 annual plan, we gave the agency
credit for addressing the need to coordinate with other federal
agencies having related strategic goals or performance goals. However,
we further noted that Education could build on its foundation by
identifying performance goals that reflect activities being undertaken
to support programs of a crosscutting nature and specifying the
activities each agency would undertake and what it expects to achieve
within the fiscal year. While selected action steps in the 2004 plan
refer to instances where the department will coordinate or cooperate
with other federal agencies, the plan does not include steps or goals
for most crosscutting issues identified in Education's strategic plan.
For example, for a crosscutting issue identified in the strategic plan
on safe and drug-free schools and communities, Education said it
partners with the Departments of Justice (Justice) and HHS to promote
drug and alcohol education programs and to disseminate information to
schools and private organizations. The department also coordinates
closely with the Office of National Drug Control Policy, and works
closely with the Office of Juvenile Justice and Delinquency Prevention
Programs to share innovative ideas and promote prevention strategies
and programs. The relevant annual plan sections in both volumes 1 and 2
do not identify goals or action steps related to these interactions.
Additionally, according to our report on Education's June 1997 draft
strategic plan,[Footnote 98] the department has numerous crosscutting
programs and activities, such as those related to early childhood and
employment training, and the 2004 annual plan does not address them
all.
Education's 2004 annual plan discusses applicable goals, measures, and
strategies for two governmentwide major management challenges regarding
strategic human capital management and information security, as well as
three of the four major management challenges we identified for
Education in our January 2001 Performance and Accountability
Series.[Footnote 99] For example, for its student financial assistance
programs, the department has developed performance targets for fiscal
year 2004 related to being removed from our high-risk list, increasing
the default recovery rate, and decreasing grant overpayments to
students. Also, a key strategy under its goal to improve the strategic
management of the department's human capital is to develop a 5-year
human capital plan, including developing a recruitment plan and
relevant training programs. For the fourth of Education's major
management challenges--promoting coordination with other federal
agencies and school districts to help build a solid foundation of
learning for all children--the 2004 plan did not include specific goals
or measures, but it did discuss some related strategies and steps, such
as using partnerships with other federal programs to promote
development of intervention strategies and methods to address the high
incidence of learning disabilities and illiteracy among adolescents
attending high schools.
Education's Fiscal Year 2004 Performance Plan Provides a General
Discussion of Strategies and Resources:
In our review of Education's 1999 annual plan, we found that the plan
had a limited discussion of how the department's strategies and
resources would help achieve its annual performance goals. The 2004
plan includes strategies and resources under each of its long-term
goals to be used to achieve its annual performance goals in volume 1,
including the projected distribution of fiscal year 2004 funding and
staffing, in both dollars and full-time-employees (FTE), for each long-
term goal, under which the annual performance goals are organized.
However, the resources and many of the strategies are not directly
linked to the achievement of individual annual performance goals and no
strategies or resources are designated for the goals and measures in
the program performance plans in volume 2. Overall, the plan does not
discuss how resources were allocated to each goal, a rationale for how
the resources will contribute to improving performance, or the
relationship of capital asset investments, including those for
information technology (IT), to the achievement of specific goals.
However, the plan does include a performance measure and goal for the
cost and schedule of IT investments and a strategy for completing the
department's enterprise architecture, which is to be used to guide IT
capital decisions. In addition, the department has a plan for human
capital management and a new performance appraisal system that is meant
to link employee performance standards to the department's strategic
priorities, but neither had been fully implemented.
In our review of the 1999 plan, we said that external factors that
could affect performance were not discussed and that such factors are
important for a department like Education because much of what it hopes
to achieve depends on others and external events. In its 2004 plan,
Education clearly acknowledges that improving support for its state,
local, and institutional partners, who have the direct ability to
influence outcomes the department seeks, is a major challenge. The plan
contains numerous activities to handle this challenge, including, for
example, to provide support and technical assistance, improve grant
monitoring, and fund an annual survey of states' efforts. Moreover,
although not labeled as addressing external factors, the plan has
specifically related strategies and/or action steps for most external
factors identified in Education's 2002-2007 strategic plan.
Education's Fiscal Year 2004 Performance Plan Provides General
Confidence That Performance Data Will Be Credible:
In our review of Education's 1999 annual plan, we found that it did not
provide sufficient confidence that its performance information would be
credible. For example, the 1999 plan did not sufficiently recognize
limitations in Education's data for its elementary and secondary
education programs. In comparison, Education's 2004 plan recognizes
limitations in Education's data for many of its elementary and
secondary education programs, as well as for other programs. In many of
these cases, the plan also discusses plans to address these
limitations. Also, the plan includes an appendix containing an
abbreviated form of its "Information Quality Guidelines" and a sample
checklist for statistical data from its complete guidelines. The
appendix recognizes data quality as a major challenge to the
department's successful implementation of GPRA and says that the
improvement of data quality is a top priority. The checklist includes
several steps related to the verification and validation of data, such
as evaluating data quality, including known limitations; addressing the
reliability of data sources; and ensuring reproducibility of findings
using the same data and methods of analysis. In addition, the plan
usually identifies the sources of data and, although not in volume 1,
includes a column on sources and data quality for each measure in
volume 2, usually with an item entitled "Validated By" and, in some
cases, "Limitations." In the end, the lack of direct control over the
implementation of its programs, including the collection of data, is a
significant data quality challenge that Education must face.
The 2004 annual plan also contains several action steps on new or
changing information systems that relate to improving the collection of
information for measuring performance. For example, under its strategy
to reduce Education's partners' data reporting burden, the plan
includes an action step to develop and implement the Performance-Based
Data Management Initiative collection system. This step is directly
related to the plan's measure to reduce the burden hours of Education
program data collections per year.
Observations on Changes in the Quality of DOE's Annual Performance
Plan:
Compared to the fiscal year 1999 plan we reviewed, DOE's fiscal year
2004 performance plan continued to provide a limited picture of
intended performance. Although the plan included more results-oriented
annual performance measures, it still provided a limited linkage
between its reported annual goals and its mission, strategic plan
goals, and program activities within its budget request. Furthermore,
the 2004 plan provided a general discussion of strategies and
resources, similar to our 1999 findings. Finally, the 2004 plan
provided a limited level of confidence that data will be credible by
making little progress in reporting on the procedures it uses to ensure
data quality or identifying significant data limitations, which is
consistent with our 1999 findings.
DOE's Fiscal Year 2004 Performance Plan Provides a Limited Picture of
Intended Performance:
While DOE has improved its development of annual performance measures-
-referred to as targets--by making them more results oriented, the
overall picture of performance is limited by the lack of alignment
between its annual and strategic goals and minimal discussion of
coordination with other agencies. Our review of DOE's 1999 plan found
that many measures were unclear, appeared limited in scope, or were not
very useful indicators of performance. We found these problems in the
performance plans for subsequent years as well. For the 2004 plan,
however, the majority of the performance measures related to each goal
were results oriented and pertained specifically to the performance of
fiscal year 2004. An example of one measure requires DOE to train 4,000
federal employees by the end of fiscal year 2004 in energy management
best practices that support National Energy Policy education goals.
DOE provided a limited link between its reported annual goals and its
mission, strategic plan goals, and program activities within its budget
request. While the 2004 annual performance plan goals address all of
the major program activities in DOE's budget, the goals and mission of
the 2004 plan do not align with the mission and goals for the 2003
draft strategic plan. This represents a set back because in our review
of DOE's 1999 annual performance plan, we found that DOE clearly linked
its annual goals to the agency's mission, strategic plan goals, and its
program activities within its budget request. DOE officials told us the
lack of linkage between the performance plan and the strategic plan was
a matter of timing. According to these officials, the department
originally updated its strategic plan at the same time as the annual
performance plan, which was finalized in the early months of 2003, and
the goals of each plan coincided, but the draft strategic plan goals
were revised in the latter part of the year and no longer align with
the 2004 performance plan.
DOE's ability to show coordination with other agencies is also limited.
In 1999, we reported that DOE did not adequately show that it
coordinated with other agencies that have related strategic or
performance goals. DOE's 1999 plan contained very little evidence of
specific goals and measures that addressed crosscutting programs and
only briefly described coordination with other agencies. The 2004 plan
does not specifically describe how coordination is taking place among
crosscutting programs, but does identify groups that it is
collaborating with on certain programs. In response, DOE officials told
us the plan does not discuss what specific collaboration activities are
taking place because it would require reporting too much detail for a
performance plan. DOE officials stated that it collaborates at the
program level, rather than the agency level, because the program plans
pertain to an organizational layer lower than the annual performance
plan.
Finally, the plan briefly mentions that the department has been
identifying challenges and working on ways to address them. According
to DOE officials, when developing annual targets for the department,
management challenges are considered but not mentioned specifically in
the report. Our review of management challenges in 2002 found that DOE
had addressed all eight of its challenges in its 2003 annual
performance plan.[Footnote 100] In comparing these challenges to the
2004 plan, we found that DOE continues to have goals that address the
eight challenges we identified in 2002 and the additional challenges
that we identified in our 2003 performance and accountability
report.[Footnote 101]
DOE's Fiscal Year 2004 Performance Plan Provides a General Discussion
of Strategies and Resources:
DOE provided a general discussion of the strategies and resources that
it will use to achieve its annual performance goals. DOE's 1999 plan
partially provided clear and reasonable strategies for achieving
performance goals, how strategies would contribute to achieving the
performance goals, and key external factors that might affect
performance. For each of the 2004 annual performance goals, DOE
included a "Means and Strategies" section in the plan that described
how each goal will be achieved. For example, one strategy identified to
meet its goal of contributing unique, vital facilities to the
biological and environmental sciences is to conduct peer reviews of the
facilities to assess the scientific output, user satisfaction, and the
overall cost-effectiveness of each facility's operations, and their
ability to deliver the most advanced scientific capability. The 2004
plan includes a brief discussion of the department's overall needs,
particularly in the areas of human capital, financial, and logistical
resources. The plan also identified budget amounts for each of its
goals. DOE's 1999 plan partially identified the resources needed to
accomplish annual performance goals.
The plan also provided a general discussion of the external factors
that could affect achievement of the goals, but it did not specifically
discuss actions on how the external factors will be addressed. For
example, the plan states that external factors related to DOE's goal of
achieving reliable, affordable, and environmentally sound energy
supplies, such as renewable fuels, include program funding, the state
of the economy, the availability of conventional supplies, the cost of
competing technologies, and the continuation of federal tax incentives
and other national-level policies.
DOE's Fiscal Year 2004 Performance Plan Provides Limited Confidence
That Performance Data Will Be Credible:
DOE has made limited progress on reporting the procedures it uses to
ensure data quality. Its 1999 plan described how DOE would ensure that
its performance information is sufficiently complete, accurate, and
consistent, but did not discuss in detail DOE procedures on how to help
ensure the quality of data or the process of collecting the data. The
plan also did not identify significant data limitations and how they
may affect DOE's ability to achieve performance goals. However, the
2004 plan showed some improvement over the 1999 plan by describing
credible procedures to verify and validate performance information and
specific program evaluations are mentioned for each goal. The plan also
discusses that DOE acquired new commercial software for performance
tracking through remote data entry, monitoring, and oversight by
program offices and managers. The 2004 plan only identifies data
limitations and any new or modified systems very briefly for a few
relevant goals. According to DOE officials, with a few exceptions, its
plans do not discuss data limitations because DOE writes goals that are
not affected by data limitations. The goals are written to ensure that
the data will be there to meet performance targets. However, our 2003
performance and accountability series identified several DOE management
challenges where data quality was a concern, such as further upgrades
needed for cyber security to ensure adequate protection of data and
information systems and additional information on the results of
contractors' performance to keep projects on schedule and within
budget.
Observations on Changes in the Quality of HUD's Annual Performance
Plan:
HUD's annual performance plan for fiscal year 2004 improves upon areas
where we previously reported shortcomings and generally meets the
criteria set forth in GPRA. HUD's 2004 plan provides a general picture
of intended performance by covering all the programs contained in HUD's
budget and linking program activities to strategic goals and
objectives. The plan also improved by providing specific information on
HUD's strategies and activities along with performance measures it will
use to assess progress toward its goals and discussing relevant
external factors that could affect the attainment of certain program
objectives. HUD also provides greater confidence that performance data
will be credible by thoroughly discussing the data it will use for
measuring progress toward its goals. Nevertheless, the plan could be
further enhanced if it included more specific information on how funds
will be allocated to achieve program objectives, explain how HUD will
contribute to crosscutting efforts along with other agencies, and what
steps it will take to mitigate the impact of external factors on its
programmatic objectives.
HUD's Fiscal Year 2004 Performance Plan Provides a General Picture of
Intended Performance:
Since our review of HUD's annual performance plan for fiscal year
1999,[Footnote 102] HUD has made progress in developing an annual
performance plan that generally reflects the department's mission and
provides a general picture of intended performance. HUD's most recent
performance plan covers the program activities contained in its budget,
and generally links program activities to strategic goals and
objectives, key items missing from its plan for fiscal year 1999. HUD
has also improved the quality of its performance plan by including
performance measures that generally indicate how the department will
gauge progress toward achieving its goals. For example, the performance
plan lists a series of performance measures for each objective that can
be used to indicate progress towards the department's goals and
expected performance. These measures are also objective and a number of
them have been quantified, another key area where HUD has improved
since its first performance plan. For example, activities supporting
HUD's long-term strategic objective to "Improve the Physical Quality
and Management Accountability of Public and Assisted Housing" include,
among other things, eliminating 100,000 units of the worst public
housing. According to the current plan, the department intends to
demolish 10,000 of these units in fiscal year 2004.
While HUD's most recent annual performance plan generally identifies
other agencies it will coordinate with to address crosscutting efforts,
it does not discuss how it plans to work with these agencies to address
these crosscutting activities. For example, the plan states that the
Interagency Working Group on Limited English Proficiency will ensure
that persons with limited English proficiency will have meaningful
access to funded and federally conducted programs and activities.
However, the plan does not discuss what HUD's contribution to this
multiagency effort will be, what strategies it will employ, or how it
will measure progress toward achieving the strategies of this
multiagency effort.
HUD's current performance plan is an improvement compared to its fiscal
year 1999 plan as it describes steps HUD will take to address major
management challenges. One of HUD's strategic goals, "Embrace High
Standards of Ethics, Management, and Accountability," identifies five
objectives that cover management challenges, some of which have been
raised by GAO and the HUD IG. These objectives discuss plans to rebuild
HUD's human capital and diversify its workforce; improve HUD's
management, internal controls, and resolve audit issues; improve
accountability, service delivery, and customer service; ensure program
compliance; and improve internal communication and employee
involvement.
HUD's Fiscal Year 2004 Performance Plan Provides a General Discussion
of Strategies and Resources:
HUD's most recent performance plan also improves upon earlier plans we
reviewed in providing readers an idea of the strategies that HUD will
employ to carry out its goals. Each strategic goal in the annual
performance plan contains a section titled "Means and Strategies,"
which describes activities HUD will pursue to support that goal. For
example, to support its goal of "Increasing Homeownership
Opportunities," HUD will fund low-income homeowner assistance programs
to provide approximately 40,000 families with down payments and closing
costs on their homes, 473,199 families with home purchase and
homeownership counseling, and about 232,370 families with rental
counseling.
HUD's performance plan also discusses the strategies it will employ to
address the department's human capital issues, such as the upcoming
potential wave of employees planning to retire and the need to equip
staff with the desired knowledge and skills. For example, HUD completed
a staff resource estimation and allocation system in 2002, and it will
conduct a comprehensive workforce analysis in 2004 to serve as the
basis to fill mission-critical skill gaps through succession planning,
hiring, and training initiatives in its Five-Year Human Capital
Management Strategy.
Although HUD has made progress in linking its resources to strategies,
it could improve the discussion by linking funding allocations to
specific performance goals, thus making the plan more informative. The
plan discusses budget and staff allocations for programs supporting
each strategic goal. For instance, portions of HUD's Community
Development Block Grants Fund and Home Investment Partnership Program,
with a combined budget authority for fiscal year 2004 of more than $2.5
billion and staff of 203, support the strategic goal of promoting
"Decent Affordable Housing." However, it is unclear what resources will
be used to pursue specific performance targets for each program.
Additionally, HUD does not mention in its plan how IT and capital
resources will be used to support its programs.
Anticipating that some aspects of the department's strategic goals are
intertwined with broader phenomena, the performance plan also discusses
several external factors relevant to each strategic goal that could
affect HUD's ability to meet its objectives. For example, for its
strategic goal "Promote Decent and Affordable Housing," HUD states that
broad economic factors can affect opportunities for low-income workers
relying on the department for rent assistance to make progress towards
self-sufficiency. However, it is unclear from the performance plan what
actions, if any, HUD has put in place to mitigate the effect of these
external factors.
HUD's Fiscal Year 2004 Performance Plan Provides General Confidence
That Performance Data Will Be Credible:
HUD has also made significant progress in providing assurance that the
department will be able to use credible data to gauge progress towards
achieving its goals. HUD identifies the steps it (or others) will take
to verify and validate the performance data to ensure that what is
reported on HUD's performance will be credible. For example, for its
objective "Increasing Minority Homeownership," HUD will rely on, among
other indicators, the rate of minority homeownership from the Current
Population Survey conducted monthly by the U.S. Census Bureau. HUD will
not verify the data because the Bureau performs that task.
Additionally, HUD also includes in its most recent performance plan a
discussion of the inherent limitations of the data it will use and
generally discusses steps it will take to improve the measure,
providing the reader with a clearer expectation of what HUD will be
able to report.
Observations on Changes in the Quality of SBA's Annual Performance
Plan:
SBA's 2004 performance plan shows progress made over the agency's 1999
performance plan. In contrast to our review of SBA's 1999
plan,[Footnote 103] the 2004 plan provides a general picture of
intended performance by discussing coordination between SBA and other
federal agencies on crosscutting activities. Resource analysis sections
throughout the plan provide a general discussion of how SBA has
previously used its resources to achieve its goals and how it intends
to use future resources for the same purposes. The 2004 plan also
provides general confidence that SBA's performance data will be
credible by including more detail on how SBA verifies and validates its
data, as well as by identifying data limitations. However, several
areas of the plan could be improved, such as clearly linking SBA's
performance indicators, performance goals, and programs.
SBA's Fiscal Year 2004 Performance Plan Provides a General Picture of
Intended Performance:
SBA's fiscal year 2004 performance plan provides a general picture of
intended performance. The performance goals and performance indicators
in the plan are generally objective, measurable, and quantified.
Performance indicators are listed throughout the plan by related
programs and strategic programmatic goals. Performance goals and
outcome goals are listed for each strategic programmatic goal. In our
review of SBA's fiscal year 1999 performance plan, we noted that SBA's
performance goals were objective and measurable, its performance
measures were generally objective and quantified, and that the
performance goals in the plan were clearly linked to SBA's strategic
goals and objectives.
Like the performance measures contained in its 1999 plan, the 2004
plan's performance indicators will be useful in assessing progress
towards SBA's performance goals. For example, the performance indicator
"Regulatory Cost Savings to Small Business" will adequately show
progress for the corresponding performance goal "Cost savings for small
business due to the efforts of the Office of Advocacy." In this
example, the performance indicator, which is listed under the Advocacy
Program, can be linked to a performance goal because of a crosswalk
that relates outcome goals, performance goals, and programs. However,
there is not always such a clear link between all of SBA's performance
indicators and performance goals because indicators are listed by
program instead of by performance goal. The BusinessLaw.gov program is
linked to three performance goals: "number of users of
BusinessLaw.gov," "reduced cost to businesses and regulatory agencies,"
and "increased rate of compliance." While the first two performance
goals appear related to the first two indicators listed in the
BusinessLaw.gov program section, there is no clear relationship between
any of the other performance indicators for this program and the third
performance goal, "increased rate of compliance.":
SBA's 2004 performance plan contains annual performance goals that
generally cover the agency's budget activities. The 2004 performance
plan contains a budget crosswalk that "shows how the goals relate to
specific and general program areas." This is an improvement over the
1999 plan, which we noted contained a budget crosswalk, but the
categories in it did not match SBA's budget accounts or activities by
name or account number. However, the performance plan does not seem to
cover all of SBA's programs. Three "advocacy" programs listed in the
crosswalk do not seem to be contained in the plan: Business.gov,
Disability Initiative, and National Women's Business Council.
Each strategic programmatic goal section in the 2004 plan contains a
discussion of crosscutting issues. Several examples of coordination
efforts are given, such as SBA working with the Department of Defense
to integrate the PRO-Net system with the Central Contractor Registry
and SBA partnering with the Federal Acquisition Institute to develop
on-line training courses for small business programs. In contrast,
SBA's 1999 performance plan provided little information on SBA's
coordination efforts with other entities whose programs and activities
crosscut those of SBA.
SBA's 2004 performance plan generally addresses performance and
accountability challenges we have previously identified. For example,
we have previously stated that SBA needs to strengthen its performance
in human capital management. The 2004 plan includes outcome goals,
performance goals, and programs to address SBA's strategic management
of human capital in a section on the PMA.
SBA's Fiscal Year 2004 Performance Plan Provides a General Discussion
of Strategies and Resources:
The 2004 performance plan provides a general discussion of the
strategies and resources SBA will use to achieve its goals. Each
strategic programmatic goal and each of the goals for the PMA contains
a discussion of the strategies for accomplishing the goals. These
discussions provide a broad overview of the strategies used at the
strategic programmatic goal level. For example, the plan includes a
strategy for SBA's strategic management of human capital. The strategy
lays out SBA's Transformation and Human Capital plans, which will be
used to implement a new vision of SBA. In its 1999 performance plan,
SBA discussed strategies for most of its performance goals, although
for some of the goals the strategies were missing.
Each strategic programmatic goal and several of the PMA goals contain
brief discussions of external factors that could affect the achievement
of SBA's goals. These discussions include actions to address external
factors, such as working with an Interagency Acquisition Working Group
under the Procurement Executives Council to develop supplemental
performance measures to better evaluate the success of its programs. In
1998, we noted that SBA's 1999 plan recognized certain external factors
and contained a discussion of actions SBA could take to mitigate the
effects of such factors for one of its strategic goals. We stated that
it would be useful for SBA to include a similar discussion of external
factors and mitigation strategies for its other strategic goals.
Each program listed throughout the plan has a resource analysis section
that describes how resources were used in fiscal year 2002. Some of
these analyses also include planned resources for fiscal year 2004. For
example, the resource analysis section for the Small Business
Development Centers (SBDC) program states that for fiscal year 2004 SBA
requested approximately the same level of funding as in fiscal year
2002. In 2002, 85 percent of the funds were for grants, while the other
15 percent covered field support, program management, and overhead
costs such as rent, legal services, human resources, and information
technology support. Some of the resource analyses also contained pie
charts of the breakdown of costs. This is an improvement over SBA's
1999 performance plan, which we found did not specifically identify the
human or technological resources that SBA would need to achieve its
performance goals.
SBA's Fiscal Year 2004 Performance Plan Provides General Confidence
That Performance Data Will Be Credible:
SBA's 2004 performance plan provides general confidence that its
performance data will be credible. An appendix of the performance plan
contains verification and validation information, as well as data
limitations and remedies for these limitations for most of SBA's
performance indicators. However, the appendix does not include this
information for the performance indicators of the disaster loan
program, nor are any of the PMA performance indicators discussed in the
appendix.
Generally, the discussions of SBA's verification and validation
processes for its indicators in the 2004 plan are one-or two-sentence
statements. For one of the indicators, "number of jobs created and
retained by the 7(a) loan program," SBA states that it does not have
access to the data for verification purposes. SBA also notes that it
does not independently verify some of the external data it gathers, as
is stated in the verification discussion of the indicator, "504 loans
to emerging market firms." This is an improvement over SBA's 1999
performance plan, which included brief descriptions, often only one or
two words, on the means it used to verify and validate its data. We
noted in our report on the 1999 plan that these appeared to be sources
of data for the measures rather than means to verify and validate the
data.
The data limitations contained in SBA's 2004 performance plan are
generally one-sentence statements and the same limitations are used for
multiple indicators. For example, several limitations, such as "the
measure is based on the number of approved loans" or "information is
derived from loan approval data," are used for multiple indicators. The
appendix also lists remedies for the data limitations for each
indicator. For example, a limitation of SBA's indicator "small business
appointments conducted with procurement officials" is that the
indicator may not capture unscheduled appointments. The remedy for this
limitation is to keep track of both scheduled and unscheduled
appointments. The discussion of data limitations and their remedies in
the 2004 plan shows progress over SBA's 1999 plan, which did not
contain a discussion of data limitations.
Observations on Changes in the Quality of SSA's Annual Performance
Plan:
Compared to the fiscal year 1999 plan we reviewed, SSA's performance
plan provided a clear picture of intended performance by (1) defining
expected performance, (2) offering trend data, which helps track
progress toward performance goals, and (3) using objective, measurable,
and quantifiable performance measures. SSA also provided general
information on its strategies and resources, somewhat better than our
1999 findings. Finally, the plan provided a general level of confidence
that data will be credible by describing the Inspector General's (IG)
involvement in data testing, providing data sources and definitions,
and identifying some data weaknesses, an improvement over the 1999
plan. However, SSA's performance plan still does not fully discuss the
agency's coordination with other agencies, identify performance goals
that clearly cover all the program activities, address how SSA plans to
use the information from the evaluations to improve program results,
identify the resources needed to address each performance goal, and
discuss data verification and validation procedures for its internal
systems.
SSA's Fiscal Year 2004 Performance Plan Provides a Clear Picture of
Intended Performance:
Overall, SSA's fiscal year 2004 performance plan has improved over its
1999 plan.[Footnote 104] The 1999 plan only provided a partial picture
of SSA's intended performance across the agency. In June 1998, we
reported that SSA's 1999 Annual Performance Plan contained performance
goals, many of which were measurable and linked to the agency's
strategic goals[Footnote 105] and objectives; some of the performance
goals related to particular strategic goals were objective, measurable,
and quantifiable. However, other goals were not measurable or
quantifiable and did not define the level of performance to be
achieved, thus making it difficult to see how SSA would assess success.
SSA's fiscal year 2004 plan provides a much clearer picture of intended
performance through (1) defining expected performance, (2) offering
trend data, which helps track progress toward performance goals, and
(3) the use of objective, measurable, and quantifiable performance
measures. For example, as part of the strategic objective to "Prevent
fraudulent and erroneous payments and improve debt management," SSA
provided historical data on the outcome measure "Percent of SSI
payments free of preventable error (overpayments and underpayments)"
from fiscal years 1999-2001 and projected goals for fiscal years 2002-
2004.
While we found significant improvements in SSA's 2004 annual
performance plan over its 1999 plan, we also found some weaknesses. For
example, coordination efforts with other entities, such as federal
agencies, state and local entities, and others, are not well
identified. According to SSA officials, SSA coordinates with other
federal agencies, such as the IRS and the Immigration and
Naturalization Service, as well as Veterans Administration, on
information-sharing initiatives. However, these types of coordination
efforts are mentioned only briefly, if at all, in the 2004 annual
performance plan.
In its 2004 plan, SSA includes a list of major program evaluations it
plans to conduct during 2003-2004, with a brief description of the
evaluations, their associated strategic goals, and projected completion
dates. However, there is no indication how SSA plans to use the
information from the evaluations to improve program results. The plan
could be enhanced if the descriptions of these evaluations included the
manner in which SSA planned to use the information gathered to improve
its programs.
SSA's Fiscal Year 2004 Performance Plan Provides a General Discussion
of Strategies and Resources:
SSA's 1999 plan had little discussion of the relationship between SSA's
mission, goals, and budget activities. Throughout the document, the
fiscal year 2004 plan included clearer discussions of the linkage
between SSA's mission and goals. It also provided performance data
dating back to fiscal year 1999, essential to making comparisons
between prior and proposed levels of performance. The 2004 performance
plan noted that the Limitation on Administrative Expenses account,
SSA's basic administrative account, is an annual appropriation that
covers everything from salaries and benefits of SSA federal employees
(excluding IG) to systems and telecommunications activities. SSA
provided information on the funding sources of this account, including
some of its budget accounts.
In its fiscal year 2004 plan, SSA provided information on the
strategies it plans to use in addressing its key strategic objectives.
The plan included a summary chart, showing the strategic objectives
associated with each strategic goal, as well as the performance
measures under each objective. In addition, the "means and strategies"
section associated with each strategic objective identified strategies
that support items in the PMA, GAO and IG major management challenges,
and Social Security Advisory Board recommendations. In our October 2002
report Performance and Accountability: Reported Agency Actions and
Plans to Address 2001 Management Challenges and Program Risks,[Footnote
106] we noted that SSA identified directly related goals and measures
for five of its six challenges, and had strategies (without goals or
measures) for the sixth challenge.
It is difficult to determine whether or not the annual performance plan
identifies annual performance goals that cover all of the program
activities in the agency's budget, as well as the financial, human
capital, and information technology resources needed to address each
individual goal. General human capital requirements and goals are
identified as part of SSA's strategic goal to strategically manage and
align staff to support SSA's mission. The plan is neither structured by
program activity nor account. SSA noted that it aligned its strategic
goals, performance measures, and budget with its major functional
responsibilities rather than by program accounts since direct service
and support employees provide services linked to these functional
responsibilities, as opposed to a specific program. However, SSA does
not indicate what it means by "functional responsibilities," nor does
it show a clear link between its strategic goals and such
responsibilities.
As in the fiscal year 1999 plan, the fiscal year 2004 plan included a
discussion of external factors that could affect the achievement of its
goals.[Footnote 107] SSA identified strategies to alleviate some, but
not all of, the factors. For example, SSA plans to mitigate the loss of
institutional knowledge through SSA's "retirement wave" through the use
of employee development programs, redeploying positions to direct
service, hiring Presidential Management Interns, and the increased use
of hiring flexibilities. However, the discussion of factors affecting
SSA's solvency strategic goal merely notes that Social Security
programs must respond to related developments.
SSA's Fiscal Year 2004 Performance Plan Provides General Confidence
That Performance Data Will Be Credible:
SSA's 1999 plan stated that the Office of the Inspector General was
responsible for reviewing the data systems underlying its performance
measures, but did not provide further details that would assure the
reader that SSA is taking the steps necessary to ensure data integrity.
In contrast, SSA's fiscal year 2004 plan provided data sources and
definitions for each performance measure. SSA's fiscal year 2004 plan
identified data limitations related to performance measures, as well as
some efforts to correct or address data weaknesses. When performance
indicators and goals are not quantified, SSA describes its benchmarks
for goal achievement. For example, for the outcome measure "Developing
new performance management systems," SSA defines "Implementing the new
SES system" as its goal for 2003.
As in the fiscal year 1999 plan, SSA notes that the IG's office is
involved in the data system reliability process. In the fiscal year
2004 plan, SSA went further to explain the IG's four-point approach to
reviewing performance measures, including assessing whether the
reported performance measure data are valid. SSA also noted that
performance data for its quantifiable measures are generated by
automated management information and workload measurement systems, as a
by-product of routine operations. However, there is no discussion of
verification and validation procedures for data generated by these
systems.
Observations on Changes in the Quality of DOT's Annual Performance
Plan:
DOT's annual performance plan for fiscal year 2004 showed evidence of
improvements in areas that we previously identified had shortcomings in
our 1998 review of DOT's 1999 performance plan.[Footnote 108] The 2004
plan provides a clear picture of intended performance with DOT's
measures and performance goals now being clearly linked to the
department's strategic objectives. A specific discussion of DOT's
strategies and resources in the plan includes numerous and detailed
strategies for achieving DOT's performance goals, and the resources
needed for those strategies. Procedures to verify and validate data, as
well as known data limitations, are described for each performance
measure providing full confidence in the credibility of DOT's
performance data. Still, the performance plan could be improved by
including a discussion of, and performance measures for, each of DOT's
program activities and by more consistently describing DOT's role in
crosscutting programs.
DOT's Fiscal Year 2004 Performance Plan Provides a Clear Picture of
Intended Performance:
DOT's fiscal year 2004 performance plan shows evidence of many of the
same strengths as, and a few improvements over, its fiscal year 1999
performance plan and provides a clear picture of intended performance.
The 2004 plan lists outcome goals, performance goals, and measures by
strategic objective, all of which are generally objective,
quantifiable, and can show progress toward DOT's strategic objectives.
For example, the measure "fatalities per 100 million vehicle-miles of
travel" will gauge progress toward the performance goal "reduce highway
fatalities per 100 million vehicle-miles traveled to no more than 1.0
in 2008, from 1.7 in 1996." The data gathered by the measure will also
show progress toward the related outcome of "reduce the number of
transportation-related deaths" for DOT's "safety" strategic objective.
This is an improvement over DOT's 1999 plan in which we found that
DOT's performance goals typically covered only a portion of the
strategic goals and the link between annual performance goals and
strategic goals could be improved.
DOT's plan also presents trend and baseline data for each performance
measure and goal. For the example given above, the plan indicates the
performance targets from 1999 to 2004 and also presents actual data for
these targets for 1999 to 2002. In its 1999 plan, DOT had provided
baseline data for most of its performance goals and measures as well.
This information, along with the clearly linked performance goals and
strategic objectives, helps to show DOT's progress in achieving its
goals.
As in the 1999 plan, DOT's 2004 performance plan generally covers each
program activity in its budget request for fiscal year 2004. An
appendix to the performance plan lists DOT's program activities and
indicates the proposed funding level for each program by strategic
objective. However, as in its 1999 plan, a few programs do not seem to
be linked to the strategic objectives elsewhere in the plan. Capital
grants to the National Passenger Rail Corporation (Amtrak) and the
Bureau of Transportation Statistics' Office of Airline Information are
both linked to the "mobility & economic growth" strategic objective in
the budget crosswalk, but they do not appear in the discussions
contained within that strategic objective section. When the 2004 plan
was published in February 2003, DOT had not yet released its new reform
strategy for Amtrak, which was made public in July 2003. Still, the
inclusion of information on Amtrak, as well as a discussion of the
Bureau of Transportation Statistics' Office of Airline Information,
would provide a clearer picture of how DOT intends to achieve its
goals.
The discussions of each performance goal have sections entitled "Other
Federal Programs with Common Outcomes." In this section, the plan
describes crosscutting programs and other agencies with which DOT
works. For example, the plan states that the Research and Special
Programs Administration of DOT continues to develop the National
Pipeline Mapping System with the Federal Energy Regulatory Commission,
the National Oceanic and Atmospheric Administration, the Department of
Energy, the U.S. Geological Survey and others, in order to help analyze
risks to environmentally sensitive and populated areas. This supports
DOT's efforts to reduce pipeline incidents. Yet for several goals,
coordination efforts are not described. One example of this is in the
highway congestion section where the plan states that the Federal
Highway Administration works closely with the Department of the
Interior, Department of Agriculture, and Department of Defense agencies
to improve mobility on federally owned lands. However, the plan does
not describe the specific actions that are being taken to improve
mobility. Our 1998 report stated that DOT's contribution or role was
not described in many of the crosscutting programs listed in DOT's 1999
performance plan.
The 2004 performance plan generally addresses performance and
accountability challenges we previously identified. The discussions of
these management challenges are included in the plan by the performance
goal and programs to which they are related. For example, in discussing
highway safety, DOT addresses our concerns on transportation safety,
specifically through the use of safety belts. The strategies include
continuing the National Highway Traffic Safety Administration's safety
belt outreach to high-risk populations and encouraging states to
embrace "Click It or Ticket" as the message or theme for their Buckle
Up Campaigns. A performance measure related to this management
challenge included in the plan is "percentage of front seat occupants
using safety belts." However, not all of the management challenges have
related measures and goals. For example, we have previously identified
building human capital strategies as a management challenge for DOT. A
section within the plan focuses on an "organizational excellence"
objective to implement the PMA. Strategic management of human capital
strategies is discussed in this section but no goals or measures are
given to show DOT's progress with these strategies. Still, this shows
some improvement over the 1999 plan, which generally covered management
challenges, but did so in a separate appendix without explaining how
the challenges were related to the rest of the plan. We noted that this
area could be improved by including goals and measures related to
resolving these challenges.
DOT's Fiscal Year 2004 Performance Plan Provides a Specific Discussion
of Strategies and Resources:
DOT's 2004 performance plan shows several improvements over its 1999
performance plan, providing a specific discussion of strategies and
resources. Discussions of each performance goal include a section
titled "Strategies and Initiatives to Achieve 2004 Target." These
sections include a variety of means by which DOT intends to accomplish
its performance goals. One example would be for DOT's performance goal
to "reduce pipeline hazmat (hazardous materials) spilled 30 percent by
2006, from the last five years' average spill rate." The strategies for
this goal include enforcing operator qualification requirements,
expanding monitoring technology that can help prevent construction-
related damage to pipelines, and developing regulatory standards for
leak detection technology. This shows progress from when we reported
that DOT's 1999 performance plan lacked sufficient information to
clearly link the strategies to performance goals in many cases.
In contrast to its 1999 performance plan, DOT's 2004 performance plan
generally discusses the human, capital, information, and other
resources needed to meet its performance goals. Each performance goal
section of the performance plan includes a graph showing the enacted
funding for fiscal year 2002, and the proposed funding for fiscal years
2003 and 2004. The organizational excellence objective for DOT
describes DOT's human capital and information technology resources and
strategies. For example, one of DOT's strategies for strategic
management of human capital is to "establish a corporate approach to
target recruitment efforts, with special emphasis on cross-modal,
mission-critical occupations," which includes a pilot program for
centrally recruiting and training entry-level employees for one or more
mission-critical occupations.
The 2004 plan also discusses external factors that could hamper DOT's
ability to achieve its performance goals. In our review of DOT's 1999
performance plan, we noted that the plan could be improved by
recognizing more external factors and by discussing actions that DOT
could take to mitigate the effects of these factors. In contrast,
external factors are listed for most of the performance goals in the
2004 plan. For its transportation accessibility goals, DOT states that
as the population ages, more people will require accessible public
transit, for which states and local agencies decide how best to
allocate federally provided resources. One of the strategies that
addresses this external factor is the Special Needs of Elderly
Individuals and Individuals with Disabilities grants, which DOT states
will help meet the transportation needs of the elderly and persons with
disabilities when regular transportation services are unavailable,
insufficient, or inappropriate to meet their needs.
DOT's Fiscal Year 2004 Performance Plan Provides Full Confidence That
Performance Data Will Be Credible:
The 2004 plan provides full confidence that DOT's performance data will
be credible. As in the 1999 performance plan, the 2004 performance plan
contains a section, entitled "Performance Data and Performance
Measurement," that discusses the means that DOT uses to verify and
validate its data. But unlike the 1999 plan in which this discussion
was broad and not linked to specific goals and measures, the 2004 plan
also contains an appendix that provides the following for each of DOT's
measures: the source of the data, limitations of the data, observations
on the quality of the data, work planned or ongoing to improve data
quality, and any known biases. Finally, DOT has compiled source and
accuracy statements,[Footnote 109] which provide more detail on the
methods used to collect the data, sources of variation and bias in the
data, and methods used to verify and validate the data.
The presentation of data limitations in DOT's 2004 performance plan
also shows progress from its 1999 plan. The Performance Data and
Performance Measurement section includes a general discussion of DOT's
data limitations. This discussion includes limitations for the internal
and external data used by the department. Specific limitations for
internal data can be found in the aforementioned source and accuracy
compendium, while details on the limitations of external data are given
in the appendix on performance measures. In our report on DOT's 1999
performance plan, we stated that information on data limitations was
lacking for most measures and that the plan could be improved by more
consistently addressing the data limitations throughout the plan.
[End of section]
Appendix V: Observations on Agencies' Annual Performance and
Accountability Reports:
To help Congress and the President determine agencies' actual
performance and progress in achieving strategic plan goals, GPRA
requires each agency to prepare a report on program performance for the
previous fiscal year.[Footnote 110] One of our objectives was to assess
the overall quality of agencies' annual performance and accountability
reports and the extent to which selected elements of agency reporting
have improved. To meet this objective, we judgmentally selected six
agencies--Education, DOE, HUD, SBA, SSA, and DOT--using criteria, such
as agency size, primary program type, and previous GAO reviews. To
assess the overall quality and improvements made to the agencies'
performance and accountability reports, we relied on requirements and
guidance contained in GPRA and accompanying committee report
language,[Footnote 111] guidance to agencies from OMB for developing
performance reports,[Footnote 112] interviews with agency officials,
the Chief Financial Officers Act,[Footnote 113] our previous
reports,[Footnote 114] and our knowledge of agencies' operations and
programs. To assess the quality of the six agencies' performance and
accountability reports, we categorized each report based on the degree
to which it addressed three characterizations: (1) picture of
performance, (2) link between resources and results, and (3)
credibility of performance information.
To assess the degree to which an agency's report provided a clear
picture of performance across the agency, we reviewed the extent to
which the report addressed elements required by GPRA. The annual
performance report should:
* describe the performance indicators established in the agency's
annual performance plan, along with the actual program performance
achieved compared with the performance goals expressed in the plan for
that fiscal year;
* review the success of achieving the performance goals of the fiscal
year;
* provide actual results for the 3 preceding fiscal years;
* evaluate the performance plan for the current fiscal year relative to
the performance achieved toward the performance goals in the fiscal
year covered by the report;
* explain and describe where a performance goal has not been met or a
corresponding level of achievement if an alternative form is used, as
well as why the goal was not met, plans and schedules for achieving the
established performance goal, and if the performance goal is
impractical or infeasible;
* describe the use and assess the effectiveness of achieving
performance goals of any waivers; and:
* include the summary findings for those program evaluations completed
during the fiscal year covered by the report.[Footnote 115]
We also looked at the extent to which the reports clearly discussed
progress achieved in addressing the major management challenges
previously identified by us or others. For agencies that choose to
issue a performance and accountability report, the Reports
Consolidation Act of 2000 requires that the report include a summary of
the most serious management and performance challenges facing the
agency, as identified by their IG, and a brief assessment of the
agency's progress in addressing those challenges.
In assessing the clarity of the performance information, we also looked
at selected qualitative characteristics used by the Association of
Government Accountants, in conjunction with the Chief Financial
Officers Council, in assessing performance and accountability reports
for the Certificate of Excellence in Accountability Reporting.[Footnote
116] These characteristics included (1) whether there was a clear
relationship between the performance information in the report and the
goals and objectives contained in the strategic and annual performance
plans, (2) the extent to which the agency limited the measures it
discussed to those that were most significant for its programs, and (3)
the extent to which the report was user friendly by being well-
organized, concise, readable, and making effective use of graphics to
ease understanding of narrative information. We characterized the
clarity of each report in one of four ways: (1) clear, (2) general, (3)
limited, or (4) unclear, based on the extent to which the 2002 report
addressed the elements required by GPRA and the other informative
practices we described.
Both GPRA and the CFO Act emphasized the importance of linking program
performance information with financial information as a key feature of
sound management and an important element in presenting to the public a
useful and informative perspective on federal spending. Similarly, the
current administration's ambitious agenda for performance budgeting,
calling for agencies to better align budgets with performance goals and
focus on capturing full budgetary costs and matching these costs with
output and outcome goals, suggests that agencies need to develop
integrated financial and performance management systems that will
enable the reporting of the actual costs associated with performance
results. Although linking resources to performance goals is not a
requirement of GPRA, the committee report for GPRA suggested that
developing the capacity to relate the level of program activity with
program costs, such as costs per unit of result, costs per unit of
service, or costs per unit of output, should be a high priority. We
have reported that world-class financial management practices call for
enterprisewide systems to integrate financial and operating data to
support both management decision making and external reporting
requirements. To assess the degree to which an agency's report
discussed the relationship between resources and results, we
characterized each report as having a (1) clear relationship, (2)
general relationship, (3) limited relationship, or (4) no relationship.
Finally, to assess the degree to which an agency's plan provided
confidence that the agency's performance information would be credible,
we examined how each report discussed the quality of the data
presented. To help improve the quality of agencies' performance data,
Congress included a requirement in the Reports Consolidation Act of
2000 that agencies assess the completeness and reliability of their
performance data. Under the act, agencies were to begin including this
assessment in the transmittal letter with their fiscal year 2000
performance reports. Agencies were also required to discuss in their
report any material inadequacies in the completeness and reliability of
their performance data and discuss actions to address these
inadequacies.
We have previously reported on other practices that enhance the
credibility of performance data that are not specifically required by
GPRA.[Footnote 117] For instance, discussions of standards and methods
used by agencies to assess the quality of their performance data in
their performance reports provides decision makers greater insight into
the quality and value of the performance data. We also reported on
additional practices, in several agencies' performance reports, that
would help foster transparency to the public and assist decision makers
in understanding the quality of an agency's data. The additional
practices we observed included (1) discussions of data quality,
including known data limitations and actions to address the limitations
and (2) discussions of data verification and validation procedures. To
address the extent to which a report provided confidence that
performance information was credible, we characterized each report as
providing (1) full confidence, (2) general confidence, (3) limited
confidence, or (4) no confidence.
In conducting our reviews, to the extent information was available in
prior assessments, we compared our findings of agencies' fiscal year
2002 reports to our assessments of reports for fiscal year
1999.[Footnote 118] A more detailed discussion of our scope and
methodology and the criteria we used can be found in appendix I. Table
13 shows the results of our assessment of the six agencies' reports.
Table 13: Characterizations of Agencies' Fiscal Year 2002 Annual
Performance and Accountability Reports:
Department/agency: Department of Education;
Characterizations: Picture of performance: (unclear, limited, general,
clear): Limited;
Characterizations: Resources linked to results: (no, limited, general,
clear): Clear;
Characterizations: Data credible: (no, limited, general, full):
General.
Department/agency: Department of Energy;
Characterizations: Picture of performance: (unclear, limited, general,
clear): General;
Characterizations: Resources linked to results: (no, limited, general,
clear): Limited;
Characterizations: Data credible: (no, limited, general, full):
Limited.
Department/agency: Department of Housing and Urban Development;
Characterizations: Picture of performance: (unclear, limited, general,
clear): General;
Characterizations: Resources linked to results: (no, limited, general,
clear): No;
Characterizations: Data credible: (no, limited, general, full):
General.
Department/agency: Small Business Administration;
Characterizations: Picture of performance: (unclear, limited, general,
clear): Limited;
Characterizations: Resources linked to results: (no, limited, general,
clear): General;
Characterizations: Data credible: (no, limited, general, full):
General.
Department/agency: Social Security Administration;
Characterizations: Picture of performance: (unclear, limited, general,
clear): General;
Characterizations: Resources linked to results: (no, limited, general,
clear): Limited;
Characterizations: Data credible: (no, limited, general, full):
General.
Department/agency: Department of Transportation;
Characterizations: Picture of performance: (unclear, limited, general,
clear): General;
Characterizations: Resources linked to results: (no, limited, general,
clear): No;
Characterizations: Data credible: (no, limited, general, full): Full.
Sources:
U.S. Department of Education, U.S. Department of Education FY 2002
Performance and Accountability Report; (Washington, D.C. 2003);
U.S. Department of Energy, Performance and Accountability Report,
Fiscal Year 2002; (Washington, D.C. 2003);
U.S. Department of Housing and Urban Development, Fiscal Year 2002
Performance and Accountability Report; (Washington, D.C. 2003);
Small Business Administration, Fiscal Year 2002 Performance and
Accountability Report; (Washington, D.C. 2003);
Social Security Administration, Performance and Accountability Report,
Fiscal Year 2002; (Washington, D.C. 2002);
U.S. Department of Transportation, Fiscal Year 2002 Performance and
Accountability Report; (Washington, D.C. 2003).
[End of table]
The remainder of this appendix discusses our observations on the
quality of the agencies' annual performance and accountability reports
we reviewed and, to the extent information was available from our prior
reviews, how the quality has changed since the agencies submitted their
first reports on fiscal year 1999 performance. We did not independently
verify or assess the information we obtained from agency annual
performance reports. If an agency chose not to discuss its efforts
concerning elements in the report, it does not necessarily mean that
the agency is not implementing those elements.
Observations on the Quality of Education's Fiscal Year 2002 Performance
and Accountability Report:
Education's fiscal year 2002 Performance and Accountability Report
comprises two volumes--the main volume and a second volume including
performance reports for the agency's individual programs. In our
assessment, we did not review the second volume, which includes very
detailed, discrete, and disaggregated performance information with over
350 individual measures for the Office of Civil Rights, IG, and 117
Education programs in 60 clusters.
Although Education's report included many features designed to present
its performance information clearly, the overall clarity was limited by
the significant amount of performance information that was unavailable
to show Education's performance results. In contrast, Education's
report very clearly related its performance to its costs by using both
graphics and text to provide the agency's estimate of appropriations
associated with achieving each of its six strategic goals, 24
objectives (long-term goals), and individual programs. Finally,
Education provided a general level of confidence in the quality of its
data, primarily because of its recognition of the challenges it faces
on the timeliness, reliability, and validity of its data. Education's
recent efforts in undertaking a performance-based data management
initiative in partnership with state leaders to allow timely and ready
access to high-quality achievement and other performance data, which
the IG said would address many of the related concerns identified
during IG audits, also aided in the level of confidence in the data.
Education's Fiscal Year 2002 Report Provided a Limited Picture of
Performance:
Education's 2002 performance report is directly aligned with the goals
and measures in the agency's 2002-2007 strategic plan and its 2002-2003
annual plan. Of the 210 measures included in the agency's strategic
plan and annual plan, 120 were identified for measurement in fiscal
year 2002, and all of these are addressed in the performance report.
The report contains sections on changes planned to enhance performance
on the basis of results. For each measure, the performance report
includes trend data, with a table showing actual data from fiscal years
1999 through 2002; in some cases, the table also includes data from
fiscal year 1998. When data are not provided, the table indicates that
they were not applicable or not available. Overall, the report contains
clear, succinct figures and a table summarizing the status of all 120
measures, as summarized in figure 19.
Figure 19: Summary of Education's Performance Indicators for Fiscal
Year 2002:
[See PDF for image]
[End of figure]
However, while Education's 2002 report does review the levels of
success for its performance goals[Footnote 119] for fiscal year 2002,
there is a critical limitation. As we observed in our review of
Education's 1999 report,[Footnote 120] data were not yet available for
many measures in the 2002 report. Specifically, 2002 data were
available for only 41 of the 120 measures; the rest were characterized
as "Pending: Data Not Yet Available" (63) or "Incomplete: Data Not
Expected" (16). Despite the numerous strengths in Education's 2002
performance report, the picture of performance for Education presented
in this report is limited mainly because of the lack of data for so
many of its 2002 targets. However, Education recognizes the challenges
created by its limited access to timely, reliable data:
We still face significant challenges to meeting our national education
goals. Primary among these challenges is access to timely, reliable
data on our performance in meeting our goals and implementing our
programs. Our efforts to identify effective and ineffective programs in
the Department are severely limited by the frequent absence of useful
data about them. In FY 2002 we designed a performance-based data
management initiative which will provide much more robust information
about our programs and strategic objectives, as well as provide a
strong foundation for educational research and evaluation.
This data management initiative is being undertaken in partnership with
state leaders and the software industry and is expected to result in an
electronic data system that will allow ready access to high-quality
achievement and other performance data in a timely and seamless manner
in the future. Because the lack of data for so many of its targets
blurs the picture of performance, Education should take every possible
step to complete, as quickly as possible, its newly established
performance-based data management initiative.
While lacking data for so many of its measures, Education's 2002 report
provides an explanation for measures with pending or incomplete data.
For pending data, the report states that comparisons to targets will be
made in the subsequent performance and accountability report, in
addition to citing the department's performance-based data management
initiative. The report further indicates that measures with incomplete
data were so characterized because methods to collect data were not
ready in time to measure fiscal year 2002 results, data collection did
not occur, or data collection was delayed. The report goes on to say
that, for these measures, Education will put methods in place to
measure fiscal year 2003 results, develop other data sources, or revise
its measures to measure results differently. In addition, for each
incomplete measure, the report clearly describes why data are
incomplete and what will be done to address the situation. For example,
for its measure on the percentage of states with complete school
accountability systems in place, as required by the No Child Left
Behind Act,[Footnote 121] the report explains that the requirements
under this act are more extensive than in the past, that states that
had met prior requirements may not yet meet new requirements, that the
department had decided regulation would be necessary, and that
regulations had not been finalized to define a complete school
accountability system.
In addition, the report almost always included explanations of
performance and sometimes provided information on why targets were not
met when that was the case. However, such information was not always
easy to find, as the report did not always include it in the same area
as the related measure.
For most measures, including those that did not meet their targets, the
report provided information on the steps Education is taking or plans
to take to improve or enhance performance. For example, the 2002 target
for a measure on the number of children attending charter schools was
690,000 and the actual result was 575,000. To improve performance on
this measure, Education's report says that it is distributing guidance
and information to encourage parents to consider charter schools, using
both publications and its Web site to promote charter school
enrollment, and sponsoring a charter schools Web site with information
on federal assistance for charter schools. However, it was not always
clear how the steps cited would improve or enhance performance. For
example, for four measures on advanced placement (AP) achievement that
were not met in 2002, the strategy given for improving performance is
to continue to support increasing AP achievement through the Advanced
Placement Incentives program, but the report does not include an
explanation of, or any information on, this incentives program.
Education's report included an appendix entitled "Findings from FY 2002
Evaluations" and included summaries of the findings from nine GAO
reports and eight other studies completed in fiscal year 2002. For
example, the Education for Homeless Children and Youth Program:
Learning to Succeed report comprised two studies that found that
homeless students are best served when promising practices are
implemented as part of a comprehensive and coordinated education
program for the homeless.
The 2002 performance and accountability report also contained an
appendix consisting of the IG's summary of serious management
challenges, including financial management, federal student aid
programs, information technology, program performance and
accountability, and human capital. For each of these challenges, the IG
provided information on Education's progress in addressing them. Under
program performance and accountability, for example, the IG pointed out
that a significant amount of the data used to measure education
programs were provided by state and local education entities and that
it is imperative that these data are accurate, so as to provide
Congress, OMB, and the public with an objective measure of the success
of education programs. The IG said that Education has recognized the
importance of improving data quality and addressed this issue in its
performance plan.
The IG's summary of serious management challenges also included
references to GAO's high-risk list with respect to federal student aid
programs and human capital management, and Education's report included
measures related to both of these areas. Specifically, for its 2002
measure to have Federal Student Aid (FSA) leave the GAO high-risk list
by 2003 and not return, the report states that the department "almost
met" its 2002 target by achieving 94 percent of its FSA High Risk Plan,
and it described the shortfall as not significant or material. In
contrast, in our review of Education's 1999 Performance
Report,[Footnote 122] we noted that the department did not have goals,
objectives, or measures related to problems with its student assistance
programs. In addition, for the six measures in Education's 2002 report
under its strategic goal to improve the strategic management of its
human capital, the department reports that four targets were pending,
one was incomplete, and one had set the baseline. With respect to the
GPRA requirement[Footnote 123] that agencies' performance reports
include a review of the performance goals and evaluation of the
performance plan relative to the department's strategic human capital
management, Education's report discusses its human capital management
strategic goal and related performance goals in the context of its
human capital management plan, One-Ed.
Education's 2002 Report Showed a Clear Relationship between Resources
and Results:
Education's 2002 report included information for each of its six
strategic goals and 24 objectives that clearly linked the department's
resources with its efforts to achieve specific results. While the
department was not able to break the costs down by each of its measures
and targets, the report used both graphics and text to provide the
department's estimate of appropriations associated with achieving each
of its six strategic goals, 24 objectives (long-term goals), and
individual programs. For example, for each of its objectives, the
report used a pie chart to show the percentage of the department's
appropriation that supports the objective, the rest of the strategic
goal the objective falls under, and the other five strategic goals. An
example is shown in figure 20 for the objective to ensure that all
students read on grade level by the third grade.
Figure 20: Inputs: Allocating Funds for Education's Objective to Ensure
That All Students Read on Grade Level by the Third Grade:
[See PDF for image]
[End of figure]
The text accompanying each chart listed the dollar amount supporting
the objective's activities, the percentage of the strategic goal's
allocation that amount represented, the individual programs that
supported the objective, and the dollar amount from salaries and
expenses that was included in the dollar amount for the objective. In
addition, in the report's appendixes, a table summarizing fiscal year
2002 appropriations and staffing allocated by goal and objective also
included the FTEs under staffing for each strategic goal and objective.
Another table provided a percentage breakdown of each objective's
appropriations by 146 agency programs.
Education's Fiscal Year 2002 Report Provided General Confidence That
Performance Data Were Credible:
Education's 2002 performance report provided general confidence that
the agency's data were credible because of its recognition of the
challenges it faces on the timeliness, reliability, and validity of its
data; its straightforward disclosure of these challenges; and its
recent efforts to address them. In Education's transmittal letter, the
department secretary said that the information contained in the report
is "as complete and reliable as we have available." However,
Education's report recognized that one of the agency's significant
challenges to meeting its national education goals is access to timely,
reliable data on performance and that the lack of useful data severely
limits efforts to identify effective and ineffective programs. The
report further explained that 97 percent of the department's funding is
awarded to third parties, including, for example, state and local
agencies, that have an impact on the measurement of results, especially
the timing of data collection. Thus, Education recognized in its report
that it had limited control over the data it must use to report
results. Similarly, in the IG's summary of serious management
challenges, the report noted that Education needs to improve its
controls over the timeliness, reliability, and validity of data.
Moreover, Education's report included information on recent steps it
has taken to address its data challenges. In addition to a discussion
of its data management initiative to develop an electronic data system
providing access to timely, high-quality data, the report included an
appendix with the agency's Information Quality Guidelines and draft
Data Quality Standards presented in an abbreviated format. The
discussion of data quality standards recognizes the importance of data
quality concepts to the process of developing high-quality performance
measures. The eight standards provided are: validity, accurate
definitions, accurate counts, editing, calculation, timeliness,
reporting, and burden reduction. To facilitate the use of the
standards, Education reported that it created a data quality checklist
and regularly held classes to teach staff how to apply the standards.
The IG's summary of serious management challenges gave the department
credit for this effort, pointing out that these guidelines address many
of the concerns identified during IG audits and that the department
plans to disseminate these guidelines to the chief state school
officers.
The report also provided explanations of data sources and data quality
for most measures. For example, for the measures on the percentages of
12thgrade students scoring at or above the basic and proficient levels
on the National Assessment for Educational Progress (NAEP) reading
test, the source is given as: "U.S. Department of Education; National
Center for Education Statistics, (NAEP); The Nation's Report Card,
Reading." Under data quality, the report states that NAEP data are
validated using rigorous National Center for Education Statistics
statistical standards. For most of the measures, the explanations of
data quality contain similar information on data validation. However,
the data quality information only sometimes identifies what limitations
are relevant, if any. For example, for a measure on the percentage of
managers satisfied with services received from Education's Office of
Management when hiring staff, the department relied on an internal
survey of managers for its data. Although the response rate for this
survey was 22 percent, the report did not say whether this was a
limitation to the data collected. For a few of Education's measures,
the report stated that no data limitations had been noted. It would be
better if the agency clearly stated whether the data for each measure
had limitations or not, and, if so, what they were.
Observations on the Quality of DOE's Fiscal Year 2002 Annual
Performance and Accountability Report:
DOE's 2002 Annual Performance and Accountability Report provided a
general picture of intended performance by explaining in detail the
progress made in meeting performance measures and addressing management
challenges. It also provided a limited discussion of the costs incurred
to achieve DOE's performance goals by organizing its report by major
program activities, the costs of these activities, and their
corresponding goals. Finally, the report provided a limited level of
confidence that data will be credible because the report did not
include a discussion on data limitations.
DOE's Fiscal Year 2002 Report Provided a General Picture of
Performance:
DOE's 2002 Performance and Accountability Report provided a general
picture of performance in meeting its goals and measures. The report
contained a detailed explanation of progress for each of its
performance measures, which were referred to as "targets," by
identifying whether each measure was met, not met, or had mixed
results, as shown in figure 21. In addition, the report identified the
operational processes, technology, human capital, and other resources
used to achieve each performance measure. The results for the past 3
years of performance measures related to each goal were also reported
so that performance trends could be identified. However, the report did
not clearly explain how the results of the performance measures
reported contributed to achieving the performance goals in DOE's annual
performance plan and strategic plan.
Figure 21: Summary of DOE's Performance Indicators for Fiscal Year
2002:
[See PDF for image]
[End of figure]
Each performance measure that was not met or had mixed results
contained a plan of action to achieve the measure in the future. In
addition, the majority of fiscal year 2002 measures that were not met
contained a clear explanation as to why they were not met or had mixed
results. For example, a measure that required Southeastern Power
Administration to meet its planned repayment of principal of federal
investment was not met due to severe drought. DOE's report explained
that to achieve the measure in the future, Southeastern plans to change
its rate design, propose rate increases to obtain greater revenue, and
increase cost recovery from fixed charges.
DOE also discussed the progress it made in addressing performance and
accountability challenges. The report identifies significant issues for
fiscal year 2002 that, as stated in the report, merit a higher level of
attention and focus in the department. Each of the challenges were
linked to a goal and its related performance measure(s). In addition,
actions taken to address each challenge and the progress made on those
actions were identified. For example, one of the challenges identified
was the need for DOE to meet federal requirements for improved and more
cost-effective use of information technology. A related goal to deal
with this challenge was for DOE to promote the effective management of
information technology resources in the department. To address this
challenge, the report stated that DOE realigned its management
structure for information technology issues, established an
enterprisewide license for Microsoft software, and launched an e-
government applications task force to identify high-priority e-
government investments, among other actions. In our prior review of
DOE's 2001 performance report, we also found that DOE had made progress
in addressing all eight of its major management challenges.[Footnote
124]
It is unclear, however, how program evaluations were used to assess
performance because DOE did not include a summary of program evaluation
findings in either the fiscal year 1999 or 2002 reports. According to
DOE officials, a section on program evaluations was not included in
fiscal year 2002 and one is not planned for fiscal year 2003 in order
to limit the amount of detail included in the report.
DOE's Fiscal Year 2002 Report Showed a Limited Relationship between
Resources and Results:
DOE's 2002 performance report provided a limited discussion of how its
resources were related to its performance. As in our review of the 1999
report, DOE's 2002 report, which also includes DOE's financial
information, continued its practice linking the department's
performance information to the costs of its program activities. For the
majority of its program activities, the 2002 report included the
program activity's net costs and related performance information for
fiscal years 1999 through 2002. During our review of DOE's 1999
performance report, we were supportive of DOE's efforts to link its
performance goals and measures to the program activities in the
President's Budget. However, DOE has not moved beyond presenting its
performance and cost information by program activity, instead of by
strategic or annual performance goal or objective. A report that
presented cost and performance information by performance goals in
addition to other presentations would more clearly identify the costs
associated with the achievement of each goal. According to DOE
officials, the department plans to link its individual performance
measures to the costs of program activities in future reports.
DOE's Fiscal Year 2002 Report Provides Limited Confidence That
Performance Data Are Credible:
DOE's reporting on data credibility has improved but is still limited.
Based on our review of DOE's 1999 performance report, a key improvement
made to the 2002 performance report was the department's ability to
report on its data validation and verification processes. DOE's 1999
report did not discuss the implementation of DOE's verification and
validation plan or provide any evidence that the data quality was
sufficient for assessing the department's performance. The 2002 report
met some requirements of the Reports Consolidation Act by including a
statement in the report's transmittal letter assessing the completeness
and reliability of the data. The letter did not discuss any material
inadequacies with DOE's performance data. The report also included a
high-level discussion on how DOE will validate and verify its data and
refers the reader to its 2003 annual performance plan for further
details. For example, the report stated that DOE's end-of-year
reporting process includes certification by heads of organizational
elements on the accuracy of reported results, followed by a review for
quality and completeness by DOE's Office of Program Analysis and
Evaluation.
Although the department has improved on reporting its data verification
and validation processes, it has not improved on reporting any existing
data limitations. Neither the 1999 nor the 2002 report included an
overall discussion of the limitations to the data or steps DOE would
take to address those limitations, although the 2002 performance report
did identify minor data limitations for a few specific goals. DOE
officials stated that a discussion on data limitations was not included
in the 2002 report because the department already reports on this
information in the annual performance plan and they thought it was
redundant to put it in the report.
Observations on the Quality of HUD's Fiscal Year 2002 Annual
Performance and Accountability Report:
Compared to its fiscal year 1999 performance report, HUD's fiscal year
2002 Performance and Accountability Report provides a general picture
of what the department accomplished by including, among other things, a
report card listing its performance indicators with the corresponding
achievement, a list of program evaluations concluded during the fiscal
year, trend information for some of its performance indicators, a
discussion of the department's attempts to address its performance and
accountability challenges, and visual aids to illustrate information on
its performance. In a few instances, the report mentions whether
certain performance goals are impractical or unfeasible. However, the
report is not as clear as it could be because it does not (1) explain
how it plans to address performance targets that were not met during
the fiscal year; (2) include an evaluation of the fiscal year 2003
performance plan relative to the performance information presented in
the performance report for the fiscal year; or (3) include an
evaluation of the fiscal year 2003 performance plan relative to the
performance information presented in the performance report for fiscal
year 2002. The report does not show the relationship between resources
and results by linking expended dollar amounts to specific program
objectives. The report provides general confidence to the reader that
the data presented are credible by providing background information on
each performance indicator and discussing the results and analysis of
the most recent data.
HUD's 2002 Report Provided a General Picture of Performance:
Overall, HUD's fiscal year 2002 Performance and Accountability Report
provides a general understanding of what the department's mission is
and what it accomplished during the previous fiscal year.[Footnote 125]
Since we first reviewed its report for fiscal year 1999, HUD has made
progress in developing its performance report to comply with
GPRA.[Footnote 126] In reviewing HUD's fiscal year 1999 performance
report, we noted that it only contained performance information for
three of the department's four outcome measures. HUD's report for
fiscal year 2002 includes a report card for each strategic goal listing
performance targets that were met. The report card also provides an
explanation for many performance targets that were not marked as being
met during the fiscal year. Although the report includes visual aids to
enhance readers' understanding of progress made toward attaining
performance targets, HUD could enhance the report by providing a
summary of performance targets for all strategic objectives met (or
not) during the fiscal year.
HUD's performance report does not meet GPRA's requirement of including
an evaluation of its fiscal year 2003 performance plan relative to the
performance attained by the department in fiscal year 2002. Including
this evaluation could provide readers some assurance that HUD takes
into account prior performance information, such as unmet goals, to
manage its performance in the fiscal year already under way.
The report suggests that some of the performance targets that were not
met during the fiscal year were impractical or unfeasible. For example,
for its goal of "Increase the Rate of Homeownership," HUD mentions that
the indicator can be resistant to increases above an undetermined level
because homeownership is not practical or desirable for all households.
Broad economic conditions, including employment, incomes and interest
rates can affect homeownership rates. Likewise, HUD will no longer
track a performance indicator that measures the percentage of low-
income housing units containing threats to health and safety, such as
exposed wiring, unvented heaters, holes in the floor, and rodents. HUD
mentions that this indicator is not included in the fiscal year 2003
annual performance plan because of the difficulty of attributing the
results to its programs.
In several instances, HUD's annual performance report lacks a
discussion of how it plans to address unmet performance targets as
required by GPRA. For example, while HUD substantially met almost half
of its performance targets in fiscal year 2002, the report does not
mention what steps the department will take to address some of its
unmet performance targets (see fig. 22).
Figure 22: Summary of HUD's Performance Indicators for Fiscal Year
2002:
[See PDF for image]
[A] Rather than stating if some performance targets were met or not,
HUD provided the following explanations: data not available; no
performance goal for this fiscal year; third quarter of calendar year
(last quarter of fiscal year, not entire fiscal year); calendar year
ending in the current fiscal year; calendar year ending the previous
fiscal year; other reporting period; results too complex to summarize;
and baseline newly established.
[End of figure]
HUD continued to build upon the strengths of its earlier report by
including trend information for some performance indicators it used to
measure progress toward its targets during the past fiscal year. While
not presented consistently throughout the report, trend information
provides a context to understand HUD's performance and helps to show
the extent to which HUD exceeded or fell short of expectations set for
its performance targets. For instance, for HUD's performance indicator
that tracks increases in the share of welfare families residing in
public housing that move from welfare to work each year, the report
mentions that in fiscal year 2002 the rate was 13.1 percent compared to
19.9 percent in fiscal year 2001. In preparing to implement the
Temporary Assistance for Needy Families program, HUD originally
estimated this indicator to be around 6.5 percent in fiscal year 1997.
HUD's fiscal year 2002 report also mentions that it concluded several
program evaluations during the fiscal year, a key requirement absent
from its performance report for fiscal year 1999. The report also
provides a brief summary of the main findings of these program
evaluations. In fiscal year 2002, HUD concluded and published reports
on 21 program evaluations covering five of its strategic goals.
Similar to our findings on HUD's previous performance reports,[Footnote
127] HUD's fiscal year 2002 report discusses the steps the department
took to address decade-long management challenges. For example, while
HUD's report mentions that deficiencies remain in its financial
management systems, in fiscal year 2002 the department initiated a
project to design and implement an integrated financial system.
Similarly, to address staffing imbalances and human capital challenges,
HUD implemented the last phase of its Resource Estimation and
Allocation Process in January 2002 and started to implement the Total
Estimation and Allocation Mechanism, a tool that collects actual
workload accomplishments and staff usage within the various operating
components at HUD.
HUD's 2002 Report Showed No Relationship between Resources and Results:
While HUD has made some improvements in how it presents cost
information in its report, it is still not useful for linking program
objectives to specific dollar expenditures. HUD's report provides a
summary of the cost of operations by each reporting segment, such as
the total amount of money spent by the Federal Housing Authority and
the Public and Indian Housing programs, and states that the total cost
for fiscal year 2002 operations was $33 billion. However, the report
does not reconcile these costs to specific program performance
objectives, limiting the reader's ability to understand how HUD used
its resources to carry out its objectives during the fiscal year.
HUD's 2002 Report Provided General Confidence That Performance Data Are
Credible:
HUD's fiscal year 2002 performance report generally informs the reader
on critical issues about the reliability of its performance data, an
issue that was not discussed in detail in its earlier report. In its
transmittal letter, HUD briefly discusses that in some instances the
data used in the report were either incomplete and/or unreliable. The
report includes background information, results, analysis, and a
discussion of the data used for each performance indicator during the
fiscal year. In discussing the data, in some instances HUD points out
issues concerning data validity and accuracy and mentions steps HUD
will take to correct problems. For example, to address problems with
its indicator on the number of homeowners who have been assisted with
the HOME program, HUD has established a team of managers, technical
staff, and contractors to make a series of improvements to the
Integrated Disbursement and Information System beginning in fiscal year
2003, which should reduce the need for data cleanup.
Observations on the Quality of SBA's Fiscal Year 2002 Annual
Performance and Accountability Report:
SBA's fiscal year 2002 annual performance report shows several
improvements over the agency's initial report. The report shows a
general relationship between resources and results by including an
analysis of resources used by each program in fiscal year 2002. A
section on data validation and verification, which includes data
limitations and remedies for those limitations, provides a general
level of confidence in the credibility of SBA's performance data. While
the 2002 report includes a scorecard to show the agency's overall
performance, the report provides a limited picture of performance due
to a lack of plans to meet unmet goals in the future and data that were
unavailable to show progress towards a large share of SBA's performance
goals, among other reasons.
SBA's 2002 Report Provided a Limited Picture of Performance:
SBA's 2002 performance report[Footnote 128] provides a limited picture
of its performance. SBA's 2002 report includes a scorecard that shows
overall agency performance for its 2002 goals, including trend data
(when available) from fiscal year 1999 to fiscal year 2002, the fiscal
year 2002 goal, and a column showing the percentage of the fiscal year
2002 goal achieved. However, based on the performance information
provided in the fiscal year 2002 performance report, it can be
difficult to gauge SBA's progress in achieving its goals. This is
similar to our findings on SBA's 1999 report, which we noted was
unclear as to how well SBA performed in achieving several of its
performance goals for two of the key outcomes addressed in our 2000
report.[Footnote 129] Figure 23 summarizes SBA's progress on its 19
performance goals for fiscal year 2002.
Figure 23: Summary of SBA's Performance Goals for Fiscal Year 2002:
[See PDF for image]
[End of figure]
Data were unavailable for 10 of SBA's 19 performance goals in 2002. For
the nine goals that had performance data available, SBA met seven.
SBA's 2002 performance report included explanations for all of its
goals that were unmet, deemed infeasible, or for which data were not
available. For example, the report states that "homes restored to pre-
disaster conditions" and "businesses restored to pre-disaster
conditions" are no longer goals because SBA is reviewing its outcome
measures for the disaster loan program. Also, data were not available
for the "customer satisfaction" goal in the disaster assistance program
because a Customer Service Survey for disaster loan recipients was not
issued during the fiscal year due to having not received final
clearance from OMB. This contrasts to our findings on SBA's 1999 report
when the agency did not provide explanations for not meeting several of
its performance goals. However, for the two goals that were unmet in
2002, "start-ups receiving 7(a) and 504 financing" and "jobs created
and retained by SBIC clients," the report does not describe any plans
for achieving the goals in the future. The lack of information for over
half of SBA's performance goals and the absence of plans for achieving
unmet goals limits our ability to assess the overall progress the
agency made in fiscal year 2002, as well as the likelihood that it will
improve its performance in the future.
Several other factors limited our ability to evaluate SBA's performance
in its fiscal year 2002 report. The report presents performance data in
accordance with the goal structure of SBA's 2003-2008 draft strategic
plan. The goal structure contained in the fiscal year 2002 performance
report does not directly correspond with the goal structure presented
in the 2002 performance plan. Only one of the report's strategic goals,
"Help Families and Businesses Recover from Disasters," directly
corresponds to the 2002 performance plan. Similarly, not all of the
performance goals listed in performance scorecards in both documents
correspond. For example, the performance goal "start-ups receiving 7(a)
and 504 loans viable 3 years after receiving loan," listed in the
scorecard in the 2002 report, is not listed in the 2002 performance
plan. The report states that based on 2002 results SBA made
"substantial modifications" to its fiscal year 2003 goals, but the
report does not specifically discuss how the performance achieved in
2002 could affect the achievement of the 2003 goals. Finally, the
report does not include the findings of program evaluations completed
in fiscal year 2002. SBA states that it was unable to conduct program
evaluations in 2002 due to a lack of funding and that the agency has
requested funding for program evaluations in fiscal years 2003 and
2004.
Similar to our findings on SBA's 1999 performance report, the agency
continues to provide information indicating the agency's progress in
addressing management challenges that have been previously identified.
For example, we have previously observed that SBA needs to improve the
quality of the performance measures that it uses for the disaster loan
program. SBA states in its 2002 report that the agency is in the
process of reevaluating its measures for the disaster loan program, and
specifically that the methodology for measuring the number or
percentage of homes and businesses restored through the program will be
addressed by this review.
SBA's 2002 performance report also discusses the agency's strategic
management of human capital. One section of the report relating to the
PMA describes SBA's transformation plan, which is to realign the
agency's "organization, operation, and workforce to better serve its
small business customers." An appendix to the report identifies fully
developing and implementing the agency's human capital management
strategy as one of SBA's major challenges. This section lists the
actions that SBA needs to take to address this challenge as well as the
progress the agency has made in implementing these actions. However,
the report does not include a review of the performance goals and
evaluation of the performance plan relative to the agency's strategic
human capital management, as required by the Homeland Security Act of
2002.
The report also contains two broad overviews and an appendix of GAO
audits and recommendations, as well as a description of management
challenges identified by the agency's Inspector General. One chart,
entitled "Status of GAO Reviews Conducted at SBA in FY 2002," shows the
review title, status of the review (open or closed), and the number of
recommendations that came from these reviews. Another chart, entitled
"Number of Open GAO Recommendations at End of FY 2002," lists the GAO
report number and title, the year it was issued, and the number of
recommendations remaining open. Further detail is provided in an
appendix to the performance report, which lists GAO's outstanding
recommendations, the status of the recommendations, and the estimated
date of completion. Another appendix includes a report from SBA's
Acting IG that describes the most serious management challenges SBA
faced in fiscal year 2002.
SBA's 2002 Report Showed a General Relationship between Resources and
Results:
SBA's 2002 performance report contains analyses of resources and
results for SBA's programs that show a general relationship between
resources and results. In the description of each program's
performance, the report includes an analysis of the resources used by
each program. For example, the fiscal year 2002 cost of the Advocacy
Program was estimated to be $8 million, with 50 percent of the funds
going to support the Office of Advocacy, 14 percent funding economic
research, 11 percent for SBA's executive direction support, 16 percent
for fixed costs, and 9 percent going to human resources, information
technology, and procurement. The report contains crosswalks that show
the relationship between SBA's strategic goals, outcome goals,
performance goals, and programs. The resources used by each program can
then be linked through this crosswalk to performance goals to generally
show the resources needed for the results achieved towards the goals.
However, the connection of resources to results could be more
explicitly stated in the report if results and resources were also
presented by performance goal.
SBA's 2002 Report Provided General Confidence That Performance Data Are
Credible:
SBA's 2002 performance report provides general confidence that the
agency's performance data are credible. In the letter transmitting its
2002 performance report, SBA states that the performance data for its
credit and procurement assistance programs are complete and reliable,
based on a systematic review of these data. SBA further states that it
is "working to improve the completeness and reliability of the
performance data for the advice provided to small business through
SBA's resource partners." Data for this aspect of SBA's performance are
collected through surveys, which the agency notes are neither
consistent nor comparable, and from which client responses are
difficult to obtain. This could be seen as a material inadequacy, of
which a discussion is required by the Reports Consolidation Act. SBA
discusses the actions it will take to address the quality of the
surveys by stating in the transmittal letter that it is working to
improve the survey instruments it uses to obtain performance data.
SBA provides a detailed discussion of each performance indicator in a
section of the report on data validation and verification. For each
indicator, the report provides a definition, source, information on
validation, and means for data verification. The verification process
for several measures includes audits, independent reviews, and
consistency checks. However, for these measures this is the only
discussion of verification procedures and no further details are
provided. Also, for several measures, such as "number of start-up firms
financed by 7(a) & 504" and "regulatory cost savings to small
businesses," SBA states that it does not independently verify the data.
The report also addresses limitations to its data in the section on
data validation and verification. In this section SBA states that it
faces many challenges in acquiring high-quality data on both outputs
and outcomes, from both internal and external sources. The strategies
that SBA will use to address the shortcomings of its data quality are
contained in this section as well, which include ensuring the validity
of performance measures and data, fostering organizational commitment
and capacity for data quality, assessing the quality of existing data,
responding to data limitations, and building quality into the
development of performance data.
The section on data validation and verification in the 2002 report
discusses how SBA plans to remedy the limitations for each indicator.
For example, for the "customer satisfaction rate" measure of the
disaster loan program, the report states that the surveys used for this
measure only determine the satisfaction of those who received disaster
loans and therefore do not address the satisfaction of those who did
not receive the loans. The remedy listed for this limitation is to
expand the survey to include all applicants. This is an improvement
from SBA's 1999 report, which we noted did not discuss data limitations
that could affect the quality of data used by SBA to assess its
performance.
Observations on the Quality of SSA's Fiscal Year 2002 Annual
Performance and Accountability Report:
On the whole, SSA's 2002 performance report generally showed the
agency's progress towards its annual goals for fiscal year
2002.[Footnote 130] It showed continued emphasis on outcome-oriented
goals and identified relevant results that were linked to individual
strategic objectives. It also provided trend information, typically
back to fiscal year 1999, and contained a brief discussion of the
program evaluations completed during fiscal year 2002. SSA's strategic
goals were linked to financial resources at a very high level, but none
of the performance goals were associated with costs; thus, the cost of
achieving (or not achieving) a particular goal was not clear.
Additionally, the SSA Commissioner certified that SSA's data
presentation was credible, but missing data and a lack of documentation
of the methods and data used to measure its performance reduced the
overall quality of the document.
SSA's 2002 Report Provided a General Picture of Performance:
SSA's 2002 performance report exhibited a general description of
performance, including the identification of 14 key performance
indicators out of a total of 69 indicators. Similar to our review of
SSA's fiscal year 1999 report,[Footnote 131] we found that many of
SSA's fiscal year 2002 goals and indicators were outcome oriented. In
the fiscal year 2002 report, SSA plainly summarized progress on its 69
performance indicators, as shown in figure 24.
Figure 24: Summary of SSA's Performance Goals for Fiscal Year 2002:
[See PDF for image]
[End of figure]
In SSA's 1999 annual performance report, performance measures focused
on activities rather than results, so it was difficult to determine the
agency's real progress in achieving results. For example, the measures
directly related to the outcome "long-term disability benefits are
reduced because people return to the workplace" did not sufficiently
track progress toward this key outcome. One of the measures was to
"begin implementation of the 'Ticket to Independence' program,
contingent upon enactment of supporting legislation in FY
1998."[Footnote 132] This measure was neither quantifiable nor
measurable, and did not measure the number of beneficiaries achieving
this outcome.
In the 2002 report, SSA identified relevant results that are linked to
strategic objectives. For example, one of SSA's objectives related to
the strategic goal "deliver citizen-centered, world-class service" was
to "maintain the accuracy, timeliness, and efficiency of service to
people applying for Old Age and Survivors Insurance (OASI) and
Supplemental Security Income (SSI) Aged benefits." SSA reported on the
timeliness of OASI and SSI claims, as well as the implementation of
software and infrastructure for paperless processing of claims, as the
relevant results.
In its 1999 performance report, SSA noted that a number of its goals
were not met, such as those relating to accurate and timely disability
determinations. Also, data on the accuracy of decisions at the initial
level were not available, and accuracy at the appellate level was not
measured. In its 2002 report, 10 percent of the goals were not met and
10 percent were almost met. SSA's report provided explanations for 17
performance goals SSA did not meet. However, not all of the
explanations actually identified the reasons for SSA's not meeting its
goals. For example, SSA did not meet its goals for the performance
indicators "percent of 800-number calls handled accurately--payment"
and "percent of 800-number calls handled accurately--service." The
explanation noted that several quality initiatives were implemented,
but SSA did not provide explanations as to why the goals were not met.
SSA also noted that some of its performance indicators were being
eliminated in favor of more focused and outcome-based goals. In some
cases, SSA identified future plans to improve performance.
In SSA's 2002 performance and accountability report, trend information
was generally available for comparison of data back to fiscal year
1999. This information was helpful in making an assessment whether SSA
was making progress towards its goals. SSA noted that it addressed all
the IG's list of major management challenges in its report, and that it
addressed the major management challenges we identified in its annual
performance plan. SSA also addresses the progress it made against
certain challenges GAO and the IG identified during fiscal year 2002 in
its performance and accountability report.[Footnote 133] For example,
SSA highlights components of its SSI Corrective Action Plan that are
geared to improve the administration of the SSI program and get it
removed from our high-risk list. The IG's report noted that SSA needs
to have performance goals and measures that address the major
management challenges facing SSA, as they are not all addressed. For
example, performance measures were not established to address problems
with the Earnings Suspense File and the integrity of the representative
payee process.
Finally, SSA's performance and accountability report contained a
discussion of the program evaluations conducted, organized by strategic
goal. However, the program evaluations SSA identified were typically
surveys of people who did business with SSA or assessments of internal
needs, such as a survey of training effectiveness and water/air quality
surveys. While this is a slight improvement over its 1999 report, where
there was only a brief summary of program evaluations, it would be
helpful for SSA to report on whether and how its evaluations have
helped answer questions about program performance and results. We have
previously reported that evaluations can help agencies improve their
measurement of program performance and/or understanding of performance
and how it might be improved.[Footnote 134]
SSA's 2002 Report Showed a Limited Relationship between Resources and
Results:
In the fiscal year 2002 performance and accountability report, SSA's
performance goals were not aligned by budget account--rather, they were
associated with strategic goals, which in turn cross budget accounts
and programs. Thus, the monetary, human capital, and technological
resources necessary to achieve many performance goals were not
adequately described. The financial statements show a schedule of
financing and a schedule of budgetary resources for each of SSA's major
programs, and operating expenses were associated with four out of the
five strategic goals.[Footnote 135] However, these resources were not
broken down by performance goal, and were not linked to outcomes.
Additionally, as reported by the IG, SSA needs to further develop its
cost accounting system, which it began to use in fiscal year 2002; such
a system would help to link costs with performance.[Footnote 136]
SSA's 2002 Report Provided General Confidence That Performance Data Are
Credible:
While SSA provides data sources and definitions for the data supporting
its performance indicators, some data issues continue to detract from
SSA's performance report. SSA's transmittal letter noted that the
performance and financial data presented are fundamentally complete and
reliable, and that no material inadequacies were identified. Data
sources are identified for many of the performance indicators, such as
findings from evaluations and quality assurance reports. In certain
cases, data limitations are identified; for example, SSA noted that
data to support the "Percent of SSNs issued accurately" goal does not
include SSNs (social security numbers) assigned via the Enumeration-at-
Birth process and major errors identified by the Office of Quality
Assurance that do not include these SSNs result in SSN cards being
issued erroneously. Some data verification procedures are noted in the
report, but verification procedures are not consistently discussed and
data reliability and completeness is not ensured. The IG noted that of
the 21 measures it reviewed, 16 were reliable; data or documentation of
the methods used to measure SSA's performance were not available for
the other five measures. The IG went further to say that even for the
performance measures found to be reliable, SSA lacks documentation of
the methods and data used to measure its performance. Finally, data
were not available for 17 percent of the performance goals, so it was
difficult to assess whether or not progress had been made in those
areas.
Observations on the Quality of DOT's Fiscal Year 2002 Annual
Performance and Accountability Report:
DOT's fiscal year 2002 performance report provided information that
generally showed the department's performance and progress toward its
goals. Summary tables within the report showed when DOT met or did not
meet its targets and the report supplied brief analyses as to whether
or not DOT would likely meet its targets for fiscal year 2003 based on
actual performance in 2002. A separate section of the report on
performance data completeness and reliability, along with an on-line
compendium, provides a full level of confidence in DOT's performance
data. However, the report does not clearly show the relationship
between resources and results.
DOT's 2002 Report Provided a General Picture of Performance:
DOT's fiscal year 2002 performance report[Footnote 137] provides a
general picture of the department's performance. The report includes
performance summary tables, which show the progress made toward each
strategic objective. These performance summaries include actual
performance data from fiscal years 1996 to 2002 when possible, as well
as the performance targets for fiscal year 2002 and whether or not the
target was met. Similarly, we noted in our 2000 report[Footnote 138]
reviewing DOT's 1999 performance report that performance information
was clearly articulated, with summary tables listing the fiscal year
1999 goals and trend data, and checkmarks to indicate whether or not
goals were met. Figure 25 summarizes DOT's overall performance on its
40 performance targets, as reported in its 2002 report.
Figure 25: Summary of DOT's Performance Indicators for Fiscal Year
2002:
[See PDF for image]
[End of figure]
According to the report, DOT met 24 (60 percent) of its performance
targets. Fourteen (35 percent) of DOT's performance targets were not
met. The report provides explanations for why five of these targets
were unmet. For example, the target for the measure "number of
passengers (in millions) in international markets with open skies
aviation agreements" of the "mobility and economic growth" strategic
objective was unmet. The target was set at 59.7 million passengers,
while DOT's preliminary estimate for this measure indicated there were
57 million passengers. The report states that this target was unmet
because passenger travel diminished in fiscal year 2002 due to the
impact that the events of September 11, 2001, had on air travel.
However, the report did not provide explanations describing why the
nine other targets were not met. A DOT official stated that
explanations for these unmet targets were not included in the 2002
report due, in part, to time constraints. In our 2000 report of DOT's
1999 performance report, we stated that for all of its unmet goals
except transit fatalities, the department provided explanations for not
meeting the goals related to the outcomes we observed.
We noted in our report on DOT's 1999 performance report that the
department supplied strategies to achieve its unmet goals in the
future, for the areas we reviewed. However, of the 14 unmet targets in
the fiscal year 2002 report, DOT provided future plans to achieve only
two. For example, the report provided a plan for future achievement of
the unmet target "percent of environmental justice cases unresolved
after one year." The report stated that DOT's External Complaint
Tracking System was being revised "to track complaints more closely, in
a more timely way, and with a higher level of data quality." DOT is
also developing guidance requiring more intensive legal staff
involvement in external civil rights complaints, especially
environmental justice cases. A DOT official stated that future plans
were not included in the 2002 report for the other unmet targets due,
in part, to time constraints.
Data were unavailable for two of DOT's measures, "cumulative average
percent change in transit passenger-miles traveled per transit market"
and "employment sites (in thousands) made accessible by Job Access and
Reverse Commute (JARC) transportation services." The report explains
the reasons why data were unavailable for both of these measures and
includes plans to provide these data in the future. Data were
unavailable for the JARC program performance measure because DOT had
not received data from JARC grantees to verify that fiscal year 2002
program targets had been achieved. DOT states that a new reporting
system is being implemented, which should improve data gathering
performance.
Although the report does not identify any performance goals that were
impractical or infeasible, it states that the measure on transit
passenger-miles traveled had been changed in fiscal year 2002 because
"it placed excessive emphasis on increasing ridership in the Nation's
very largest urban areas." However, after using the measure for one
year, DOT concluded that the measure should once again be modified to
account for changes in the level of employment in each urban area. The
report states that a recent study found that changes in the level of
employment are a key economic factor related to changes in the level of
transit ridership.
Another strength of DOT's fiscal year 2002 performance report is an
analysis of whether or not DOT will likely meet its planned performance
targets for fiscal year 2003. Each discussion of performance goals
contains an evaluation of the fiscal year 2003 performance plan target
and whether or not it will likely be met based on the fiscal year 2002
data. For example, for its highway fatality rate targets, DOT says that
the National Highway Traffic Safety Administration and the Federal
Motor Carrier Safety Administration will be challenged to meet the
established fiscal year 2003 targets because targets had not been met
for fiscal year 2002. In other instances where DOT is sure that it will
meet its targets for fiscal year 2003, it simply states so, as in the
case of its aviation safety targets.
DOT's 2002 performance report also includes information on completed
program evaluations. There is a separate section in the report that
discusses DOT's fiscal year 2002 program evaluations. Summaries of the
findings of these evaluations are discussed in this section. For
example, an evaluation of the Noise Set-Aside Portion of the FAA
(Federal Aviation Administration) Airport Improvement Program found
that funding for the program's noise compatibility projects was
variable from year to year, making it difficult to forecast annual
population benefits.
As we found in the 1999 report, the major management challenges that
DOT faces are generally discussed in the 2002 report. These discussions
were contained within the program section to which they relate. The
report also showed the progress DOT has made in addressing its
management challenges. For example, we noted in our January 2003
report[Footnote 139] on DOT's major management challenges that FAA's
financial management systems remained at high risk. DOT's 2002 report
states that FAA created an Interim Fixed Asset System to centrally
control and account for its property and that in fiscal year 2003, FAA
will convert to use Delphi, DOT's financial accounting system.
The 2002 performance report included a discussion on strategic human
capital management as part of DOT's "organizational excellence"
strategic objective. This discussion included a brief overview of DOT's
human capital plan as well as strategies for strategic human capital
management. For example, the report noted that FAA was redirecting
37,300 employees into a results-oriented Air Traffic Organization,
"freeing most of the FAA to manage better and modernize more
efficiently." However, the report did not include a review of the
performance goals and evaluation of the performance plan relative to
the agency's strategic human capital management, as required by the
Homeland Security Act of 2002.
DOT's 2002 Report Showed No Relationship between Resources and Results:
DOT's 2002 performance report did not show the relationship between the
resources it used and the results it achieved in fiscal year 2002. The
financial portion of the report provided a statement of net cost for
each of DOT's programs in fiscal year 2002. The report could be
improved by providing net cost information for DOT's performance goals
in the performance section of the report, similar to the funding
information provided in the 2004 performance plan.
DOT's 2002 Report Provided Full Confidence That Performance Data Are
Credible:
DOT's 2002 performance report provided a full level of confidence that
the department's performance data were credible. In his transmittal
letter, the Secretary of DOT stated that the 2002 report "contains
performance and financial data that are substantially complete and
reliable." The letter also stated that a section of the report assessed
the inadequacies of DOT's performance data and provided plans to remedy
those inadequacies.
The "Performance Data Completeness and Reliability" section of the 2002
report generally discussed data completeness, reliability, and
limitations. This section discussed an overall limitation in DOT's
performance data. The report stated that much of DOT's performance data
came from external sources, and therefore, the department had no direct
control over the quality of these data. The report continues by stating
that DOT takes limitations to its external data into account when it
uses these data.
The 2002 report noted that DOT has compiled source and accuracy
statements that provide detail on the methods used to collect
performance data, sources of variation and bias in the data, methods
used to verify and validate the data, as well as data
limitations.[Footnote 140] However, the online Source and Accuracy
Compendium does not include this information for the goals and measures
related to the department's organizational excellence objective. The
compendium states that a small number of source and accuracy statements
are not yet completed and that they will be added upon completion.
Finally, the 2002 report also described strategies being undertaken to
address the quality of data used by DOT. The report stated that a DOT
intermodal working group addressed data quality issues by developing
departmental statistical standards and by updating source and accuracy
statements for all of DOT's data programs. The working group also
worked to improve quality assurance procedures, evaluate sampling and
nonsampling errors, and develop common definitions for data across
modes.
[End of section]
Appendix VI: GAO Federal Managers' Survey Data:
Q1. What is your current grade level?
GS/GM-13 or equivalent (percent): 31.4;
GS/GM-14 or equivalent (percent): 35.2;
GS/GM-15 or equivalent (percent): 24.6;
Senior Executive Service or equivalent (percent): 7.9;
Other - please specify - Continue with next question (percent): 0.9;
Number of respondents: 500.
Q1a. If you answered "Other" in question 1 above, please enter your
response below.
Writing comment (percent): 0.9;
Number of respondents: 3.
Q2. In total, for how many years have you been a supervisor and/or a
manager in the federal government?
Mean: 13.1;
Median: 13;
Minimum: 1;
Maximum: 43;
Number of respondents: 497.
Q2no. Or, if you have never been a supervisor or a manager in the
federal government, please check the box below.
Percent: 0.6;
Number of respondents: 3.
Q3. In your current role, approximately how many government employees
are you responsible for? (Please answer for your permanent position.
Please specify the total number. If none, enter 0.) Enter numeric
digits only. Employees:
Mean: 77.7;
Median: 15;
Minimum: 1;
Maximum: 11,000;
Number of respondents: 475.
Q4. Please indicate where you currently work. (If you are currently on
temporary assignment or on detail, please answer for your permanent
work location.):
Headquarters of my department or agency (percent): 28.9;
A field office of my department or agency (percent): 60.7;
Other - please specify - Continue with next question (percent): 9.5;
No answer (percent): 1.0;
Number of respondents: 503.
Q4a. If you answered "Other" in question 4 above, please enter your
response below.
Writing comment (percent): 9.5;
Number of respondents: 38.
Q5. For those program(s)/operation(s)/project(s) that you are involved
with, to what extent, if at all, do you consider your agency's
strategic goals when participating in the following activities? (Check
one box in each row.):
a. Setting program priorities;
To a very great extent (percent): 33.9;
To a great extent (percent): 45.3;
To a moderate extent (percent): 11.4;
To a small extent (percent): 4.7;
To no extent (percent): 1.8;
No basis to judge/Not applicable (percent): 2.2;
No answer (percent): 0.7;
Number of respondents: 503.
b. Allocating resources;
To a very great extent (percent): 30.3;
To a great extent (percent): 39.9;
To a moderate extent (percent): 18.9;
To a small extent (percent): 4.8;
To no extent (percent): 1.8;
No basis to judge/Not applicable (percent): 3.1;
No answer (percent): 1.3;
Number of respondents: 503.
c. Adopting new program approaches or changing work processes;
To a very great extent (percent): 33.6;
To a great extent (percent): 39.2;
To a moderate extent (percent): 16.5;
To a small extent (percent): 4.6;
To no extent (percent): 2.4;
No basis to judge/Not applicable (percent): 1.9;
No answer (percent): 1.8;
Number of respondents: 503.
d. Developing or refining program performance measures;
To a very great extent (percent): 29.0;
To a great extent (percent): 37.0;
To a moderate extent (percent): 16.3;
To a small extent (percent): 8.3;
To no extent (percent): 4.1;
No basis to judge/Not applicable (percent): 4.4;
No answer (percent): 0.9;
Number of respondents: 503.
Q6. Are there performance measures for the program(s)/operation(s)/
project(s) that you are involved with?
Yes (percent): 89.0;
No (percent): 6.2;
Do not know (percent): 3.8;
No answer (percent): 1.0;
Number of respondents: 503.
Q7. To what extent, if at all, do you agree with the following
statements as they relate to performance measures for the program(s)/
operation(s)/project(s) that you are involved with? (Check one box in
each row.) We have performance measures that.
a. Tell us how many things we produce or services we provide. (Output
measures);
To a very great extent (percent): 25.8;
To a great extent (percent): 28.7;
To a moderate extent (percent): 19.4;
To a small extent (percent): 8.4;
To no extent (percent): 4.8;
No basis to judge/Not applicable (percent): 1.8;
No answer (percent): 11.0;
Number of respondents: 503.
b. Tell us if we are operating efficiently. (Efficiency measures);
To a very great extent (percent): 15.2;
To a great extent (percent): 27.5;
To a moderate extent (percent): 25.8;
To a small extent (percent): 13.4;
To no extent (percent): 5.5;
No basis to judge/Not applicable (percent): 1.5;
No answer (percent): 11.0;
Number of respondents: 503.
c. Tell us whether or not we are satisfying our customers. (Customer
service measures);
To a very great extent (percent): 16.8;
To a great extent (percent): 29.8;
To a moderate extent (percent): 21.6;
To a small extent (percent): 12.6;
To no extent (percent): 6.1;
No basis to judge/Not applicable (percent): 1.6;
No answer (percent): 11.4;
Number of respondents: 503.
d. Tell us about the quality of the products or services we provide.
(Quality measures);
To a very great extent (percent): 16.0;
To a great extent (percent): 30.3;
To a moderate extent (percent): 23.5;
To a small extent (percent): 12.6;
To no extent (percent): 5.6;
No basis to judge/Not applicable (percent): 0.7;
No answer (percent): 11.3;
Number of respondents: 503.
e. Demonstrate to someone outside of our agency whether or not we are
achieving our intended results. (Outcome measures);
To a very great extent (percent): 19.1;
To a great extent (percent): 35.9;
To a moderate extent (percent): 19.0;
To a small extent (percent): 10.5;
To no extent (percent): 2.6;
No basis to judge/Not applicable (percent): 1.5;
No answer (percent): 11.3;
Number of respondents: 503.
f. Link our product or service costs with the results we achieve.
(Cost-benefit measures);
To a very great extent (percent): 12.5;
To a great extent (percent): 18.7;
To a moderate extent (percent): 19.7;
To a small extent (percent): 20.0;
To no extent (percent): 11.7;
No basis to judge/Not applicable (percent): 6.0;
No answer (percent): 11.3;
Number of respondents: 503.
Q8. For those program(s)/operation(s)/project(s) that you are involved
with, to what extent, if at all, do you use the information obtained
from performance measurement when participating in the following
activities?
a. Setting program priorities;
To a very great extent (percent): 13.0;
To a great extent (percent): 38.7;
To a moderate extent (percent): 21.7;
To a small extent (percent): 10.8;
To no extent (percent): 3.4;
No basis to judge/Not applicable (percent): 1.2;
No answer (percent): 11.0;
Number of respondents: 503.
b. Allocating resources;
To a very great extent (percent): 14.9;
To a great extent (percent): 36.6;
To a moderate extent (percent): 20.3;
To a small extent (percent): 10.8;
To no extent (percent): 3.7;
No basis to judge/Not applicable (percent): 1.8;
No answer (percent): 12.0;
Number of respondents: 503.
c. Adopting new program approaches or changing work processes;
To a very great extent (percent): 13.9;
To a great extent (percent): 34.8;
To a moderate extent (percent): 24.2;
To a small extent (percent): 10.2;
To no extent (percent): 4.1;
No basis to judge/Not applicable (percent): 0.9;
No answer (percent): 11.9;
Number of respondents: 503.
d. Coordinating program efforts with other internal or external
organizations;
To a very great extent (percent): 10.1;
To a great extent (percent): 32.2;
To a moderate extent (percent): 29.4;
To a small extent (percent): 10.9;
To no extent (percent): 3.6;
No basis to judge/Not applicable (percent): 2.7;
No answer (percent): 11.1;
Number of respondents: 503.
e. Refining program performance measures;
To a very great extent (percent): 14.4;
To a great extent (percent): 28.8;
To a moderate extent (percent): 24.7;
To a small extent (percent): 13.2;
To no extent (percent): 3.4;
No basis to judge/Not applicable (percent): 4.1;
No answer (percent): 11.3;
Number of respondents: 503.
f. Setting new or revising existing performance goals;
To a very great extent (percent): 16.1;
To a great extent (percent): 32.4;
To a moderate extent (percent): 21.0;
To a small extent (percent): 11.8;
To no extent (percent): 2.9;
No basis to judge/Not applicable (percent): 3.5;
No answer (percent): 12.3;
Number of respondents: 503.
g. Setting individual job expectations for the government employees I
manage or supervise;
To a very great extent (percent): 18.4;
To a great extent (percent): 33.6;
To a moderate extent (percent): 22.0;
To a small extent (percent): 9.3;
To no extent (percent): 2.8;
No basis to judge/Not applicable (percent): 2.4;
No answer (percent): 11.4;
Number of respondents: 503.
h. Rewarding government employees I manage or supervise;
To a very great extent (percent): 17.7;
To a great extent (percent): 33.3;
To a moderate extent (percent): 21.8;
To a small extent (percent): 8.2;
To no extent (percent): 4.0;
No basis to judge/Not applicable (percent): 3.0;
No answer (percent): 11.9;
Number of respondents: 503.
i. Developing and managing contracts;
To a very great extent (percent): 7.3;
To a great extent (percent): 18.8;
To a moderate extent (percent): 19.3;
To a small extent (percent): 10.4;
To no extent (percent): 8.1;
No basis to judge/Not applicable (percent): 24.4;
No answer (percent): 11.7;
Number of respondents: 503.
Q9. Based on your experience with the program(s)/operation(s)/
project(s) that you are involved with, to what extent, if at all, have
the following factors hindered measuring performance or using the
performance information?
a. Setting program priorities;
To a very great extent (percent): 14.2;
To a great extent (percent): 20.3;
To a moderate extent (percent): 32.1;
To a small extent (percent): 19.7;
To no extent (percent): 9.1;
No basis to judge/Not applicable (percent): 3.9;
No answer (percent): 0.7;
Number of respondents: 503.
b. Different parties are using different definitions to measure
performance;
To a very great extent (percent): 10.3;
To a great extent (percent): 23.5;
To a moderate extent (percent): 24.6;
To a small extent (percent): 24.6;
To no extent (percent): 10.4;
No basis to judge/Not applicable (percent): 5.6;
No answer (percent): 1.0;
Number of respondents: 503.
c. Difficulty obtaining valid or reliable data;
To a very great extent (percent): 7.5;
To a great extent (percent): 22.1;
To a moderate extent (percent): 26.0;
To a small extent (percent): 27.2;
To no extent (percent): 12.0;
No basis to judge/Not applicable (percent): 4.1;
No answer (percent): 1.0;
Number of respondents: 503.
d. Difficulty obtaining data in time to be useful;
To a very great extent (percent): 7.6;
To a great extent (percent): 17.5;
To a moderate extent (percent): 23.9;
To a small extent (percent): 26.0;
To no extent (percent): 18.3;
No basis to judge/Not applicable (percent): 4.4;
No answer (percent): 2.2;
Number of respondents: 503.
e. Lack of incentives (e.g., rewards, positive recognition);
To a very great extent (percent): 12.9;
To a great extent (percent): 15.8;
To a moderate extent (percent): 25.0;
To a small extent (percent): 24.1;
To no extent (percent): 16.6;
No basis to judge/Not applicable (percent): 4.8;
No answer (percent): 0.8;
Number of respondents: 503.
f. Difficulty resolving conflicting interests of stakeholders, either
internal or external;
To a very great extent (percent): 9.5;
To a great extent (percent): 19.3;
To a moderate extent (percent): 25.5;
To a small extent (percent): 22.0;
To no extent (percent): 14.5;
No basis to judge/Not applicable (percent): 8.1;
No answer (percent): 1.0;
Number of respondents: 503.
g. Difficulty distinguishing between the results produced by the
program and results caused by other factors;
To a very great extent (percent): 7.4;
To a great extent (percent): 18.7;
To a moderate extent (percent): 24.3;
To a small extent (percent): 28.6;
To no extent (percent): 13.7;
No basis to judge/Not applicable (percent): 6.3;
No answer (percent): 1.0;
Number of respondents: 503.
h. Existing information technology and/or systems not capable of
providing needed data;
To a very great extent (percent): 9.7;
To a great extent (percent): 19.4;
To a moderate extent (percent): 20.6;
To a small extent (percent): 26.6;
To no extent (percent): 17.0;
No basis to judge/Not applicable (percent): 5.4;
No answer (percent): 1.3;
Number of respondents: 503.
i. Lack of staff who are knowledgeable about gathering and/or analyzing
performance information;
To a very great extent (percent): 10.1;
To a great extent (percent): 17.5;
To a moderate extent (percent): 23.4;
To a small extent (percent): 26.2;
To no extent (percent): 16.5;
No basis to judge/Not applicable (percent): 5.0;
No answer (percent): 1.3;
Number of respondents: 503.
j. Lack of ongoing top executive commitment or support for using
performance information to make program/funding decisions;
To a very great extent (percent): 9.7;
To a great extent (percent): 16.0;
To a moderate extent (percent): 18.4;
To a small extent (percent): 24.0;
To no extent (percent): 23.1;
No basis to judge/Not applicable (percent): 7.9;
No answer (percent): 1.0;
Number of respondents: 503.
k. Lack of ongoing Congressional commitment or support for using
performance information to make program/funding decisions;
To a very great extent (percent): 7.1;
To a great extent (percent): 16.6;
To a moderate extent (percent): 18.1;
To a small extent (percent): 16.5;
To no extent (percent): 17.1;
No basis to judge/Not applicable (percent): 23.8;
No answer (percent): 0.7;
Number of respondents: 503.
l. Difficulty determining how to use performance information to improve
the program;
To a very great extent (percent): 5.3;
To a great extent (percent): 12.8;
To a moderate extent (percent): 30.1;
To a small extent (percent): 26.6;
To no extent (percent): 18.8;
No basis to judge/Not applicable (percent): 5.6;
No answer (percent): 0.7;
Number of respondents: 503.
m. Concern that OMB will micromanage programs in my agency;
To a very great extent (percent): 7.9;
To a great extent (percent): 10.8;
To a moderate extent (percent): 14.7;
To a small extent (percent): 19.2;
To no extent (percent): 26.6;
No basis to judge/Not applicable (percent): 19.7;
No answer (percent): 1.0;
Number of respondents: 503.
Q10. To what extent, if at all, do you agree with the following
statements? (Check one box in each row.):
a. Agency managers/supervisors at my level have the decision making
authority they need to help the agency accomplish its strategic goals;
To a very great extent (percent): 9.5;
To a great extent (percent): 30.1;
To a moderate extent (percent): 28.1;
To a small extent (percent): 25.2;
To no extent (percent): 5.6;
No basis to judge/Not applicable (percent): 0.6;
No answer (percent): 1.0;
Number of respondents: 503.
b. Agency managers/supervisors at my level are held accountable for
agency accomplishment of its strategic goals;
To a very great extent (percent): 14.6;
To a great extent (percent): 42.9;
To a moderate extent (percent): 24.1;
To a small extent (percent): 12.7;
To no extent (percent): 3.3;
No basis to judge/Not applicable (percent): 1.5;
No answer (percent): 1.0;
Number of respondents: 503.
c. Agency managers/supervisors at my level are held accountable for the
results of the program(s)/operation(s)/project(s) they are responsible
for;
To a very great extent (percent): 24.2;
To a great extent (percent): 46.5;
To a moderate extent (percent): 17.2;
To a small extent (percent): 7.2;
To no extent (percent): 3.0;
No basis to judge/Not applicable (percent): 0.9;
No answer (percent): 1.0;
Number of respondents: 503.
d. Employees in my agency receive positive recognition for helping the
agency accomplish its strategic goals;
To a very great extent (percent): 9.7;
To a great extent (percent): 27.1;
To a moderate extent (percent): 31.2;
To a small extent (percent): 22.0;
To no extent (percent): 7.2;
No basis to judge/Not applicable (percent): 1.5;
No answer (percent): 1.3;
Number of respondents: 503.
e. My agency's top leadership demonstrates a strong commitment to
achieving results;
To a very great extent (percent): 24.4;
To a great extent (percent): 37.1;
To a moderate extent (percent): 21.8;
To a small extent (percent): 9.9;
To no extent (percent): 2.9;
No basis to judge/Not applicable (percent): 2.6;
No answer (percent): 1.3;
Number of respondents: 503.
f. My agency is investing the resources needed to ensure that its
performance data is of sufficient quality;
To a very great extent (percent): 9.5;
To a great extent (percent): 21.4;
To a moderate extent (percent): 29.3;
To a small extent (percent): 21.9;
To no extent (percent): 7.4;
No basis to judge/Not applicable (percent): 9.3;
No answer (percent): 1.0;
Number of respondents: 503.
Q11. To what extent, if at all, do you agree with the following
statements? (Check one box in each row.) The following items focus on
the program(s)/operation(s)/project(s) that you are responsible for.
a. The individual I report to periodically reviews with me the results
or outcomes of the program(s) operation(s)/project(s) that I am
responsible for;
To a very great extent (percent): 18.8;
To a great extent (percent): 36.1;
To a moderate extent (percent): 23.1;
To a small extent (percent): 15.3;
To no extent (percent): 5.3;
No basis to judge/Not applicable (percent): 0.4;
No answer (percent): 1.0;
Number of respondents: 503.
b. Funding decisions for the program(s)/operation(s)/project(s) I am
responsible for are based on results or outcome-oriented performance
information;
To a very great extent (percent): 4.5;
To a great extent (percent): 20.9;
To a moderate extent (percent): 23.1;
To a small extent (percent): 26.4;
To no extent (percent): 16.5;
No basis to judge/Not applicable (percent): 7.5;
No answer (percent): 1.0;
Number of respondents: 503.
c. Staffing and personnel decisions for the program(s)/operation(s)/
project(s) I am responsible for are based on results or outcome-
oriented performance information;
To a very great extent (percent): 4.2;
To a great extent (percent): 20.9;
To a moderate extent (percent): 27.8;
To a small extent (percent): 23.2;
To no extent (percent): 19.0;
No basis to judge/Not applicable (percent): 3.7;
No answer (percent): 1.3;
Number of respondents: 503.
d. Changes by management above my level to the program(s)/operation(s)/
project(s) I am responsible for are based on results or outcome-
oriented performance information;
To a very great extent (percent): 2.6;
To a great extent (percent): 20.5;
To a moderate extent (percent): 25.2;
To a small extent (percent): 26.5;
To no extent (percent): 15.3;
No basis to judge/Not applicable (percent): 8.2;
No answer (percent): 1.6;
Number of respondents: 503.
e. It is easy to motivate employees to be more results-oriented in the
program(s)/operation(s)/project(s) I am responsible for;
To a very great extent (percent): 3.5;
To a great extent (percent): 22.0;
To a moderate extent (percent): 33.6;
To a small extent (percent): 29.0;
To no extent (percent): 7.8;
No basis to judge/Not applicable (percent): 2.8;
No answer (percent): 1.3;
Number of respondents: 503.
f. I have sufficient information on the validity of the performance
data I use to make decisions;
To a very great extent (percent): 4.7;
To a great extent (percent): 32.6;
To a moderate extent (percent): 30.2;
To a small extent (percent): 20.3;
To no extent (percent): 7.1;
No basis to judge/Not applicable (percent): 3.2;
No answer (percent): 1.8;
Number of respondents: 503.
Q12. During the past 3 years, has your agency provided, arranged, or
paid for training that would help you to accomplish the following
tasks? (Check one box in each row.):
a. Conduct strategic planning;
Yes (percent): 46.7;
No (percent): 52.3;
No answer (percent): 1.0;
Number of respondents: 503.
b. Set program performance goals;
Yes (percent): 48.8;
No (percent): 50.2;
No answer (percent): 1.0;
Number of respondents: 503.
c. Develop program performance measures;
Yes (percent): 42.9;
No (percent): 55.5;
No answer (percent): 1.6;
Number of respondents: 503.
d. Assess the quality of performance data;
Yes (percent): 35.3;
No (percent): 63.5;
No answer (percent): 1.3;
Number of respondents: 503.
e. Use program performance information to make decisions;
Yes (percent): 40.7;
No (percent): 56.8;
No answer (percent): 2.4;
Number of respondents: 503.
f. Link the performance of program(s)/operation(s)/project(s) to the
achievement of agency strategic goals;
Yes (percent): 40.8;
No (percent): 57.0;
No answer (percent): 2.2;
Number of respondents: 503.
g. Implement the requirements of the Government Performance and Results
Act (GPRA or the Results Act);
Yes (percent): 31.9;
No (percent): 66.7;
No answer (percent): 1.3;
Number of respondents: 503.
Q13. In your opinion, to what extent, if at all, do you believe you
need training (or additional training) in order to help you to
accomplish the following tasks? (Check one box in each row.):
a. Conduct strategic planning;
To a very great extent (percent): 8.2;
To a great extent (percent): 18.3;
To a moderate extent (percent): 27.8;
To a small extent (percent): 30.8;
To no extent (percent): 11.2;
No basis to judge/Not applicable (percent): 2.8;
No answer (percent): 0.9;
Number of respondents: 503.
b. Set program performance goals;
To a very great extent (percent): 8.6;
To a great extent (percent): 19.7;
To a moderate extent (percent): 27.6;
To a small extent (percent): 27.2;
To no extent (percent): 12.7;
No basis to judge/Not applicable (percent): 3.0;
No answer (percent): 1.2;
Number of respondents: 503.
c. Develop program performance measures;
To a very great extent (percent): 9.6;
To a great extent (percent): 21.1;
To a moderate extent (percent): 29.2;
To a small extent (percent): 23.7;
To no extent (percent): 11.2;
No basis to judge/Not applicable (percent): 3.6;
No answer (percent): 1.6;
Number of respondents: 503.
d. Assess the quality of performance data;
To a very great extent (percent): 8.9;
To a great extent (percent): 22.1;
To a moderate extent (percent): 25.7;
To a small extent (percent): 25.7;
To no extent (percent): 13.3;
No basis to judge/Not applicable (percent): 2.5;
No answer (percent): 1.9;
Number of respondents: 503.
e. Use program performance information to make decisions;
To a very great extent (percent): 9.5;
To a great extent (percent): 22.1;
To a moderate extent (percent): 23.7;
To a small extent (percent): 26.9;
To no extent (percent): 13.8;
No basis to judge/Not applicable (percent): 2.2;
No answer (percent): 1.9;
Number of respondents: 503.
f. Link the performance of program(s)/operation(s)/project(s) to the
achievement of agency strategic goals;
To a very great extent (percent): 12.0;
To a great extent (percent): 21.6;
To a moderate extent (percent): 24.2;
To a small extent (percent): 27.3;
To no extent (percent): 11.2;
No basis to judge/Not applicable (percent): 2.8;
No answer (percent): 0.9;
Number of respondents: 503.
g. Implement the requirements of the Government Performance and Results
Act (GPRA or the Results Act);
To a very great extent (percent): 12.0;
To a great extent (percent): 25.0;
To a moderate extent (percent): 26.4;
To a small extent (percent): 17.6;
To no extent (percent): 11.1;
No basis to judge/Not applicable (percent): 6.9;
No answer (percent): 1.0;
Number of respondents: 503.
Q14. What, in your opinion, can the Federal government do to improve
its overall focus on managing for results?
Writing comment (percent): 65.4;
Number of respondents: 503.
Q15. Prior to receiving this questionnaire , which of the following
statements best describes your awareness of GPRA?
I had never heard of GPRA. (percent): 19.5;
I had heard of GPRA but had no knowledge of its requirements.
(percent): 13.5;
I had heard of GPRA and had a low level of knowledge of its
requirements. (percent): 24.5;
I had heard of GPRA and had moderate knowledge of its requirements.
(percent): 35.4;
I had heard of GPRA and had extensive knowledge of its requirements.
(percent): 5.6;
No answer (percent): 1.5;
Number of respondents: 503.
Q16. For those program(s)/operation(s)/project(s) that you are involved
with, to what extent, if at all, do you consider the annual performance
goals set forth in your agency's GPRA annual performance plan when
participating in the following activities?
a. Setting program priorities;
To a very great extent (percent): 7.5;
To a great extent (percent): 19.8;
To a moderate extent (percent): 24.4;
To a small extent (percent): 12.7;
To no extent (percent): 13.5;
No basis to judge/Not applicable (percent): 20.8;
No answer (percent): 1.2;
Number of respondents: 503.
b. Allocating resources;
To a very great extent (percent): 5.2;
To a great extent (percent): 16.8;
To a moderate extent (percent): 23.9;
To a small extent (percent): 15.7;
To no extent (percent): 14.7;
No basis to judge/Not applicable (percent): 22.5;
No answer (percent): 1.2;
Number of respondents: 503.
c. Adopting new program approaches or changing work processes;
To a very great extent (percent): 5.9;
To a great extent (percent): 21.3;
To a moderate extent (percent): 23.8;
To a small extent (percent): 13.5;
To no extent (percent): 13.3;
No basis to judge/Not applicable (percent): 21.1;
No answer (percent): 1.2;
Number of respondents: 503.
d. Coordinating program efforts with other internal or external
organizations;
To a very great extent (percent): 5.4;
To a great extent (percent): 16.2;
To a moderate extent (percent): 26.4;
To a small extent (percent): 15.7;
To no extent (percent): 12.8;
No basis to judge/Not applicable (percent): 22.3;
No answer (percent): 1.2;
Number of respondents: 503.
e. Developing or refining program performance measures;
To a very great extent (percent): 5.7;
To a great extent (percent): 18.1;
To a moderate extent (percent): 21.5;
To a small extent (percent): 15.9;
To no extent (percent): 14.7;
No basis to judge/Not applicable (percent): 22.9;
No answer (percent): 1.2;
Number of respondents: 503.
f. Setting individual job expectations for the government employees I
manage or supervise;
To a very great extent (percent): 5.7;
To a great extent (percent): 22.0;
To a moderate extent (percent): 19.7;
To a small extent (percent): 15.4;
To no extent (percent): 15.0;
No basis to judge/Not applicable (percent): 20.5;
No answer (percent): 1.6;
Number of respondents: 503.
g. Rewarding government employees I manage or supervise;
To a very great extent (percent): 5.9;
To a great extent (percent): 19.5;
To a moderate extent (percent): 20.4;
To a small extent (percent): 16.1;
To no extent (percent): 16.2;
No basis to judge/Not applicable (percent): 20.5;
No answer (percent): 1.5;
Number of respondents: 503.
h. Developing and managing contracts;
To a very great extent (percent): 3.3;
To a great extent (percent): 12.3;
To a moderate extent (percent): 14.9;
To a small extent (percent): 13.3;
To no extent (percent): 16.4;
No basis to judge/Not applicable (percent): 37.1;
No answer (percent): 2.7;
Number of respondents: 503.
Q17. During the past 3 years, have you been involved in these
activities? (Check one box in each row.) GPRA-related activities:
a. Developing ways to measure whether program performance goals are
being achieved;
Yes (percent): 45.8;
No (percent): 52.7;
No answer (percent): 1.5;
Number of respondents: 503.
b. Gathering and analyzing data to measure whether programs are meeting
their specific performance goals;
Yes (percent): 50.6;
No (percent): 47.9;
No answer (percent): 1.5;
Number of respondents: 503.
c. Using measures for program performance goals to determine if the
agency's strategic goals are being achieved;
Yes (percent): 42.6;
No (percent): 55.9;
No answer (percent): 1.5;
Number of respondents: 503.
d. Assessing the quality of data used in measuring performance;
Yes (percent): 39.7;
No (percent): 58.2;
No answer (percent): 2.1;
Number of respondents: 503.
Q18. To what extent, if at all, do you agree with the following
statements? (Check one box in each row.) Extent I agree with the
following statements:
a. The objectives of my program(s)/operation(s)/project(s) are in
alignment with my agency's strategic plan under GPRA;
To a very great extent (percent): 10.1;
To a great extent (percent): 29.4;
To a moderate extent (percent): 15.8;
To a small extent (percent): 7.2;
To no extent (percent): 1.2;
No basis to judge/Not applicable (percent): 35.0;
No answer (percent): 1.2;
Number of respondents: 503.
b. The costs associated with implementing GPRA have taken time or funds
away from other important activities or projects;
To a very great extent (percent): 4.8;
To a great extent (percent): 7.7;
To a moderate extent (percent): 13.1;
To a small extent (percent): 15.1;
To no extent (percent): 10.6;
No basis to judge/Not applicable (percent): 46.1;
No answer (percent): 2.5;
Number of respondents: 503.
c. The benefits to my agency that are achieved by implementing GPRA are
worth the costs incurred in doing so (e.g., in time, money, and
effort);
To a very great extent (percent): 3.8;
To a great extent (percent): 11.3;
To a moderate extent (percent): 15.8;
To a small extent (percent): 12.4;
To no extent (percent): 7.4;
No basis to judge/Not applicable (percent): 48.0;
No answer (percent): 1.3;
Number of respondents: 503.
d. GPRA strategic and annual performance plans are mostly a repackaging
of goals, measures, and objectives that were already being used within
my agency;
To a very great extent (percent): 5.7;
To a great extent (percent): 22.7;
To a moderate extent (percent): 19.1;
To a small extent (percent): 7.7;
To no extent (percent): 2.8;
No basis to judge/Not applicable (percent): 40.1;
No answer (percent): 1.9;
Number of respondents: 503.
e. Managerial effectiveness is impeded by the lack of integration
between GPRA and other federal management programs;
To a very great extent (percent): 3.8;
To a great extent (percent): 7.7;
To a moderate extent (percent): 16.3;
To a small extent (percent): 13.4;
To no extent (percent): 6.8;
No basis to judge/Not applicable (percent): 50.5;
No answer (percent): 1.5;
Number of respondents: 503.
f. My agency considers contributions to and comments on GPRA plans or
reports from managers/supervisors at my level: ;
To a very great extent (percent): 3.5;
To a great extent (percent): 10.6;
To a moderate extent (percent): 15.6;
To a small extent (percent): 15.0;
To no extent (percent): 11.7;
No basis to judge/Not applicable (percent): 42.1;
No answer (percent): 1.5;
Number of respondents: 503.
g. Agency managers/supervisors at my level use GPRA annual performance
plans to manage their program(s)/operation(s)/project(s);
To a very great extent (percent): 3.6;
To a great extent (percent): 7.8;
To a moderate extent (percent): 18.5;
To a small extent (percent): 16.4;
To no extent (percent): 17.7;
No basis to judge/Not applicable (percent): 33.9;
No answer (percent): 2.1;
Number of respondents: 503.
h. GPRA's planning and reporting requirements impose a significant
paperwork burden;
To a very great extent (percent): 2.9;
To a great extent (percent): 7.4;
To a moderate extent (percent): 17.1;
To a small extent (percent): 15.9;
To no extent (percent): 7.3;
No basis to judge/Not applicable (percent): 47.7;
No answer (percent): 1.8;
Number of respondents: 503.
i. GPRA has caused agency managers/supervisors at my level to place a
greater emphasis on getting input from appropriate stakeholders on
their interests and expectations;
To a very great extent (percent): 2.5;
To a great extent (percent): 7.0;
To a moderate extent (percent): 20.8;
To a small extent (percent): 13.1;
To no extent (percent): 12.2;
No basis to judge/Not applicable (percent): 42.2;
No answer (percent): 2.1;
Number of respondents: 503.
Q19. To what extent, if at all, do you believe that GPRA has improved
your agency's ability to deliver results to the American public.
To a very great extent (Continue with question 19a. ) (percent): 1.6;
To a great extent (Continue with question 19a. ) (percent): 6.9;
To a moderate extent (Continue with question 19a. ) (percent): 14.5;
To a small extent (percent): 25.2;
To no extent (percent): 11.6;
No basis to judge/Not applicable (percent): 38.0;
No answer (percent): 2.1;
Number of respondents: 503.
Q19a. Please briefly describe how GPRA has improved your agency's
ability to deliver results to the American public.
Writing comment (percent): 23.0;
Number of respondents: 139.
Q20. To what extent, if at all, do you believe the following persons or
entities pay attention to your agency's efforts under GPRA? (Check one
box in each row.):
a. Department Secretary (if applicable);
To a very great extent (percent): 10.3;
To a great extent (percent): 18.3;
To a moderate extent (percent): 10.6;
To a small extent (percent): 6.2;
To no extent (percent): 3.6;
No basis to judge/Not applicable (percent): 47.1;
No answer (percent): 3.9;
Number of respondents: 503.
b. Agency head other than Department Secretary (if applicable);
To a very great extent (percent): 11.4;
To a great extent (percent): 20.6;
To a moderate extent (percent): 12.9;
To a small extent (percent): 6.9;
To no extent (percent): 2.1;
No basis to judge/Not applicable (percent): 42.9;
No answer (percent): 3.2;
Number of respondents: 503.
c. The individual I report to;
To a very great extent (percent): 8.5;
To a great extent (percent): 16.0;
To a moderate extent (percent): 19.3;
To a small extent (percent): 14.0;
To no extent (percent): 13.2;
No basis to judge/Not applicable (percent): 26.6;
No answer (percent): 2.4;
Number of respondents: 503.
d. Managers and supervisors at my level;
To a very great extent (percent): 5.1;
To a great extent (percent): 13.2;
To a moderate extent (percent): 19.7;
To a small extent (percent): 19.5;
To no extent (percent): 16.0;
No basis to judge/Not applicable (percent): 24.1;
No answer (percent): 2.4;
Number of respondents: 503.
e. Employees who report to me;
To a very great extent (percent): 2.5;
To a great extent (percent): 8.3;
To a moderate extent (percent): 13.8;
To a small extent (percent): 18.1;
To no extent (percent): 30.7;
No basis to judge/Not applicable (percent): 24.0;
No answer (percent): 2.7;
Number of respondents: 503.
f. Office of Management and Budget;
To a very great extent (percent): 14.1;
To a great extent (percent): 16.7;
To a moderate extent (percent): 11.4;
To a small extent (percent): 4.0;
To no extent (percent): 1.8;
No basis to judge/Not applicable (percent): 49.0;
No answer (percent): 2.9;
Number of respondents: 503.
g. Congressional committees;
To a very great extent (percent): 7.8;
To a great extent (percent): 14.5;
To a moderate extent (percent): 13.0;
To a small extent (percent): 8.2;
To no extent (percent): 3.5;
No basis to judge/Not applicable (percent): 50.3;
No answer (percent): 2.6;
Number of respondents: 503.
h. The audit community (e.g., GAO, Inspectors General);
To a very great extent (percent): 11.7;
To a great extent (percent): 16.9;
To a moderate extent (percent): 11.5;
To a small extent (percent): 7.3;
To no extent (percent): 1.2;
No basis to judge/Not applicable (percent): 48.2;
No answer (percent): 3.2;
Number of respondents: 503.
i. The general public;
To a very great extent (percent): 1.5;
To a great extent (percent): 2.5;
To a moderate extent (percent): 6.1;
To a small extent (percent): 17.5;
To no extent (percent): 26.4;
No basis to judge/Not applicable (percent): 42.2;
No answer (percent): 3.9;
Number of respondents: 503.
Q21. To what extent, if at all, do you believe that efforts to
implement GPRA to date have improved the program(s)/operation(s)/
project(s) in which you are involved?
I have not been sufficiently involved in GPRA to have an opinion.
(percent): 48.7;
To a very great extent (percent): 2.9;
To a great extent (percent): 5.2;
To a moderate extent (percent): 14.7;
To a small extent (percent): 16.2;
To no extent (percent): 10.0;
No answer (percent): 2.4;
Number of respondents: 503.
Q22. To what extent, if at all, do you believe that efforts to
implement GPRA to date have improved your agency's programs/operations/
projects?
I have not been sufficiently involved in GPRA to have an opinion.
(percent): 47.9;
To a very great extent (percent): 2.3;
To a great extent (percent): 6.2;
To a moderate extent (percent): 17.0;
To a small extent (percent): 16.3;
To no extent (percent): 8.7;
No answer (percent): 1.5;
Number of respondents: 503.
Q23. To what extent, if at all, do you believe implementing GPRA can
improve your agency's programs/operations/projects in the future?
To a very great extent (percent): 3.4;
To a great extent (percent): 11.9;
To a moderate extent (percent): 23.7;
To a small extent (percent): 16.7;
To no extent (percent): 5.6;
No basis to judge (percent): 35.7;
No answer (percent): 3.0;
Number of respondents: 503.
Q24. If you have been involved to any extent in implementing GPRA for
the program(s)/operation(s)/project(s) you are involved with, what has
been your greatest difficulty, and in what ways, if any, do you think
this difficulty could be addressed?
Writing comment (percent): 26.0;
Number of respondents: 503.
Q25. If you have additional comments regarding any previous question or
any comments/suggestions concerning GPRA, please use the space provided
below.
Writing comment (percent): 16.5;
Number of respondents: 503.
Note: Percents reported are weighted percents based on the population
size. Unweighted N reported for each item.
[End of table]
[End of section]
Appendix VII: Agencies Subject to the Chief Financial Officers Act:
The Chief Financial Officers Act of 1990 (the CFO Act) created the
position of Chief Financial Officer in each executive department and in
each major executive agency in the federal government. The agencies
covered by the CFO Act are:
1. Agency for International Development:
2. Department of Agriculture:
3. Department of Commerce:
4. Department of Defense:
5. Department of Education:
6. Department of Energy:
7. Department of Health and Human Services:
8. Department of Housing and Urban Development:
9. Department of the Interior:
10. Department of Justice:
11. Department of Labor:
12. Department of State:
13. Department of Transportation:
14. Department of the Treasury:
15. Department of Veterans Affairs:
16. Environmental Protection Agency:
17. Federal Emergency Management Agency[Footnote 141]
18. General Services Administration:
19. National Aeronautics and Space Administration:
20. National Science Foundation:
21. Nuclear Regulatory Commission:
22. Office of Personnel Management:
23. Small Business Administration:
24. Social Security Administration[Footnote 142]
[End of section]
Appendix VIII: Comments from the Office of Management and Budget:
EXECUTIVE OFFICE OF THE PRESIDENT:
OFFICE OF MANAGEMENT AND BUDGET
WASHINGTON, D.C. 20503:
DEPUTY DIRECTOR FOR MANAGEMENT:
February 2, 2004:
Ms. Pat Dalton:
Director:
Strategic Issues:
General Accounting Office:
441 G Street, NW:
Washington, DC 20548:
Dear Ms. Dalton:
Thank you for the opportunity to comment on the draft GAO report on
GPRA implementation (Results-Oriented Government: GPRA has Established
a Solid Foundation for Achieving Greater Results, GAO-04-38).
We appreciate GAO's extensive review of the Executive Branch's
implementation of the Government Performance and Results Act (GPRA). We
agree that GPRA provides a foundation for making the Federal government
more results-oriented. However, too often, neither Congress nor the
Executive Branch was using GPRA performance information to manage or
oversee programs. As you know, we are giving enhanced attention to
agency and program results and identifying the corrective action
necessary to overcome any shortcomings in program management or
performance. Our goal should be to implement those corrective actions,
which will help us realize the results-orientation envisioned by GPRA.
Your recommendations will enhance our efforts to make the government
more results-oriented. For instance, we will continue to use the
Program Assessment Rating Tool (PART) to improve agency performance
measurement practices and share those practices across
government. And we will revise our guidance, both for GPRA and the
PART, to clarify the integrated and complementary relationship between
the two initiatives. We will also work with agencies to ensure they are
provided adequate training in performance management, a key objective
of the PMA.
With respect to your recommendation to produce a government-wide
strategic and performance plan, we believe that the President's annual
Budget represents the Executive Branch's government-wide strategic and
performance plan. The President's Budget provides a
strategic cross-cutting view of the President's priorities, and the
Budget is providing increasing emphasis on performance. The Budget
includes a discussion of agency performance against goals, as well as
agency status and progress on the President's Management Agenda. As the
Administration works to reach its goal of applying the PART to all
programs administered by the Federal government, we will be able to
include in the President's Budget outcome-oriented goals for all the
government's programs.
Thank you for the opportunity to review and comment on your draft
report. I appreciate your willingness to take our oral and written
comments into consideration in the final draft. I look forward to
working with you to improve the ways in which we are creating a
results-oriented government.
Sincerely,
Signed by:
Clay Johnson III:
[End of section]
Appendix IX: Comments from the Department of Energy:
Department of Energy:
Washington, DC 20585:
January 15, 2004:
Ms. Patricia A. Dalton, Director:
Strategic Issues:
U.S. General Accounting Office:
Washington, DC 20548:
Dear Ms. Dalton:
We are forwarding our comments on the draft audit entitled, "Results -
Oriented Government, GPRA has Established a Solid Foundation for
Achieving Greater Results, GAO-04-38, January 2004." In this report,
GAO reviewed the draft strategic plans, annual performance plans, and
annual performance reports for the Department of Energy (DOE) and five
other agencies. The DOE took a different approach to preparing these
documents than some other agencies. Below we have demonstrated how we
complied with GPRA guidelines in preparing the required documents, yet
in some cases GAO did not give us the proper credit.
In regard to Annual Performance Plans (APPS), GAO compared agencies'
APPs for 2004 with 1999. The report (pages 8-9) indicates that DOE was
the only agency of the six reviewed that did not:
* Maintain or improve our efforts to provide a clear picture of intended
performance.
* Specify the strategies and resources we intended to use to achieve our
performance goals.
* Report on how we would ensure performance data will be credible.
DOE disagrees with all three comments.
* In the introduction section of the FY 2004 APP, we provide a clear
picture of intended performance. Each of the 17 General Goals contains
a description of the program. A good example of providing a clear
picture of performance can be found on page 12 of our FY 2004 APP in
General Goal 1 where each activity (hydrogen, wind, hydropower, solar,
etc.) is separately addressed.
* Each General Goal contains a paragraph on the means and strategies
that will be used to achieve our goals. In addition, funding is
included at the General Goal level.
* The introduction contains a section on "Criteria for Assessing our
Results." Under the Validation and Verification section on page 4 of
our FY 2004 APP, we discuss the Joule software package used for
documenting and tracking performance data.
On page 1 I of the report, GAO states that "The clarity of Energy's
plan remained limited because its annual goals were not clearly linked
to its mission, the long-term goals in its strategic plan, or the
program activities in its budget request." This is not a accurate
statement. As was explained to the GAO team that visited with DOE, when
the FY 2004 APP was prepared, the draft strategic plan contained 17
General Goals and matched the published version of the FY 2004 APP. As
a result of coordinating our draft strategic plan within DOE and with
the Office of Management and Budget, the draft strategic plan used for
this report had been reduced to seven General Goals. Had the draft of
the strategic plan in its place at the time the FY 2004 APP was
prepared been used, there would have been a one-to-one correlation
between the goals in the APP and the strategic plan.
The portion of the draft GAO report dedicated to Annual Performance
Reports for FY 2002 (i.e., Performance and Accountability Report (PAR))
states on page 19, that six agencies (including DOE) did not
consistently report the reasons for not meeting their goals. In the DOE
FY 2002 PAR, each target that was not met provided a plan of action for
addressing the causes of not meeting the target. An example can be
found on page 32 of the FY 2002 PAR, under SC3-1, where an action plan
is provided for a target that was not met.
Also in reference to the DOE FY 2002 PAR, page 105 of the draft GAO
report indicates that DOE did not provide a discussion of the
relationship between the strategic plan, performance plan, and the
performance report. However, page 80 of the FY 2002 PAR under the
introductory section of the Detailed Performance Results contains a
paragraph dedicated to providing that linkage.
We are hopeful that our comments will lead to a fair and accurate
presentation of DOE in your final report. Our staff is available to
meet with you to discuss these points in further detail. In order to
arrange such a meeting, please contact Van Jones on 202-586-4050.
Signed by:
James G. Powers:
Director, Office of Program Assessment and Evaluation/OMBE:
The following are our comments on DOE's letter dated January 15, 2004.
GAO Comments:
The Director of the Office of Program Assessment and Evaluation
forwarded written comments from DOE on a draft of this report. DOE
disagreed with several of our conclusions concerning its 2004 Annual
Performance Plan and 2002 Performance and Accountability Report. We
incorporated the additional information and perspectives of DOE into
our report as appropriate.
1. We stated that, when compared to the 1999 Annual Performance Plan,
DOE's 2004 Annual Performance Plan continued to provide a limited
picture of intended performance. DOE disagreed, stating that the plan
provided a clear picture of intended performance because a description
of each program was contained in each of the general goals. We agree
that a description of programs is provided for each goal, and we also
stated in our draft report that improvement was made in developing
results-oriented performance measures that pertained specifically to
fiscal year 2004. However, in our view, describing each program does
not sufficiently explain DOE's expectations for intended performance.
More specifically, we found the overall picture of intended performance
was limited because DOE did not specifically describe how it
coordinates with other agencies to accomplish crosscutting programs and
did not provide a clear link between its annual goals and its mission
and strategic goals. In our draft report we acknowledged that a link
did not exist between the performance plan and the strategic plan
because the strategic plan was revised after the performance plan was
finalized. Nevertheless, DOE did not revise the final 2004 performance
plan to reflect its changes in strategic goals. The lack of alignment
between the performance plan and strategic plan goals limits the
usefulness of the performance plan to support managers and staff in
their day-to-day activities in achieving DOE's long-term strategic
goals.
2. In response to our observation that DOE provided a general
discussion of the strategies and resources needed to achieve its
performance goals, DOE stated that its 2004 annual performance plan
provided specific strategies and resources that will be used to achieve
performance goals. DOE also noted that funding was included at the
general goal level. We agree that funding was included at the general
strategic goal level. However, better plans relate resources to the
achievement of performance goals. DOE did not provide resource
information at the performance goal level. Furthermore, while DOE
discussed external factors that could affect its ability to achieve its
performance goals at a high level, it did not discuss any specific
strategies to mitigate those factors.
3. DOE disagreed with our characterization that its 2004 annual
performance plan provided limited confidence that performance data will
be credible. The department stated that the introduction section of its
plan contained specific sections on assessing results and validating
and verifying data, as well as discussed a software package used to
document and track performance. In our draft report, we stated that
DOE's plan showed some improvement over its 1999 plan by describing
credible procedures to verify and validate performance information and
by mentioning specific program evaluations for each goal. We also noted
that DOE acquired new commercial software for performance tracking
through remote data entry, monitoring, and oversight by program offices
and managers. However, we concluded that DOE's reporting of credible
performance data was limited because its plan does not specifically
identify data limitations overall or for each of its goals. As we
stated in our report, we found this to be of particular concern
because, as we mentioned in our 2003 performance and accountability
series, DOE has several management challenges where data quality is a
concern.
4. Concerning its 2002 Annual Performance and Accountability Report,
DOE stated that it provided a plan of action for addressing the causes
of targets that were not met. We agree, and in our draft report we
state that all targets that were not met or had mixed results contained
a plan of action to achieve the target in the future. We also found
that the majority of DOE targets that were not met or had mixed results
contained clear explanations. We revised our text in the final version
of this report to make these findings more evident.
5. Finally, DOE disagreed with our finding that it did not provide a
discussion of the relationship between the strategic plan, performance
plan, and the performance report in its 2002 Performance and
Accountability Report. DOE stated that the introductory section of the
report contains a paragraph that discusses the linkages between these
three reports. However, although the performance and accountability
report links these documents by organizing its results section
according to strategic goals, associated program performance goals and
targets, it did not succinctly demonstrate how the results relate to
the annual and long-term strategic goals. We modified the draft
accordingly to clarify this point.
[End of section]
Appendix X: Comments from the Department of Housing and Urban
Development:
DEPARTMENT OF HOUSING AND URBAN DEVELOPMENT
WASHINGTON, DC 20410-3000:
OFFICE OF THE CHIEF FINANCIAL OFFICER:
JAN 9 2003:
Ms. Patricia A. Dalton:
Director:
Strategic Issues:
U.S. General Accounting Office:
451 G Street NW:
Washington, DC 20548:
The Department of Housing and Urban Development (HUD) appreciates the
recognition of the substantial progress HUD has made in meeting the
Government Performance and Results Act requirements in your draft
report entitled "Results-Oriented Government: GPRA has Established
a Solid Foundation for Achieving Greater Results." All of the areas
that GAO has suggested for further improvement are in process or under
consideration at HUD.
Per your request, we are providing the enclosed comments for your
consideration in clarifying or correcting the draft report. Please
contact James Martin at (202) 708-0638 ext. 3706 or Kenneth M.
Leventhal (202) 708-0614 ext. 6849 if you have any questions.
Sincerely yours,
Signed by:
De W. Ritchie:
Deputy Chief Financial Officer:
Enclosure:
ENCLOSURE:
Comments for GAO:
Strategic Plan:
Comment 1: In the last paragraph on page 4 of the draft report, we
question why HUD is omitted from the list of agencies whose strategic
plans contained long-term goals that: demonstrated improvements in the
quality of their 1997 goals; covered their missions; were results-
oriented; and were expressed in a manner that could be used to gauge
future success. HUD's March 2003 Strategic Plan addresses all these
issues. The omission of HUD from this listing appears inconsistent with
the HUD progress recognized by GAO in Table 1 on page 3 of the draft
report.
In addition, HUD does not believe that the improvements in both long-
term numerical and intermediate goals are fully reflected. We would
first comment that the nature of a Strategic Plan has certain
limitations as to the level of details and that the linkages flow in
large part from the corresponding development of the Annual Performance
Plan. Furthermore, we do not agree with the assessment on page 41 that
the link between long-term and intermediate goals is difficult to
discern. For example, in the last paragraph of page 41, the general
point is not apparent. The statement that "long-term performance
measures are not consistently linked to a corresponding intermediate
measure" seems to miss the point that both long-term and intermediate
measures are aligned with strategic goals. The need for a direct
linkage with each other is not apparent. Second, the example given
seems to be characterized as a problem simply because the baseline has
not yet been determined. It is not clear how this fact illustrates the
general point. Finally, the statement about the number of respondents
to be surveyed seems out of place; the appropriate consideration for a
Strategic Plan would seem to be whether a statistically reliable
measure currently exists or is planned. HUD suggests that extraneous
material be edited out of the paragraph to clarify the point, and that
GAO further consider the following points.
* HUD worked diligently to improve the quality of its long-term
strategic goals and substantially increased the number of long-term
goals and advanced the quality by committing to many numerical goals
that can be reviewed, as was requested during Congressional
consultation. The provision of numerical goals makes HUD's goals
particularly meaningful since they can then be measured. In addition,
many of the long-term performance measures reflect a maturing of HUD's
efforts and ability to measure performance accurately. As an example,
the PHAS and SEMAP systems have now been in place for sufficient time
to provide over the course of the strategic plan accurate measurement
of progress in our Public Housing and Section 8 programs which
represent close to 70 percent of the Department's overall resources.
* HUD also feels strongly that the long-term performance measurements
under Strategic Goal: Embrace High Standards of Ethics, Management and
Accountability reflect substantial progress in capturing the major
management challenges facing the Department as identified by GAO and IG
reviews.
Comment 2: On page 40, HUD is given credit for explaining how it will
coordinate with other agencies but credit should also be recognized on
page 7. Paragraph 3, states that none of the agencies provided any
details regarding crosscutting activities. Following are citations from
HUD's Strategic Plan, for GAO's consideration:
* On page 7. HUD cites its intention to "Work with the Department of
Treasury to enact a tax credit for developers of affordable single-
family housing."
* On page 17, under "Increase-housing opportunities for persons with
disabilities HUD cites "...a pilot program in 11 states called Project
Access, a joint effort between HUD and the Department of HHS designed
to ease the transition of non-elderly persons with disabilities from
nursing homes into community living and also cites...."
* "Signing a Memorandum of Understanding with the Department of Justice
and IRS to ensure that multifamily properties within the Low-Income
Housing Tax Credit Program, the number one producer of low-income
housing, comply with federal accessibility standards."
* On page 26, HUD details the interagency approach in fighting
homelessness reflected in the reauthorized Interagency Council on the
Homeless and the coordination between the Secretaries of HUD, HHS, and
Veterans Affairs.
* On page 34, HUD cites the cooperative efforts with the Department of
Justice to enforce the Fair Housing Act and with the Departments of
Treasury and Justice to ensure the Low-Income Housing Tax Credit
projects are in compliance with the Fair Housing Act.
* On page 50 the Plan also refers to the conducting of interagency
events and conferences designed to educate and train faith-based and
community organizations on equal access issues and partnership
opportunities.
* Finally, the Department has a more extensive discussion of Interagency
Partnerships on page 53, which highlights 6 additional interagency
efforts.
Comment 3: On page 42, GAO states that, "HUD's current strategic plan
does not describe how program evaluations were used to develop its
strategic goals or other components of its plan, and does not include a
schedule for future evaluations." On page 3, HUD requests that in Table
1 and in the discussion on page 3 GAO will recognize HUD's fulfillment
of the Evaluations element in our Strategic Plan. For example, the
closing cost research discussed on page 9 is the only feasible way to
address the issue of home purchase transaction costs. The research
directly supports the long-term performance measure and the strategic
objective, and the discussion specifies a timetable for replicating the
evaluation for long-term performance assessment. Similarly, the
"awareness of fair housing laws" discussion (page 35, as well as 33-34)
shows how research provided a baseline and was used to establish a
goal, and also provides a timetable for replicating the study. Numerous
similar examples appear.
HUD was not recognized for the Plan's extensive discussion of policy
development and program evaluation supporting strategic goals (pages
57-60). For example, the "increasing minority homeownership" and
"Affordable housing goals for GSEs" discussions on
page 57 explain how research helped establish the goals and in turn how
research is being shaped to address policy issues that are critical to
goal achievement. Program evaluation also is mentioned throughout the
document, and plays a central role in the establishment of performance
goals; a number of the long-term performance measures and intermediate
measures explicitly rely on research efforts to supplement the
performance metrics available from Administrative data for the reasons
discussed above, HUD requests that GAO modify Table 111. 4 and the
discussion on page 39 to recognize the inclusion of evaluations
material in HUD's Strategic Plan. The plan explains how evaluations
were used to establish and measure goals in the Strategic Plan, and
identifies timetables for several evaluations. In light of this
material, it seems appropriate to change the "No" in Table 111. 4 to
"Yes."
HUD believes that the second paragraph of page 42 is factually
incorrect for the reasons discussed above. It would seem appropriate
for GAO to recognize the ways in which HUD's Strategic Plan fulfills
the GPRA evaluations requirement and to identify ways in which it falls
short of GAO's criteria for a successful presentation of evaluation
information.
Comment 4: On page 42 the draft rises the issue of how HUD will
"discern between those jobs created by CDBG and those created by other
means", this question is more appropriate to a data review or data
audit and seems unnecessary for this report.
Annual Performance Plans:
Comment 1: HUD included a detailed Means and Strategy discussion for
each of our Strategic Goals and we believe we should be added to the
DOE citation on pages 12-13.
Comment 2: On page 75, HUD's goal is misstated: it should read,
"Promote decent affordable housing." A strategic objective is misstated
on the same page: "Increase minority homeownership."
Performance and Accountability Report:
Comment 1: On page 13, for the FY 2002 PAR, HUD for the first time
provided a table akin to the SSA table with results shown in the front
of each Strategic Goal Section. The table/shows whether HUD
"substantially met" (a check mark) or did not meet (no check mark) and/
or provided extensive footnotes where necessary to clarify the result.
In addition, HUD had approximately 190 data elements to report on
compared to the 69 cited for SSA.
We recognize that the number of indicators required by GPRA guidance
regarding full coverage of program and activities does require a
thorough effort to review HUD's inclusive performance results. Thus,
the multi-pronged nature of HUD's mission and programs to some extent
precludes simple summaries of overall performance. HUD believes that
the summary tables for each strategic goal provided the best and most
feasible balance of simplicity, transparency and completeness for the
Department's strategic framework and programmatic structure.
Comment 2: On page 19, the report says that HUD, SBA and DOT did not
discuss plans or strategies to achieve unmet goals in the future and
that HUD and several other agencies did not report the reasons for not
meeting their goals. HUD believes that we consistently tried to report
on reasons for not meeting goals as well as discussing strategies to
achieve unmeet goals in the future. We would concede that we may not
have done this in every case and that is in part because it is often
difficult to assess why some goals were not met and therefore what
actions might improve performance. As an example of our consistent
effort, on the very first goal and indicator in which we did not meet a
goal, pages 2-13, indicator 1. 1.f., the first paragraph under
"Results" gives a reason for missing the target and the following
paragraph describes our planned efforts to improve results. This same
approach is on the very next indicator on pages 2-14, indicator 1.1.13
and is replete throughout the report.
Comment 3: Regarding the middle paragraph on page 108 of the draft
report, HUD requests that GAO consider: 1) the impacts of acceleration
of the performance report issuance on an agency's ability to include an
evaluation of the new fiscal year's performance plan relative to the
performance attained in the just completed fiscal year, and 2)
reasonable alternative actions to fulfill this requirement. At HUD, the
issuance of the annual performance report was initially completed in
six months or 180 days after the end of the fiscal year and this has
been accelerated to 120 days last year, 80 days this year and will be
in only 45 days next year. As a reasonable alternative, HUD generally
considers prior year performance in conjunction with updating the
annual performance plan for each upcoming year.
The intended point of the last paragraph on page 108 needs to be
clarified for HUD and other users of this report. Whereas the previous
paragraph contends that HUD's FY 2002 performance report did not
include an evaluation of the FY 2003 performance plan relative to
performance attained in FY 2002, the paragraph in question cites two
examples where HUD's FY 2002 performance report reflects an adjustment
of FY 2003 plans based on an assessment of prior year performance. If
the examples are to be retained in the final report, HUD requests that
they be put in proper context because the current wording incorrectly
implies that HUD no longer has an interest in establishing and
measuring goals for homeownership or housing quality. HUD did eliminate
the two indicators in question because they were too broad and beyond
the direct scope or influence of HUD's programs. However, the draft
report fails to indicate that HUD's performance plans and reports
retained other indicators that capture the direct impact of HUD
programs on homeownership and housing quality.
Comments 4: On page 110, we disagree with the chart's depiction that 35
percent of HUD's performance targets were "undetermined" and that 15
percent had "no explanation." While it is true that detailed
explanations of missed goals were not provided in the summary charts at
the beginning of each strategic goal section, many of the missed
targets had explanations in the individual indicator write-ups that
followed the summaries. We request that GAO reassess the basis for the
percentages and labels reflected in this chart to accurately reflect
the content of HUD's performance report. We are concerned that GAO has
not recognized the complexity in reporting whether indicator
performance goals have been substantially met or missed. The estimate
of 47 percent is of concern since HUD has not done that measurement and
there can be very legitimate differences as to how to score indicators
where "results are too complex to summarize" or data is unavailable. We
are also concerned how you determined that 2 percent of data is "not
reliable."
More specifically, on page 110, the footnote to Figure V.4 appears to
mischaracterize the purpose of the footnotes in HUD's performance
report cards. The footnotes are intended to briefly and transparently
disclose attributes of the performance data in the report cards that
could mislead a reader who does not read the full discussion of each
performance indicator. In a number of instances, the explanations
appear along with a statement of whether targets were missed. HUD
requests that GAO revise the statement "Rather than stating if some
performance targets were met or not, HUD provided the following
explanations" to capture our efforts to better inform the reader.
In Figure V.4, it is unclear why HUD indicators are characterized as
"undetermined," "no explanation," and "not reliable," as GAO has not
defined these terms or their criteria. The preponderance of
"undetermined" and "no explanation" designations is especially puzzling
and suggests that GAO may have misinterpreted the performance report
cards. HUD requests that GAO review these indicators for
recategorization as "data not available" (as denoted by footnote "a" or
"goal not achieved." No indicator should be classified as
"undetermined" simply because HUD described it with a footnote.
Agencies also appear to be treated inconsistently in the use of
categories and the way indicators are assigned to categories. For
example, a number of HUD indicators did not have a performance target
for fiscal year 2002, usually because data sources were not scheduled
to be available. These indicators were denoted by footnote "b" in HUD's
performance report card. GAO characterized these indicators as "data
not expected" for the Department of Education (Figure V.1), but has not
used this category for HUD. HUD requests that GAO review all indicators
with footnote "b" for recategorization as "data not expected."
Comment 5: On page 93, first paragraph, the first sentence seems to be
confusing the annual performance "plan" with the annual performance
"report," as it is unclear what the report could do to make the plan
more credible.
Comment 6: On page 107, second paragraph, it appears that "fiscal"
should be changed to "fiscal year."
The following are our comments on HUD's letter dated January 9, 2004.
GAO Comments:
HUD provided written comments and disagreed with several of our
observations, which we address below. HUD also mentioned that all of
the areas we suggested for further improvement were already in process
or being considered. Where appropriate, we incorporated HUD's comments
and perspectives to clarify our report.
1. HUD did not agree with our observation that the link between long-
term and intermediate goals in its strategic plan is difficult to
discern. The department mentioned that the need for a direct link
between long-term and intermediate goals is not apparent, as they are
aligned with strategic goals. GPRA requires that an agency's strategic
plan contain, among other things, a description of the relationship
between the long-term goals and objectives and the annual performance
goals. In addition, OMB's June 2002 Circular A-11 states that the
strategic plan should briefly outline how annual performance goals
relate to the long-term, general goals, and how they help determine the
achievement of the general goals. Federal agencies can help readers
understand how they move from general goals to specific, measurable
outcomes by discussing how they plan to measure progress in achieving
the long term-goals in their strategic plan. For example, for its
strategic goal of "Increase Homeownership Opportunities," HUD mentions
that one of its long-term performance measures is to combat predatory
lending. Readers can review the intermediate measures listed under that
goal to get a sense of how HUD plans to accomplish this objective. For
example, HUD mentions that beginning in the third quarter of fiscal
year 2003, field offices will report all activities related to
predatory lending to headquarters each quarter. However, not all long-
term measures listed in the strategic plan have a corresponding
intermediate performance measure.
2. HUD disagreed with our observation that it did not explain in its
strategic plan how it used the results of program evaluations to update
the current plan and did not include a schedule for future evaluations.
As we have previously reported, program evaluations are individual,
systematic studies that use objective measurement and analysis to
answer specific questions about how well a program is working and,
thus, may take many forms. Where a program aims to produce changes that
result from program activities, outcome or effectiveness evaluations
assess the extent to which those results were achieved. Where complex
systems or events outside a program's control also influence its
outcomes, impact evaluations use scientific research methods to
establish the causal connection between outcomes and program activities
and isolate the program's contribution to those changes. A program
evaluation that also systematically examines how a program was
implemented can provide important information about why a program did
or did not succeed and suggest ways to improve it.[Footnote 143] In its
strategic plan, HUD provides a few examples of how it modified
performance measures as a result of program evaluations. However, we
found that 38 of the 41 performance measures discussed in the strategic
plan did not mention how, if at all, HUD revised and/or updated them as
the result of program evaluations. Elsewhere in the plan, HUD discussed
program evaluation activities carried out by its Office of Policy
Development and Research; however, a significant number of those
evaluations will take place in the future and there is no fixed
timetable for when HUD will issue reports on its findings.
3. HUD questioned an example we used to show that its strategic plan
did not always provide a clear picture of how it will be able to
measure progress toward its strategic goals. We chose this example
because HUD used the number of jobs created or retained to measure its
progress in achieving the results of the Community Development Block
Grant (CDBG) program. As HUD discusses in it strategic plan, there are
factors external to the CDBG program, such as broad macro-economic
trends and HUD's limited control over how grant recipients use the
funding, which can significantly affect job creation in a community.
Therefore, it is difficult to establish the contribution of the CDBG
program--apart from the other factors--to HUD's stated goal.
4. HUD also disagreed with our observation that in its annual
performance report it did not state the steps it would take to address
unmet performance goals. We recognize that in some instances HUD
mentioned how it would address unmet goals. GPRA requires that agencies
explain and describe, where a performance goal has not been met, why
the goal was not met, schedules for achieving the established
performance goal, and whether or not the performance goal is
impractical or unfeasible. However, our review of HUD's performance
report found that of the 93 unmet performance targets for fiscal year
2002, 74 lacked an explanation of how HUD would address them in fiscal
year 2003.
5. In commenting on our observation that HUD did not include an
evaluation of its fiscal year 2003 performance plan relative to the
performance attained by the department in fiscal year 2002, HUD
mentioned that we should consider the impact of the acceleration of
reporting deadlines on the department's ability to include an
evaluation of the new fiscal year's performance plan relative to the
performance attained in the just completed fiscal year and reasonable
alternative actions to fulfill this requirement. While we acknowledge
that changes in the reporting deadlines can create challenges for
federal agencies, these deadlines are governmentwide and not specific
to HUD. In our review of agency plans we found that some agencies, such
as DOT, were able to collect performance information for 95 percent of
their performance indicators and were able to predict future
performance, despite not having complete performance information and
facing the same deadlines. DOT provided an evaluation of whether or not
fiscal year 2003 performance targets would be met for each of its 40
performance goals based on fiscal year 2002 results. These evaluations
were included for the two performance goals for which data were
unavailable. For example, for the measure "Number of employment sites
(in the thousands) that are made accessible by Job Access and Reverse
Commute (JARC) transportation services," DOT could not characterize
performance since data had not yet been received from JARC grantees.
The 2002 performance report stated that a new easier to use reporting
system is being implemented that should improve data gathering
performance. The report further stated that DOT would meet this target
in fiscal year 2003.
6. HUD also disagreed with how we presented the performance information
in its summary report cards (see fig. 22). HUD noted that many of the
results were explained in the individual indicator write-ups that
followed the summary information. Our review of HUD's reports included,
among other things, qualitative aspects of how the information was
presented, such as its usefulness to inform the average reader with
little to no exposure on the subject matter, and the extent to which it
presented summarized performance information that was complete and
user-friendly. Our analysis of HUD's performance information was
largely based on a review of the information and terms used in the
performance report cards. We characterized some of HUD's performance
indicators as being "undetermined," given that HUD did not clearly
indicate whether or not a goal was achieved. Instead, HUD provided
footnotes, such as "results too complex to summarize." We also
characterized some performance targets as having "no explanation,"
given that information was missing from the report card to determine
whether HUD had reached its desired target. To develop the graphic
summarizing HUD's performance information, we compiled the results of
HUD's performance indicators across all five report cards contained in
the report.
[End of section]
Appendix XI: Comments from the Social Security Administration:
SOCIAL SECURITY:
The Commissioner:
January 16, 2004:
Ms. Patricia A. Dalton:
Director, Strategic Issues:
U.S. General Accounting Office Room 2440C:
441 G Street, NW
Washington, D.C. 20548:
Dear Ms. Dalton:
Thank you for the opportunity to review and comment on the draft report
"Review Results-Oriented Government: Government Performance Results
Act (GPRA) has Established a Solid Foundation for Achieving Greater
Results" (GAO-04-38). Our comments on the report contents and
suggestions are enclosed.
If you have any questions, please contact Candace Skurnik, Director,
Audit Management and Liaison Staff at (410) 965-4636. Staff questions
should be directed to Laura Bell at (410) 965-2636.
Sincerely,
Signed by:
Jo Anne B. Barnhart:
Enclosure:
SOCIAL SECURITY ADMINISTRATION BALTIMORE MD 21235-0001:
COMMENTS ON THE GENERAL ACCOUNTING OFFICE (GAO) DRAFT REPORT "REVIEW
RESULTS-ORIENTED GOVERNMENT: GOVERNMENT PERFORMANCE RESULTS ACT (GPRA)
HAS ESTABLISHED A SOLID FOUNDATION FOR ACHIEVING GREATER RESULTS" (GAO-
04-38):
Thank you for the opportunity to review and comment on the draft
report. I am proud to see that the Social Security Administration (SSA)
was once again the only Agency, out of the six reviewed, where GAO
found that our Strategic Plan (SP) contained all required elements
during both the 1997 and the 2002 review. The figure on page 28 clearly
demonstrates our commitment and continued progress in our GPRA planning
efforts.
It should be noted that our performance measures are more outcome
oriented and that we have made, and are continuing to make, progress in
integrating SSA's budget with performance. For example, we are able to
align costs and workyears with overarching performance goals in our SP
(e.g., eliminating backlogs) as part of the budget development process.
In summary, we agree for the most part with your conclusions and
appreciate the suggestions for improvement. We will incorporate them
where appropriate in our future planning efforts.
Regarding the specific references to our SP, Annual Performance Plan
(APP), and Performance and Accountability Report (PAR) please see my
comments below. In addition, I am providing some technical comments
that should be included in the final report.
Strategic Plan:
On page 47 the report suggests that we could further improve our SP
through:
2) Explicitly describing the effect of external factors on goal
attainment.
Response:
We believe that the relationship between external factors and our
goals, objectives and strategies are clear in most cases. Our SP
clearly identifies four categories of environmental factors impacting
our future operations: Demographics, Health and Disability Trends,
Technological Advances and Workforce Trends. Each strategic goal
includes an introduction and description of related issues that
explains the interrelationship between these factors and our strategic
priorities. For example, under Service we explain how changing
demographics (page 15), health and disability trends, including the
desire of disabled people to return to work (page 16), and improvements
in technology (page 18), impact our service
delivery strategies. The impact of environmental factors is equally
clear in the context of our Solvency and Staff goals. Our Stewardship
goal, driven by the need for the effective management of Agency
finances and assets, is the only goal not explicitly linked to
environmental factors.
3) Providing timetables or schedules for achieving results.
Response:
Generally speaking, unless specified, we expect to achieve our long-
term outcomes within the 5-year period covered by the SP. In some
instances, shorter time frames are specified for key outcomes. More
detailed plans and timetables for specific activities are featured in
our APP.
4) Providing details on how each performance and accountability
challenge will be addressed.
Response:
Presumably this is a reference to the Major Management Challenges (MMC)
identified by the GAO and our Office of Inspector General (OIG). We are
not aware of a specific requirement in the Office of Management and
Budget's (OMB) A-1 I instructions to address every one of these
challenges in our long-term SP's. While our current SP does address
some of the challenges that were identified at the time of its
publication, because the MMC's are updated every year, we believe that
they are more appropriately addressed in the context of our APP and our
annual PAR. Therefore, we have made more of an effort to address the
MMC beginning with our 2004 APP and our 2003 PAR.
6) Discussing the manner in which SSA coordinated with other agencies,
especially those that serve the same beneficiaries.
Response:
While we agree that such a discussion would be useful, we note that it
is not one of the OMB requirements for preparing a strategic plan. Our
SP does discuss in several places our interaction with other Government
agencies including our partnership with the State Disability
Determination Services, and collaboration with the Department of Labor
on Ticket to Work initiatives. Again, while we believe that this is not
an OMB requirement for preparing a strategic plan, we will make a
greater effort to provide details regarding such collaborations in
future SP's.
Annual Performance Plan:
This section highlights our improvement in identifying data sources and
definitions for each performance measure but observes that there is no
discussion of verification and validation procedures for data generated
by these systems.
Response:
We will continue to improve our verification and validation procedures
and ensure that there is adequate documentation in the APP. Currently
Appendix C (pages 69 to 70) of our APP outlines our validation efforts.
We maintain internal controls on our data gathering systems and make
significant efforts to validate those controls by use of a variety of
audits, both internal and external. Most notably,
PricewaterhouseCoopers (PwC) conducts an audit of our financial
statements each year. This audit includes a very thorough review of the
15 or so key performance measures highlighted in our annual PAR. We
also work closely with our OIG to validate the remaining performance
measures through ongoing reviews.
Annual Performance Report (Performance and Accountability Report-PAR):
Page 118 states that missing data and a lack of documentation of the
methods used to measure our performance reduced the overall quality of
the document.
Response:
As mentioned previously, in addition to our own internal controls, we
work with PwC and our OIG to validate our data sources. Our annual PAR
has received the Certificate of Excellence in Accountability Reporting
for five consecutive years, in large part because of this
collaboration. Beginning in FY 2003, we began placing greater emphasis
on documenting the processes and validation methods used in our data
gathering activities. We will do more to report on these efforts in
future PAR'S.
Also, on page 121 in the program evaluation section, GAO implies the
need for us to report on how our evaluations have helped answer
questions about program performance and results as opposed to merely
describing the surveys we conduct with those who do business with us or
based on our internal needs assessment surveys.
Response:
We disagree with this assessment. Many of the evaluations directed at
our efforts to deliver high quality, citizen-centered service rely on
these surveys. This assessment seems to minimize the importance of
these surveys; however, we feel that they are an extremely important
part of our service delivery efforts. Furthermore, the program
evaluation section of our report lists about a dozen:
other evaluations that are of great importance to our ongoing
operations. For example, our continuing disability and redetermination
reviews reduce program outlays by billions of dollars every year. In
the future, we will make more of an effort to explain how these and
other evaluations help improve program performance.
Technical Comments:
References on pages 13 and 83 state that we did not structure our
performance and strategic plans by program activity or account. The
references on page 83 then go on to say that "SSA noted that it aligned
its strategic goals, performance measures, and budget with its major
functional responsibilities rather than by program accounts since
direct service and support employees provide services linked to these
functional responsibilities, as opposed to a specific program. However,
SSA does not indicate what it means by `functional responsibilities,'
nor does it show a clear linkage between its strategic goals and such
responsibilities." We believe that we do attempt to describe our
functional responsibilities in Parts I and II of our FY 2004 APP. We
will take under advisement GAO's concern and consider including clearer
descriptions in future APPs.
On page 21 we believe that footnote 10 should be included as part of
the narrative in the first paragraph and that the sentence be modified
as follows: "According to the Inspector General, SSA began to implement
an improved cost accounting system in FY 2002, which was to be phased
in over the next 3 to 4 years." The current language makes it appear
that we did not have a cost accounting system prior to 2002.
The following are our comments on SSA's letter dated January 16, 2004.
GAO Comments:
In general, SSA agreed with our conclusions. SSA also agreed to
incorporate the suggestions for improvement in its future planning
efforts. SSA made several points of clarification and disagreed with
our assessment in one area.
1. In our draft report, we noted that SSA did not explicitly link
external factors that may affect its programs to its general goals and
state how these factors could affect goal attainment. SSA attests that
its four categories of environmental factors are discussed under each
of the strategic goals, as appropriate, and the relationship between
these factors and SSA's strategic priorities is described. This general
discussion of the environmental factors is useful in understanding the
challenges SSA faces in working toward its broad strategic goals.
However, SSA provides little or no discussion of these challenges in
its discussion of the agency's performance goals. Thus, the range of
challenges facing the agency in meeting each of its performance goals
is not fully explained.
2. In our draft report, we noted that SSA does not provide timetables
or schedules for achieving all the results in its strategic plan. SSA
noted that it expects to achieve its long-term outcomes within the 5-
year period covered by the strategic plan; in selected instances,
shorter time frames are specified. SSA noted that more detailed plans
and timetables are featured in its annual performance plan. GPRA
requires agencies to furnish a schedule of significant actions in their
strategic plans; however, SSA does not clearly articulate its
timetables and schedules for achieving each of its long-term outcomes
in its strategic plan.
3. We noted that SSA's strategic plan could be improved by providing
details on how each performance and accountability challenge will be
addressed. SSA asserted that the strategic plan addresses some of the
challenges, but because the challenges are updated every year, they are
more appropriately addressed in the annual performance plan and
performance and accountability report. As noted in our discussion of
the criteria used to analyze agencies' strategic plans, it is
particularly important that agencies develop strategies that address
management challenges that threaten their ability to meet long-term
strategic goals, as one of the purposes of GPRA is to improve the
management of federal agencies.
4. In our draft report, we observed that SSA's discussion of its
interactions with other agencies, especially those that serve the same
beneficiaries, was limited. SSA noted that such a discussion would be
useful, but is not an OMB requirement. While we agree that this is not
an OMB requirement, we have reported that given scarce resources and
competing priorities, it would be useful to identify agency efforts to
maximize its effect through cooperation and coordination across the
federal government. Better strategic plans not only identify the need
to coordinate with other agencies, but also discuss how agencies intend
to coordinate common or complementary goals and strategies with other
agencies.
5. With regard to SSA's performance and accountability report, we noted
that SSA did not clearly state how program evaluations were used to
answer questions about program performance and results and how those
results can be improved. SSA disagreed with our observation, stating
that many of its evaluations rely on surveys, and these surveys form
the basis for its efforts to deliver high-quality service. SSA also
noted that it listed other evaluations that are of great importance to
its ongoing operations. We do not discount the usefulness of SSA's
surveys in assessing its day-to-day management of programs. Rather, as
we noted in the report, it would be helpful for SSA to clearly identify
the range of evaluations conducted and how each of them contributed to
improved program performance. For example, we recently recommended that
SSA evaluate a new initiative to improve the integrity of Social
Security number issuance to noncitizens; the description of such an
evaluation would be helpful for SSA to determine how it can be best
positioned to ensure the integrity of its enumeration process.[Footnote
144] Additionally, our September 2000 report on program evaluation
states that GPRA recognizes the complementary nature of program
evaluation and performance measurement. Strategic plans are to describe
the program evaluations that were used in establishing and revising
goals and to include a schedule for future program evaluations.
Agencies are to summarize the findings of program evaluations in their
annual performance reports.
Additionally, SSA made technical comments that we incorporated into the
report, as appropriate.
[End of section]
Appendix XII: GAO Contact and Staff Acknowledgments:
GAO Contact:
Patricia Dalton, (202) 512-6806:
Acknowledgments:
In addition to the persons mentioned above, Thomas Beall, Daniel
Bertoni, Kay Brown, Joyce Corry, Elizabeth Curda, David Dornisch,
William Fenzel, Kimberly Gianopoulos, Evan Gilman, Katie Harris, Susan
Higgins, Benjamin Licht, William McKelligott, James Noel, Carol
Petersen, Carolyn Samuels, Teresa Spisak, Daren Sweeney, Carolyn
Taylor, Michael Volpe, Lynn Wasielewski, and Steven Westley made key
contributions to this report.
[End of section]
Related GAO Products:
GPRA/Managing for Results:
Results-Oriented Government: Using GPRA to Address 21st Century
Challenges,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-1166T];
(Washington, D.C. Sept. 18, 2003).
Results-Oriented Management: Agency Crosscutting Actions and Plans in
Border Control, Flood Mitigation and Insurance, Wetlands, and Wildland
Fire Management,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-321];
(Washington, D.C. Dec. 20, 2002).
Results-Oriented Management: Agency Crosscutting Actions and Plans in
Drug Control, Family Poverty, Financial Institution Regulation, and
Public Health Systems,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-320];
(Washington, D.C. Dec. 20, 2002).
Performance and Accountability: Reported Agency Actions and Plans to
Address 2001 Management Challenges and Program Risks,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-225];
(Washington, D.C. Oct. 31, 2002).
Managing for Results: Next Steps to Improve the Federal Government's
Management and Performance,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-439T];
(Washington, D.C. Feb. 15, 2002).
Managing for Results: Federal Managers' Views on Key Management Issues
Vary Widely Across Agencies,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-01-592];
(Washington, D.C. May 25, 2001).
Managing for Results: Federal Managers' Views Show Need for Ensuring
Top Leadership Skills,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-01-127];
(Washington, D.C. Oct. 20, 2000).
Managing for Results: Barriers to Interagency Coordination,
Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-106];
(Washington, D.C. Mar. 29, 2000).
Management Reform: Elements of Successful Improvement Initiatives,
Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-00-26];
(Washington, D.C. Oct. 15, 1999).
Management Reform: Using the Results Act and Quality Management to
Improve Federal Performance,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-99-151];
(Washington, D.C. July 29, 1999).
Managing for Results: Opportunities for Continued Improvements in
Agencies' Performance Plans,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-99-215];
(Washington, D.C. July 20, 1999).
Agency Performance Plans: Examples of Practices That Can Improve
Usefulness to Decisionmakers,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-99-69];
(Washington, D.C. Feb. 26, 1999).
The Results Act: Assessment of the Governmentwide Performance Plan for
Fiscal Year 1999,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD/GGD-98-159];
(Washington, D.C. Sept. 8, 1998).
Managing for Results: An Agenda to Improve the Usefulness of Agencies'
Annual Performance Plans,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-98-228];
(Washington, D.C. Sept. 8, 1998).
The Results Act: An Evaluator's Guide to Assessing Agency Annual
Performance Plans.
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-10.1.20];
(Washington, D.C. April 1998).
Agencies' Annual Performance Plans Under the Results Act: An Assessment
Guide to Facilitate Congressional Decisionmaking,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD/AIMD-10.1.18];
(Washington, D.C. February 1998).
Managing for Results: Agencies' Annual Performance Plans Can Help
Address Strategic Planning Challenges,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-44];
(Washington, D.C. Jan. 30, 1998).
Managing for Results: Building on Agencies' Strategic Plans to Improve
Federal Management,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/AIMD-98-29];
(Washington, D.C. Oct. 30, 1997).
Managing for Results: Critical Issues for Improving Federal Agencies'
Strategic Plans,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-180];
(Washington, D.C. Sept. 16, 1997).
Managing for Results: Using the Results Act to Address Mission
Fragmentation and Program Overlap,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-97-146];
(Washington, D.C. Aug. 29, 1997).
Managing for Results: The Statutory Framework for Improving Federal
Management and Effectiveness,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/AIMD-97-144];
(Washington, D.C. June 24, 1997).
The Government Performance and Results Act: 1997 Governmentwide
Implementation Will Be Uneven,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-109];
(Washington, D.C. June 2, 1997).
Agencies' Strategic Plans Under GPRA: Key Questions to Facilitate
Congressional Review,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-10.1.16];
(Washington, D.C. May 1997).
GPRA: Managerial Accountability and Flexibility Pilot Did Not Work as
Intended,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-36];
(Washington, D.C. Apr. 10, 1997).
Managing for Results: Enhancing the Usefulness of GPRA Consultations
Between the Executive Branch and Congress,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-97-56];
(Washington, D.C. Mar. 10, 1997).
Managing for Results: Key Steps and Challenges in Implementing GPRA in
Science Agencies,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD/RCED-96-214];
(Washington, D.C. July 10, 1996).
Executive Guide: Effectively Implementing the Government Performance
and Results Act,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-96-118];
(Washington, D.C. June 1996).
Strategic Human Capital Management:
Human Capital: A Guide for Assessing Strategic Training and Development
Efforts in the Federal Government; (Exposure Draft),
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-893G];
(Washington, D.C. July 2003).
Results-Oriented Cultures: Creating a Clear Linkage between Individual
Performance and Organizational Success,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-488];
(Washington, D.C. Mar. 14, 2003).
Results-Oriented Cultures: Insights for U.S. Agencies from Other
Countries' Performance Management Initiatives,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-862];
(Washington, D.C. Aug. 2, 2002).
Human Capital: Key Principles From Nine Private Sector Organizations,
Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-28];
(Washington, D.C. Jan. 31, 2000).
Linking Resources to Results:
Performance Budgeting: Observations on the Use of OMB's Program
Assessment Rating Tool for the Fiscal Year 2004 Budget,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-174];
(Washington, D.C. Jan. 30, 2004).
Managing for Results: Efforts to Strengthen the Link Between Resources
and Results at the Nuclear Regulatory Commission,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-258];
(Washington, D.C. Dec. 10, 2002).
Managing for Results: Efforts to Strengthen the Link Between Resources
and Results at the Administration for Children and Families,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-9];
(Washington, D.C. Dec. 10, 2002).
Managing for Results: Efforts to Strengthen the Link Between Resources
and Results at the Veterans Health Administration,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-10];
(Washington, D.C. Dec. 10, 2002).
Managing for Results: Agency Progress in Linking Performance Plans With
Budgets and Financial Statements,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-236];
(Washington, D.C. Jan. 4, 2002).
Performance Budgeting: Fiscal Year 2000 Progress in Linking Plans With
Budgets,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-99-239R];
(Washington, D.C. July 30, 1999).
Performance Budgeting: Initial Experiences Under the Results Act in
Linking Plans With Budgets,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD/GGD-99-67];
(Washington, D.C. Apr. 12, 1999).
Performance Budgeting: Past Initiatives Offer Insights for GPRA
Implementation,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/AIMD-97-46];
(Washington, D.C. Mar. 27, 1997).
Measuring Performance:
Program Evaluation: An Evaluation Culture and Collaborative
Partnerships Help Build Agency Capacity,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-454];
(Washington, D.C. May 2, 2003).
Managing for Results: Next Steps to Improve the Federal Government's
Management and Performance,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-439T];
(Washington, D.C. Feb. 15, 2002).
Program Evaluation: Studies Helped Agencies Measure or Explain Program
Performance,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-204];
(Washington, D.C. Sept. 29, 2000).
Managing for Results: Strengthening Regulatory Agencies' Performance
Management Practices,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-10];
(Washington, D.C. Oct. 28, 1999).
Managing for Results: Measuring Program Results That Are Under Limited
Federal Control,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-99-16];
(Washington, D.C. Dec. 11, 1998).
Grant Programs: Design Features Shape Flexibility, Accountability, and
Performance Information,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-137];
(Washington, D.C. June 22, 1998).
Program Evaluation: Agencies Challenged by New Demand for Information
on Program Results,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-98-53];
(Washington, D.C. Apr. 24, 1998).
Managing for Results: Regulatory Agencies Identified Significant
Barriers to Focusing on Results,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-97-83];
(Washington, D.C. June 24, 1997).
Managing for Results: Analytic Challenges in Measuring Performance,
Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/HEHS/GGD-97-138];
(Washington, D.C. May 30, 1997).
Data Credibility:
Performance Reporting: Few Agencies Reported on the Completeness and
Reliability of Performance Data,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-02-372];
(Washington, D.C. Apr. 26, 2002).
Managing for Results: Assessing the Quality of Program Performance
Data,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-140R];
(Washington, D.C. May 25, 2000).
Managing for Results: Challenges Agencies Face in Producing Credible
Performance Information,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-52];
(Washington, D.C. Feb. 4, 2000).
Performance Plans: Selected Approaches for Verification and Validation
of Agency Performance Information,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-99-139];
(Washington, D.C. July 30, 1999).
Using Performance Information:
Managing for Results: Views on Ensuring the Usefulness of Agency
Performance Information to Congress,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/GGD-00-35];
(Washington, D.C. Jan. 26, 2000).
Managing for Results: Using GPRA to Assist Congressional and Executive
Branch Decisionmaking,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-97-43];
(Washington, D.C. Feb. 12, 1997).
Managing for Results: Achieving GPRA's Objectives Requires Strong
Congressional Role,
[Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO/T-GGD-96-79];
(Washington, D.C. Mar. 6, 1996).
(450181):
FOOTNOTES
[1] Pub. L. No. 103-62.
[2] U.S. General Accounting Office, Managing for Results: Opportunities
for Continued Improvements in Agencies' Performance Plans, GAO/GGD/
AIMD-99-215 (Washington, D.C.: July 20, 1999).
[3] PART is a diagnostic tool developed by OMB that it has been using
to rate the effectiveness of federal programs with a particular focus
on program results. OMB's goal is to review all federal programs over a
5-year period using the PART tool. OMB used the tool to review
approximately 400 programs between the fiscal year 2004 budget cycle
and the fiscal year 2005 budget cycle--234 programs were assessed last
year and 173 were assessed this year. Some reassessed programs were
combined for review for the 2005 budget, which is why the number of
programs assessed over the 2 years does not add up to exactly 400
programs.
[4] U.S. General Accounting Office, Program Performance Measures:
Federal Agency Collection and Use of Performance Data, GAO/GGD-92-65
(Washington, D.C.: May 4, 1992).
[5] Types of measures were defined in the questionnaire as follows:
performance measures that tell us how many things we produce or
services we provide (output measures); performance measures that tell
us if we are operating efficiently (efficiency measures); performance
measures that tell us whether or not we are satisfying our customers
(customer service measures); performance measures that tell us about
the quality of the products or services we provide (quality measures);
and performance measures that would demonstrate to someone outside of
our agency whether or not we are achieving our intended results
(outcome measures).
[6] U.S. General Accounting Office, Managing for Results: An Agenda to
Improve the Usefulness of Agencies' Annual Performance Plans, GAO/GGD/
AIMD-98-228 (Washington, D.C.: Sept. 8, 1998).
[7] U.S. General Accounting Office, Managing for Results: Critical
Issues for Improving Agencies' Strategic Plans, GAO/GGD-97-180
(Washington, D.C.: Sept. 16, 1997).
[8] GAO/GGD/AIMD-98-228.
[9] Pub. L. No. 103-62.
[10] The guidance on the preparation of strategic plans, annual
performance plans, and program performance reports is contained in OMB
Circular A-11, Part 6.
[11] OMB Circular A-11, July 2003.
[12] PART is a diagnostic tool developed by OMB that it has been using
to rate the effectiveness of federal programs with a particular focus
on program results. OMB's goal is to review all federal programs over a
5-year period using the PART tool. OMB used the tool to review
approximately 400 programs between the fiscal year 2004 budget cycle
and the fiscal year 2005 budget cycle--234 programs were assessed last
year and 173 were assessed this year. Some reassessed programs were
combined for review for the 2005 budget, which is why the number of
programs assessed over the 2 years does not add up to exactly 400
programs.
[13] U.S. General Accounting Office, GPRA: Managerial Accountability
and Flexibility Pilot Did Not Work as Intended, GAO/GGD-97-36
(Washington, D.C.: Apr. 10, 1997).
[14] U.S. General Accounting Office, Managing for Results: Agency
Progress in Linking Performance Plans With Budgets and Financial
Statements, GAO-02-236 (Washington, D.C.: Jan. 4, 2002).
[15] GAO/GGD/AIMD-99-215. Based on how we had rated agencies' annual
performance plans on their picture of performance, specificity of
strategies and resources, and the degree of confidence that performance
information will be credible, we assigned numeric values to each
agencies' rating (e.g., clear=3, general=2, limited=1, unclear=0) and
added them up to determine overall quality of high, medium, or low. An
agency's plan was considered high quality if its score was between 7-9,
a score of 5-6 was considered medium quality, and a score of 3-4 was
low. No agencies received a score lower than 3.
[16] GAO/GGD-92-65.
[17] U.S. General Accounting Office, Performance Budgeting:
Observations on the Use of OMB's Program Assessment Rating Tool for the
Fiscal Year 2004 Budget, GAO-04-174 (Washington, D.C.: Jan. 30, 2004).
[18] Types of measures were defined in the questionnaire as follows:
performance measures that tell us how many things we produce or
services we provide (output measures); performance measures that tell
us if we are operating efficiently (efficiency measures); performance
measures that tell us whether or not we are satisfying our customers
(customer service measures); performance measures that tell us about
the quality of the products or services we provide (quality measures);
and performance measures that would demonstrate to someone outside of
our agency whether or not we are achieving our intended results
(outcome measures).
[19] GAO/GGD/AIMD-98-228.
[20] GAO-02-236; U.S. General Accounting Office, Performance Budgeting:
Initial Experiences Under the Results Act in Linking Plans With
Budgets, GAO/AIMD/GGD-99-67 (Washington, D.C.: Apr. 12, 1999); and
Performance Budgeting: Fiscal Year 2000 Progress in Linking Plans With
Budgets, GAO/AIMD-99-239R (Washington, D.C.: July 30, 1999).
[21] U.S. General Accounting Office, Managing for Results: Efforts to
Strengthen the Link Between Resources and Results at the Veterans
Health Administration, GAO-03-10 (Washington, D.C.: Dec. 10, 2002).
[22] U.S. General Accounting Office, Managing for Results: Efforts to
Strengthen the Link Between Resources and Results at the Administration
for Children and Families, GAO-03-09 (Washington, D.C.: Dec. 10, 2002).
[23] U.S. General Accounting Office, Managing for Results: Efforts to
Strengthen the Link Between Resources and Results at the Nuclear
Regulatory Commission, GAO-03-258 (Washington, D.C.: Dec. 10, 2002).
[24] GAO/GGD-97-180.
[25] U.S. General Accounting Office, Managing for Results: Challenges
Agencies Face in Producing Credible Performance Information, GAO/GGD-
00-52 (Washington, D.C.: Feb. 4, 2000).
[26] Since 1990, GAO has periodically reported on government operations
that it identifies as "high risk" because of the greater
vulnerabilities to fraud, waste, abuse, and mismanagement. See U.S.
General Accounting Office, High-Risk Series: An Update, GAO-03-119
(Washington, D.C.: January 2003).
[27] U.S. General Accounting Office, Managing for Results: Using the
Results Act to Address Mission Fragmentation and Program Overlap, GAO/
AIMD-97-146 (Washington, D.C.: Aug. 29, 1997).
[28] GAO/GGD/AIMD-98-228.
[29] Program activity refers to the list of projects and activities in
the appendix portion of the Budget of the United States Government.
Program activity structures are intended to provide a meaningful
representation of the operations financed by a specific budget account.
[30] GAO/GGD-97-36.
[31] U.S. General Accounting Office, Major Management Challenges and
Program Risks: Department of Energy, GAO-03-100 (Washington, D.C.:
January 2003).
[32] According to OMB's Statement of Federal Financial Accounting
Standards No. 4--Managerial Cost Accounting Standards, July 31, 1995, a
responsibility segment is a component of a reporting entity that is
responsible for carrying out a mission, conducting a major line of
activity, or producing one or a group of related products or services.
In addition, responsibility segments usually possess the following
characteristics: (1) their managers report to the entity's top
management directly and (2) their resources and results of operations
can be clearly distinguished from those of other segments of the
entity. Managerial cost accounting should be performed to measure and
report the costs of each segment's outputs.
[33] SSA noted that its fifth strategic goal, "Valued Employees,"
supports the accomplishment of all its basic functions, so its
resources are inherently included in the other four goals.
[34] According to the IG, SSA began to implement an improved cost
accounting system in fiscal year 2002, which was to be phased in over
the next 3 to 4 years.
[35] U.S. General Accounting Office, Performance Reporting: Few
Agencies Reported on the Completeness and Reliability of Performance
Data, GAO-02-372 (Washington, D.C.: Apr. 26, 2002).
[36] U.S. General Accounting Office, Management Reform: Elements of
Successful Improvement Initiatives, GAO/T-GGD-00-26 (Washington, D.C.:
Oct. 15, 1999).
[37] U.S. General Accounting Office, The Government Performance and
Results Act: 1997 Governmentwide Implementation Will Be Uneven, GAO/
GGD-97-109 (Washington, D.C.: June 2, 1997) and Managing for Results:
Federal Managers' Views Show Need for Ensuring Top Leadership Skills,
GAO-01-127 (Washington, D.C.: Oct. 20, 2000).
[38] GAO/GGD-97-109 and GAO-01-127.
[39] GAO-01-127.
[40] U.S. General Accounting Office, Human Capital: A Guide for
Assessing Strategic Training and Development Efforts in the Federal
Government (Exposure Draft) GAO-03-893G (Washington, D.C.: July 1,
2003).
[41] For a complete list and discussion of the practices, see U.S.
General Accounting Office, Results oriented Cultures: Creating a Clear
Linkage between Individual Performance and Organizational Success, GAO-
03-488 (Washington, D.C.: Mar. 14, 2003).
[42] See for example, U.S. General Accounting Office, Managing for
Results: Analytic Challenges in Measuring Performance, GAO/HEHS/GGD-
97-138 (Washington, D.C.: May 30, 1997); Program Evaluation: Agencies
Challenged by New Demand for Information on Program Results, GAO/GGD-
98-53 (Washington, D.C.: Apr. 24, 1998); Managing for Results:
Measuring Program Results That Are Under Limited Federal Control, GAO/
GGD-99-16 (Washington, D.C.: Dec. 11, 1998); and Managing for Results:
Challenges Agencies Face in Producing Credible Performance Information,
GAO/GGD-00-52 (Washington, D.C.: Feb. 4, 2000).
[43] GAO/GGD-97-109, 6.
[44] U.S. General Accounting Office, Managing for Results: Key Steps
and Challenges in Implementing GPRA in Science Agencies, GAO/T-GGD/
RCED-96-214 (Washington, D.C.: July 10, 1996).
[45] U.S. General Accounting Office, Managing for Results:
Strengthening Regulatory Agencies' Performance Management Practices,
GAO/GGD-00-10 (Washington, D.C.: Oct. 28, 1999).
[46] U.S. General Accounting Office, Environmental Protection:
Assessing the Impacts of EPA's Regulations Through Retrospective
Studies, GAO/RCED-99-250 (Washington, D.C.: Sept. 14, 1999).
[47] U.S. General Accounting Office, Grant Programs: Design Features
Shape Flexibility, Accountability, and Performance Information, GAO/
GGD-98-137 (Washington, D.C.: June 22, 1998).
[48] GAO/GGD-00-52.
[49] U.S. General Accounting Office, Managing for Results: Barriers to
Interagency Coordination, GAO/GGD-00-106 (Washington, D.C.: Mar. 29,
2000).
[50] U.S. General Accounting Office, Wetlands Overview: Problems With
Acreage Data Persist, GAO/RCED-98-150 (Washington, D.C.: July 1, 1998)
and Results-Oriented Management: Agency Crosscutting Actions and Plans
in Border Control, Flood Mitigation and Insurance, Wetlands, and
Wildland Fire Management, GAO-03-321 (Washington, D.C.: Dec. 20, 2002).
[51] GAO/GGD-00-106 and U.S. General Accounting Office, Results-
Oriented Government: Using GPRA to Address 21st Century Challenges,
GAO-03-1166T (Washington, D.C.: Sept. 18, 2003).
[52] U.S. General Accounting Office, The Results Act: Assessment of the
Governmentwide Performance Plan for Fiscal Year 1999, GAO/AIMD/GGD-98-
159 (Washington, D.C.: Sept. 8, 1998).
[53] Congressional Research Service, Government Performance and Results
Act: Overview of Associated Provisions in the 106th Congress,
(Washington, D.C.: 2002).
[54] This included reports that accompanied bills passed by both the
House and Senate that were either enacted into law or vetoed by the
President.
[55] U.S. General Accounting Office, Managing for Results: Views on
Ensuring the Usefulness of Agency Performance Information to Congress,
GGD-00-35 (Washington, D.C.: Jan. 26, 2000).
[56] U.S. General Accounting Office, Human Capital: Key Principles From
Nine Private Sector Organizations, GAO/GGD-00-28 (Washington, D.C.:
Jan. 31, 2000).
[57] GAO/AIMD-97-146.
[58] U.S. General Accounting Office, Executive Guide: Effectively
Implementing the Government Performance and Results Act, GAO/GGD-96-118
(Washington, D.C.: June 1, 1996).
[59] GAO/GGD-00-35.
[60] U.S. General Accounting Office, Program Evaluation: Agencies
Challenged by New Demand for Information on Program Results, GAO/GGD-
98-53 (Washington, D.C.: Apr. 24, 1998).
[61] U.S. General Accounting Office, Program Evaluation: Studies Helped
Agencies Measure or Explain Program Performance, GAO/GGD-00-204
(Washington, D.C.: Sept. 28, 2000).
[62] GAO/GGD-00-52.
[63] For information on the design and administration of the two
earlier surveys, see GAO/GGD-97-109, GAO-01-127, and U.S. General
Accounting Office, Managing For Results: Federal Managers' Views on Key
Management Issues Vary Widely Across Agencies, GAO-01-592 (Washington,
D.C.: May 25, 2001).
[64] For example, candidates could be operations managers with hands-on
experience managing a federal program or agency officials directly
involved in carrying out the activities required under GPRA, such as
developing a strategic or annual performance plan or annual performance
report.
[65] Due to last minute circumstances, a federal manager participated
via teleconference from an agency's field office and another was unable
to attend the focus group, but mailed his answers to questions we sent
to all participants in advance of the focus groups.
[66] We read the list of comments to the manager who participated via
teleconference and voted on his behalf based on his preferences.
[67] GAO/GGD/AIMD-99-215. Based on how we had rated agencies' annual
performance plans on their picture of performance, specificity of
strategies and resources, and the degree of confidence that performance
information will be credible, we assigned numeric values to each
agencies' rating (e.g., clear=3, general=2, limited=1, unclear=0) and
added them up to determine overall quality of high, medium, or low. An
agency's plan was considered high quality if its score was between 7-9,
a score of 5-6 was considered medium quality, and a score of 3-4 was
low. No agencies received a score lower than 3.
[68] Government Performance and Results Act of 1993, Committee on
Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong.
1st Sess. (1993).
[69] Office of Management and Budget, Circular No. A-11, Part 6,
Preparation and Submission of Strategic Plans, Annual Performance
Plans, and Annual Program Performance Reports (Washington, D.C.: June
2002).
[70] U.S. General Accounting Office, Agencies' Strategic Plans Under
GPRA: Key Questions to Facilitate Congressional Review, GAO/GGD-10.1.16
(Washington, D.C.: May 1, 1997).
[71] GAO/HEHS-97-176R; GAO/RCED-97-199R; GAO/RCED-97-208R; GAO/HEHS-
97-179R; GAO/RCED-97-205R; and GAO/RCED-97-224R.
[72] GAO/HEHS-97-176R.
[73] U.S. Department of Education, Office of the Deputy Secretary,
Strategic Accountability Service, U.S. Department of Education FY 2004
Annual Plan (Washington, D.C.: March 2003). This document represents
Education's performance plan for GPRA and will be referred to
henceforth as the "annual plan."
[74] U.S. General Accounting Office, Performance and Accountability
Series: Major Management Challenges and Program Risks, Department of
Education, GAO-01-245 (Washington, D.C.: January 2001). The January
2001 assessment was the last time we assessed the Department of
Education under our Performance and Accountability Series before the
release of the agency's 2002-2007 Strategic Plan. We further reported
in October 2002 on how Education and other agencies reported responding
to their management challenges and program risks. (U.S. General
Accounting Office, Performance and Accountability: Reported Agency
Actions and Plans to Address 2001 Management Challenges and Program
Risks, GAO-03-225 (Washington, D.C.: Oct. 31, 2002).
[75] Within one of its strategies, the plan states that "managers will
be given the freedom to manage and will be held accountable for
results."
[76] Pub. L. No. 107-110, January 8, 2002. The No Child Left Behind Act
of 2001 is a reauthorization of the Elementary and Secondary Education
Act, one of the major pieces of authorizing legislation for the
Department of Education.
[77] GAO/RCED-97-199R.
[78] GAO/RCED-97-224R.
[79] U.S. General Accounting Office, Major Management Challenges and
Program Risks: Department of Housing and Urban Development, GAO-03-103
(Washington, D.C.: January 2003).
[80] GAO/RCED-97-205R.
[81] At the time of our review, the most recent SBA strategic plan was
for fiscal years 2001-2008. SBA released a new strategic plan for
fiscal years 2003-2008 in October 2003.
[82] U.S. General Accounting Office, Major Management Challenges and
Program Risks: Small Business Administration, GAO-01-260 (Washington,
D.C.: January 2001).
[83] In the 2003-2008 plan, the word "advance" replaced the word
"promote."
[84] U.S. General Accounting Office, Major Management Challenges and
Program Risks: Social Security Administration, GAO-03-117 (Washington,
D.C.: January 2003).
[85] The suspense file contains information on earnings that cannot be
matched to an individual's record due to an invalid name/Social
Security number combination.
[86] GAO/RCED-97-208R.
[87] At the time of our review, the Department of Transportation was in
the process of revising its strategic plan. A draft copy of the updated
strategic plan, dated July 1, 2003, was used for this review.
[88] U.S. General Accounting Office, Major Management Challenges and
Program Risks: Department of Transportation, GAO-03-108 (Washington,
D.C.: January 2003).
[89] Government Performance and Results Act of 1993, Committee on
Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong.
1st Sess. (1993).
[90] OMB Circular No. A-11: Part 6, Preparation and Submission of
Strategic Plans, Annual Performance Plans, and Annual Program
Performance Reports (Washington, D.C.: June 2002).
[91] GAO/GGD/AIMD-99-215 and The Results Act: An Evaluator's Guide to
Assessing Agency Annual Performance Plans, GAO/GGD-10.1.20
(Washington, D.C.: April 1998).
[92] GAO/HEHS-98-172R, GAO/RCED-98-194R, GAO/RCED-98-159R, GAO/RCED-
98-180R, GAO/RCED-98-200R, and GAO/HEHS-98-178R.
[93] The Homeland Security Act (Pub. L. No. 107-296), also requires
that agencies provide a description of how the performance goals and
objectives are to be achieved, including the operations, processes,
training, skills and technology, and the human capital, information,
and other resources and strategies required to meet those performance
goals and objectives.
[94] GAO/HEHS-98-172R, GAO/RCED-98-194R, GAO/RCED-98-159R, GAO/RCED-
98-180R, GAO/RCED-98-200R, and GAO/HEHS-98-178R.
[95] Education's plan states that its fiscal year 2004 annual plan
includes both department-level measures and program performance plans.
These are organized into two volumes: the Annual Plan Fiscal Year 2004
U.S. Department of Education includes the department-level measures and
the FY 2004 Program Performance Plan: U.S. Department of Education
includes the program performance plans with their individual program
measures. These volumes are presented in a slightly different
electronic format for the public and other parties in general, which is
available at Education's Web site. The two volumes will henceforth be
referred to as Education's 2004 annual plan, or, where applicable,
volume 1 and volume 2.
[96] Education's 2004 annual plan represents its annual performance
goals as "targets." According to GPRA, the definition for "performance
goal" is "a target level of performance."
[97] In this report, we refer to the multiyear, long-term objectives in
Education's annual plan as "long-term goals." The strategic goals
included in the plan represent overarching statements of aim or purpose
that are used to group Education's long-term goals.
[98] GAO/HEHS-97-176R.
[99] U.S. General Accounting Office, Performance and Accountability
Series--Major Management Challenges and Program Risks: A Governmentwide
Perspective, GAO-01-241 (Washington, D.C.: January 2001) and GAO-01-
245.
[100] GAO-03-225.
[101] GAO-03-100.
[102] GAO/RCED-98-159R.
[103] GAO/RCED-98-200R.
[104] Much of this improvement took place between the 1999 and 2000
plans. We reported that SSA's fiscal year 2000 performance plan showed
significant improvement over its 1999 plan in U.S. General Accounting
Office, Observations on the Social Security Administration's Fiscal
Year 2000 Performance Plan, GAO/HEHS-99-162R (Washington, D.C.: July
20, 1999).
[105] SSA reduced its strategic goals from five to four in its 2003-
2008 strategic plan.
[106] GAO-03-225.
[107] SSA refers to external factors as environmental factors.
[108] GAO/RCED-98-180R.
[109] Bureau of Transportation Statistics, Source & Accuracy
Compendium, http://www.bts.gov/statpol/SAcompendium.html (Washington,
D.C.: Aug. 15, 2003).
[110] Office of Management and Budget, Memorandum: Program Assessment
Rating Tool (PART) - Presentation in Congressional Justifications, M-
03-06 (Washington, D.C.: 2003).
[111] Government Performance and Results Act of 1993, Committee on
Governmental Affairs, United States Senate, S. Rpt. No. 58, 103d Cong.
1st Sess. (1993).
[112] OMB Circular No. A-11, Part 6, Preparation and Submission of
Strategic Plans, Annual Performance Plans, and Annual Program
Performance Reports (Washington, D.C.: June 2002).
[113] Chief Financial Officers Act of 1990 (Pub. L. No. 101-576).
[114] GAO-02-372 and Executive Guide: Creating Value Through World-
class Financial Management, GAO/AIMD-00-134 (Washington, D.C.: Apr. 1,
2000).
[115] The Homeland Security Act (Pub. L. No. 107-296) requires agencies
to include a review of the performance goals and evaluation of the
performance plan relative to the agency's strategic human capital
management.
[116] Association of Government Accountants, Certificate of Excellence
in Accountability Reporting: Reviewers Checklist, Fiscal Year 2001.
(Washington, D.C.).
[117] GAO-02-372.
[118] GAO/HEHS-00-128R, GAO/RCED-00-209R, GAO/RCED-00-211R, GAO/RCED-
00-207R, GAO/HEHS-00-126R, and GAO/RCED-00-201R.
[119] Education's annual performance goals are represented by its
targets.
[120] GAO/HEHS-00-128R.
[121] Pub. L. No. 107-110, January 8, 2002. The No Child Left Behind
Act of 2001 is a reauthorization of the Elementary and Secondary
Education Act, one of the major pieces of authorizing legislation for
Education.
[122] GAO/HEHS-00-128R.
[123] As amended by the Homeland Security Act of 2002.
[124] GAO-03-225.
[125] U.S. Department of Housing and Urban Development, Fiscal year
2002 Performance and Accountability Report (Washington, D.C.: 2003).
[126] GAO/RCED-00-211R.
[127] GAO-03-225.
[128] Small Business Administration, SBA's Performance & Accountability
Report for Fiscal Year 2002 (Washington, D.C.: 2003).
[129] Our review of SBA's fiscal year 1999 report, GAO/RCED-00-207R,
focused on our observations on only three of SBA's key outcomes, as
well as the major management challenges addressed in the performance
report. Since our review of SBA's 2002 performance report used somewhat
different assessment criteria, we could not make valid comparisons on
all aspects of the reports.
[130] Social Security Administration, Performance and Accountability
Report, Fiscal Year 2002 (Washington, D.C.: 2002).
[131] Our review of SSA's fiscal year 1999 performance report (GAO/
HEHS-00-126R) focused on our observations on five of the agency's key
outcomes, as well as the major management challenges addressed in the
performance report. Since our review of SSA's 2002 performance report
used different assessment criteria, we could not make valid comparisons
on all aspects of the reports.
[132] This program was an administration proposal to test allowing
disabled beneficiaries to choose their own public or private vocational
rehabilitation provider.
[133] In our October 2002 report GAO-03-225, we noted that SSA reported
progress on all six of its major management challenges in its fiscal
year 2001 annual performance report.
[134] We characterized program evaluations and their uses in GAO/GGD-
00-204.
[135] SSA noted that its fifth strategic goal, "Valued Employees,"
supports the accomplishment of all its basic functions, so its
resources are inherently included in the other four goals.
[136] SSA began to implement an improved cost accounting system in
fiscal year 2002, which will be phased in over the next 3 to 4 years.
[137] U.S. Department of Transportation, Fiscal Year 2002 Performance
and Accountability Report (Washington, D.C.: 2003).
[138] Our review of DOT's fiscal year 1999 report, GAO/RCED-00-201R,
focused on our observations regarding only four of the department's key
outcomes, as well as the major management challenges addressed in the
performance report. Since our review of DOT's 2002 performance report
used different assessment criteria, we could not make valid comparisons
on all aspects of the reports.
[139] GAO-03-108.
[140] Bureau of Transportation Statistics, Source & Accuracy
Compendium, http://www.bts.gov/statpol/SAcompendium.html (Washington,
D.C.: Aug. 15, 2003).
[141] The Federal Emergency Management Agency became part of the
Department of Homeland Security in March 2003.
[142] Formerly part of the Department of Health and Human Services, the
Social Security Administration once again became an independent agency
on March 31, 1995. Congress established the position of Chief Financial
Officer within SSA in the Social Security Independence and Program
Improvements Act of 1994.
[143] GAO/GGD-00-204.
[144] U.S. General Accounting Office, Social Security Administration:
Actions Taken to Strengthen Procedures for Issuing Social Security
Numbers to Noncitizens, but Some Weaknesses Remain, GAO-04-12
(Washington, D.C.: Oct. 15, 2003).
GAO's Mission:
The General Accounting Office, the investigative arm of Congress,
exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. General Accounting Office
441 G Street NW,
Room LM Washington,
D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S.
General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C.
20548: