Human Capital
Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs
Gao ID: GAO-04-291 January 30, 2004
Effective training and development programs are an integral part of a learning environment, helping improve federal workforce performance in achieving agency results. Therefore, in this report GAO was asked to identify examples of selected federal agencies' experiences and some of the key lessons they have learned in designing their training and development programs. This work focused on ways that these agencies (1) assessed agency skills gaps and identified training needs, (2) developed strategies and solutions for these training and development needs, and (3) determined methods to evaluate the effectiveness of training and development programs. GAO worked with five agencies to identify their experiences and lessons learned: the U.S. Army Corps of Engineers (USACE), Department of Defense; Fish and Wildlife Service (FWS), Department of the Interior (Interior); Internal Revenue Service (IRS), Department of the Treasury; the Office of Personnel Management (OPM); and Veterans Health Administration (VHA), Department of Veterans Affairs (VA). Agency officials provided information during interviews and furnished supporting documentation for analysis and review.
GAO identified important lessons learned from five federal agencies' experiences in designing training and development programs for their employees that could be useful to other agencies facing similar challenges. These lessons learned are related to the following three areas. Assessing Agency Skill Requirements and Identifying Training Needs: The agencies used a variety of approaches to assess current and future skill requirements, such as implementing workforce planning models and conducting knowledge and skills inventories. Generally, the agencies are transitioning to more comprehensive approaches. One of the lessons learned was to involve stakeholders and benchmark with others to identify critical skills and competencies and related training needs. For example, IRS officials believed they needed a leadership competency model directly based on the work of their agency's business units. To develop a comprehensive model, they interviewed top IRS leaders and benchmarked with leading practices in the public and private sector. Developing Strategies and Solutions for Training Needs: The agencies considered a mixture of delivery mechanisms, as well as potential sources for training and development opportunities. However, projecting costs and benefits of proposed training and development programs presented challenges for them. The agencies usually developed broad information on anticipated benefits and expected costs of potential investments, although often without tying benefits to specific performance improvements or considering all costs. One of the lessons learned was to establish mechanisms to avoid duplication or inconsistencies. Education Service Representatives in each regional VHA network, for example, coordinate training and development programs with headquarters--sharing information about successful practices and identifying areas where coordination is needed. Determining Methods for Evaluating Training Programs: Overall, the agencies relied primarily on participants' end-of-course evaluations, but they are beginning to use more comprehensive evaluation approaches, including limited use of return-on-investment analysis techniques. One of the lessons learned is to plan for the use of multiple data types and sources in order to attain a balanced assessment once the course is implemented. For example, USACE's training center incorporated pre- and post-tests on over 90 percent of its courses, as well as approaches to collect participants' and course managers' feedback, as part of the design. Four of the five agencies provided comments on a draft of this report. Interior and VA said that they generally agreed with the report's findings regarding their respective agencies. IRS and OPM said that they appreciated the opportunity to be included in the report and to share information on training activities. USACE provided no comments on the draft report.
GAO-04-291, Human Capital: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs
This is the accessible text file for GAO report number GAO-04-291
entitled 'Human Capital: Selected Agencies' Experiences and Lessons
Learned in Designing Training and Development Programs' which was
released on January 30, 2004.
This text file was formatted by the U.S. General Accounting Office
(GAO) to be accessible to users with visual impairments, as part of a
longer term project to improve GAO products' accessibility. Every
attempt has been made to maintain the structural and data integrity of
the original printed product. Accessibility features, such as text
descriptions of tables, consecutively numbered footnotes placed at the
end of the file, and the text of agency comment letters, are provided
but may not exactly duplicate the presentation or format of the printed
version. The portable document format (PDF) file is an exact electronic
replica of the printed version. We welcome your feedback. Please E-mail
your comments regarding the contents or accessibility features of this
document to Webmaster@gao.gov.
This is a work of the U.S. government and is not subject to copyright
protection in the United States. It may be reproduced and distributed
in its entirety without further permission from GAO. Because this work
may contain copyrighted images or other material, permission from the
copyright holder may be necessary if you wish to reproduce this
material separately.
Report to the Chairman, Subcommittee on Oversight of Government
Management, the Federal Workforce, and the District of Columbia,
Committee on Governmental Affairs, U.S. Senate:
January 2004:
HUMAN CAPITAL:
Selected Agencies' Experiences and Lessons Learned in Designing
Training and Development Programs:
GAO-04-291:
GAO Highlights:
Highlights of GAO-04-291, a report to the Chairman, Subcommittee on
Oversight of Government Management, the Federal Workforce, and the
District of Columbia, Committee on Governmental Affairs, U.S. Senate
Why GAO Did This Study:
Effective training and development programs are an integral part of a
learning environment, helping improve federal workforce performance in
achieving agency results. Therefore, in this report GAO was asked to
identify examples of selected federal agencies‘ experiences and some
of the key lessons they have learned in designing their training and
development programs. This work focused on ways that these agencies
(1) assessed agency skills gaps and identified training needs, (2)
developed strategies and solutions for these training and development
needs, and (3) determined methods to evaluate the effectiveness of
training and development programs.
GAO worked with five agencies to identify their experiences and
lessons learned: the U.S. Army Corps of Engineers (USACE), Department
of Defense; Fish and Wildlife Service (FWS), Department of the
Interior (Interior); Internal Revenue Service (IRS), Department of the
Treasury; the Office of Personnel Management (OPM); and Veterans
Health Administration (VHA), Department of Veterans Affairs (VA).
Agency officials provided information during interviews and furnished
supporting documentation for analysis and review.
What GAO Found:
GAO identified important lessons learned from five federal agencies‘
experiences in designing training and development programs for their
employees that could be useful to other agencies facing similar
challenges. These lessons learned are related to the following three
areas.
Assessing Agency Skill Requirements and Identifying Training Needs:
The agencies used a variety of approaches to assess current and future
skill requirements, such as implementing workforce planning models and
conducting knowledge and skills inventories. Generally, the agencies
are transitioning to more comprehensive approaches. One of the
lessons learned was to involve stakeholders and benchmark with others
to identify critical skills and competencies and related training
needs. For example, IRS officials believed they needed a leadership
competency model directly based on the work of their agency‘s business
units. To develop a comprehensive model, they interviewed top IRS
leaders and benchmarked with leading practices in the public and
private sector.
Developing Strategies and Solutions for Training Needs:
The agencies considered a mixture of delivery mechanisms, as well as
potential sources for training and development opportunities. However,
projecting costs and benefits of proposed training and development
programs presented challenges for them. The agencies usually developed
broad information on anticipated benefits and expected costs of
potential investments, although often without tying benefits to
specific performance improvements or considering all costs. One of the
lessons learned was to establish mechanisms to avoid duplication or
inconsistencies. Education Service Representatives in each regional
VHA network, for example, coordinate training and development programs
with headquarters”sharing information about successful practices and
identifying areas where coordination is needed.
Determining Methods for Evaluating Training Programs:
Overall, the agencies relied primarily on participants‘ end-of-course
evaluations, but they are beginning to use more comprehensive
evaluation approaches, including limited use of return-on-investment
analysis techniques. One of the lessons learned is to plan for the use
of multiple data types and sources in order to attain a balanced
assessment once the course is implemented. For example, USACE‘s
training center incorporated pre- and post-tests on over 90 percent of
its courses, as well as approaches to collect participants‘ and course
managers‘ feedback, as part of the design.
Four of the five agencies provided comments on a draft of this report.
Interior and VA said that they generally agreed with the report‘s
findings regarding their respective agencies. IRS and OPM said that
they appreciated the opportunity to be included in the report and to
share information on training activities. USACE provided no comments
on the draft report.
What GAO Recommends:
www.gao.gov/cgi-bin/getrpt?GAO-04-291.
To view the full product, including the scope and methodology, click
on the link above. For more information, contact George Stalcup at
(202) 512-6806 or stalcupg@gao.gov.
[End of section]
Contents:
Letter:
Results in Brief:
Background:
Agencies Used Varied Approaches in Assessing Skills and Competencies
and Identifying Related Training Needs:
Agencies Developed Strategies and Solutions for Their Training Needs:
Agencies are Considering More Sophisticated Evaluation Approaches As
Part of Designing their Training and Development Programs:
Conclusions and Observations:
Agency Comments and Our Evaluation:
Appendixes:
Appendix I: Objective, Scope and Methodology:
Appendix II: Background on Selected Agencies and Their Training and
Development Functions:
Appendix III: Core Characteristics of a Strategic Training and
Development Process:
Appendix IV: Comments from the Department of the Interior:
GAO Comment:
Appendix V: Comments from the Internal Revenue Service:
Appendix VI: Comments from the Office of Personnel Management:
Appendix VII: GAO Contacts and Staff Acknowledgments:
GAO Contacts:
Acknowledgments:
Figures:
Figure 1: Four Components of a Strategic Training and Development
Process:
Figure 2: Strategic Workforce Planning Process:
Figure 3: OPM's Five-step Workforce Planning Process:
Figure 4: IRS's Core Management Responsibilities and Leadership
Competencies:
Figure 5: Competencies in VHA's High Performance Development Model:
Figure 6: Steps for Developing Strategies and Solutions for Training
and Development Needs:
Figure 7: Steps in Determining Methods for Evaluating Training
Programs:
Letter January 30, 2004:
The Honorable George V. Voinovich: Chairman:
Subcommittee on Oversight of Government Management, the Federal
Workforce, and the District of Columbia:
Committee on Governmental Affairs:
United States Senate:
Dear Mr. Chairman:
To effectively address the nation's most urgent priorities and take
advantage of emerging opportunities, federal agencies need to continue
to build their fundamental capabilities to carry our their work in an
environment that is increasingly complex and rapidly changing. To build
their capacity, agencies should invest resources wisely to ensure that
their employees have the information, skills, and competencies they
need to succeed. As you are well aware, these investments must include
training and development efforts to continuously enhance the skills and
competencies of the federal workforce and improve the quality of
agencies' results.
As agreed with your office, this report provides information on
selected federal agencies' experiences and lessons learned in key
aspects of designing training and development programs for their
employees. Specifically, we focused on the agencies' experiences and
lessons learned related to:
* assessing current and future agency skill and competency requirements
and identifying related training and development needs,
* developing strategies and solutions for training and development
needs, and:
* determining methods to evaluate the effectiveness of training and
development programs.
For this review, lessons learned were defined as knowledge that could
be applied in the future that the agencies gained through either
positive or negative experiences. The experiences and lessons learned
from the five agencies we reviewed may well provide some valuable ideas
and useful approaches that could be adopted by other federal agencies
as they attempt to address ongoing training and development challenges-
-particularly those related to the elements within the training process
that relate to design and development.
To address these issues and as agreed with your office, we focused our
review on five federal agencies: U.S. Army Corps of Engineers (USACE),
Department of Defense; Fish and Wildlife Service (FWS), Department of
the Interior (Interior); Internal Revenue Service (IRS), Department of
the Treasury; Office of Personnel Management (OPM); and Veterans Health
Administration (VHA), Department of Veterans Affairs (VA). At the
headquarters of the five agencies--and in some field locations--we
collected and reviewed documents on the agencies' training and
development efforts and interviewed officials from the agencies' human
capital and training organizations, as well as assorted program
offices. We selected the five agencies for various reasons, including
your office's interests, the diversity of employee occupations within
the agency, and reported innovative approaches for training and
developing their employees. This agency selection process was not
designed to identify examples that could be considered representative
of all training and development efforts at the five agencies or the
federal government as a whole. Furthermore, in citing examples that
relate to the lessons learned on the design of training, we did not
assess the effectiveness of these training programs and practices.
Rather, we attempted to highlight some of the experiences and lessons
that the agencies found helped them move forward in improving their
training and development programs. Federal agencies' training and
development strategies, and how they are designed to operate in
conjunction with other strategies to improve individual and
organizational performance, continue to change and evolve.
We conducted our review from August 2002 through November 2003 in
accordance with generally accepted government auditing standards. See
appendix I for additional information on our objective, scope, and
methodology.
Results in Brief:
Officials from the five agencies in our review recognized the
importance of assessing current and future agency skill and competency
requirements to identify related workforce training needs. The agencies
used several approaches to design training and development programs
that focused on the skills and competencies their assessments indicated
needed enhanced attention. The officials emphasized that agencies are
transitioning to more formal and comprehensive planning approaches to
assess skill and competency requirements and identify related training
and development needs--primarily as part of broader efforts to
incorporate workforce planning into ongoing strategic planning and
budgeting processes focused on achieving results. The following are
lessons learned identified during this review related to assessing
skill and competency requirements and identifying training needs.
* Involve key stakeholders and benchmark with other organizations when
identifying skills and competencies to help ensure that training and
development programs are aligned with current and emerging needs and
business practices.
* Analyze existing agency data on employees' skills and competencies
and information from performance appraisals to help identify skills and
competencies that need to be addressed throughout the agency as well as
on an individual basis.
* Link the agency's workforce planning efforts with training needs
assessments to ensure consistency and enhance strategic alignment.
* Consider the training needs of staff from other organizations that
will likely use the agency's training programs or facilities to
effectively leverage training investments and meet diverse needs.
These agencies developed a wide range of strategies and solutions to
improve performance through designing training and development programs
for their employees. Officials told us they considered a mixture of
both on-the-job and other developmental programs and contemplated an
assortment of mechanisms for delivering the training as well as
potential sources for the learning need. Agency officials have found
projecting costs and benefits of proposed training and development
programs to be very challenging. Although they sometimes developed
broad information on anticipated benefits and expected costs, this
often did not involve tying anticipated benefits to specific
performance improvements or considering all related costs. Our review
identified the following lessons learned in developing strategies and
solutions when designing agency training and development programs.
* Incorporate information on employees' various competency levels and
job needs into the design of training and development programs to
increase their relevancy and timeliness.
* Assess options for using other organizations' course content, staff,
services, or facilities when designing a new training and development
program in order to develop efficient and cost-effective strategies.
* Establish mechanisms and controls to avoid unnecessary duplication or
inconsistency within and across agencies' training efforts.
* Develop and use criteria for determining the optimal mix of delivery
mechanisms to use in order to select the most effective approaches
given each learning situation.
* Ensure that employees have the needed equipment and technologies so
that they can take maximum advantage of learning opportunities.
* Plan early when developing integrated solutions that complement other
planned and ongoing strategies to improve performance so that when
implemented the strategies work effectively and are aligned to help
achieve agency goals.
* Plan for the direct participation of senior agency leaders and
experienced staff in the delivery of training and development programs
to increase buy-in and build support for organizational change.
Evaluating training programs is key to ensuring that training and
development programs are effective. Overall, the five agencies relied
primarily on standard end-of-course evaluations to obtain the
participants' reaction to, and satisfaction with, a specific training
course or learning opportunity. However, officials said that they have
begun or are planning to use more comprehensive and sophisticated
techniques for assessing the extent to which training and development
programs increased employees' knowledge and skills and enhanced
individual and organizational performance. These techniques included
pre-and post-testing, tracking changes in individual and program
performance, and some limited use of return-on-investment (ROI)
analyses. The lessons agencies learned in designing methods to evaluate
training and development programs included the following.
* Incorporate appropriate aspects of the evaluation approach when
designing training and development programs by specifying what results
are expected to better ensure the availability and use of quality
performance data.
* Consider new approaches for collecting and analyzing performance data
with the aim of increasing the quality and quantity of training
evaluation feedback.
* Plan for the use of multiple data types and sources to provide a
balanced approach in assessing the effectiveness of training and
development programs.
* Take into account all relevant factors for determining the costs of a
training and development program to better ascertain whether it is
cost-effective in relation to benefits achieved.
Four of the five selected agencies provided comments on a draft of this
report. Interior and VA said that they generally agreed with the
report's findings relating to their respective agencies. IRS said that
it was honored to share some of its lessons learned with us for
governmentwide dissemination. OPM said that it appreciated the
opportunity to be included in the report and to share information on
its training and development activities and programs. USACE informed us
that it had no comments on our draft report.
Background:
We recently issued an exposure draft of an assessment guide that
introduces a framework for evaluating a federal agency's training and
development efforts.[Footnote 1] This assessment guide consists of a
set of principles and key questions that federal agencies can use to
ensure that their training and development investments are targeted
strategically and are not wasted on efforts that are irrelevant,
duplicative, or ineffective. As detailed in our assessment guide, the
training and development process can loosely be segmented into four
broad, interrelated components: (1) planning/front-end analysis, (2)
design/development, (3) implementation, and (4) evaluation. Figure 1
depicts an overview of this process along with the general
relationships between the four components that help to produce a
strategic approach to federal agencies' training and development
efforts. Although these components can be discussed separately, they
are not mutually exclusive and encompass subcomponents that may blend
with one another. For instance, evaluation is an integral part of the
planning/front-end analysis as agencies strive to reach agreement up
front on how the success of various strategies to improve performance,
including training and development efforts, will be assessed. As noted
in the assessment guide, agencies can build on lessons learned and
performance data and feedback from previous experiences. This report
can provide a starting point for agencies to use to build on the
experiences and lessons learned by the five agencies we reviewed as
part of their efforts to design and develop training and development
programs. (See app. II for a description of the five agencies included
in this study.):
Figure 1: Four Components of a Strategic Training and Development
Process:
[See PDF for image]
[End of figure]
Our assessment guide also summarizes our observations on the core
characteristics that make a training and development process effective
and strategically focused on achieving results. These eight core
characteristics are described in more detail in appendix III, and
include:
* strategic alignment,
* leadership commitment and communication,
* stakeholder involvement,
* accountability and recognition,
* effective resource allocation,
* partnerships and learning from others,
* data quality assurance, and:
* continuous performance improvement.
A concerted effort to integrate these core characteristics can further
an agency's efforts to continually improve its training and development
process.
Agencies Used Varied Approaches in Assessing Skills and Competencies
and Identifying Related Training Needs:
Federal agencies face diverse challenges in their efforts to identify
and measure the skills and competencies that their employees must
possess to support missions and goals. Officials from the five agencies
in our review recognized the importance of assessing the need for
specific skills and competencies now and in the future in order to
identify related workforce training needs. These agencies generally
focused on the desired performance of the agency and its employees,
determined the difference between the desired and actual skill levels,
and attempted to identify the key factors contributing to performance,
including the need for enhanced workforce competencies. Officials used
a variety of approaches and tools to assist in determining the human
capital skills and competencies that are critical to achieving their
long-term goals.
An agency's ultimate goal in undertaking training and development
efforts is, of course, to optimize employee and organizational
performance. To help ensure that each training program is linked to
improving individual and agency performance, agencies first need to
analyze their strategic and performance goals so that they can
determine where training and development can most effectively enhance
goal achievement.
Organizations can evaluate the extent to which human capital approaches
support their accomplishment of current, emerging, and future strategic
goals through the use of workforce planning. Workforce planning focuses
on determining the skills and competencies needed now and in the future
to meet the agency's goals; identifying the current and projected level
of the skills and competencies of the workforce; and crafting
strategies for acquiring, developing, and retaining people to address
any identified needs. These needs include the knowledge, skills, and
abilities needed for the agency to pursue its current and future
mission as well as the size of the workforce and its deployment across
the organization. After identifying the skills and competencies that
employees need now or in the future, agencies must tackle the challenge
of determining what combination of strategies to use, such as hiring
new employees with needed skills and competencies, relying on
outsourcing, and/or enhancing employees' skills and competencies
through training and development. While agencies' approaches to
workforce planning will vary, we have identified the need for a
strategic workforce planning process to ensure that each agency's human
capital program capitalizes on its workforce's strengths and addresses
related challenges in a manner that is clearly linked to achieving the
agency's missions and goals.[Footnote 2] Figure 2 presents a model of
this strategic workforce planning process.
Figure 2: Strategic Workforce Planning Process:
[See PDF for image]
[End of figure]
The focus in this report is on the workforce strategies that involve
the design of training and development programs.
Agencies' Experiences in Assessing Skill and Competency Requirements
and Identifying Training Needs:
The five agencies used several approaches to design training and
development programs that focused on skills and competencies their
assessments indicated needed enhanced attention. One common approach
that officials used to help identify training needs was interviewing or
surveying managers, supervisors, and employees. Agencies also
established councils and held conferences, made comparisons with
leading organizations through benchmarking, and analyzed workforce data
and trends.
OPM's 2001 skills assessment, for example, relied on a survey of agency
managers and supervisors.[Footnote 3] Officials used the survey results
to identify the most important occupational competencies needed to
achieve OPM's mission, the level at which employees possessed those
competencies, and the level at which they would be needed in the
future. Their analysis identified minor gaps in the level of
competencies needed for both current and future work in mission
critical occupations. In addition, the analysis pointed out more
serious developmental needs for OPM's retirement and insurance benefits
specialists. These needs were related to the changing role of these
specialists, who increasingly need to work more closely with clients in
responding to complex issues.
Also in 2001, IRS established a workforce planning council consisting
of senior management representatives from each of the agency's
operating divisions. IRS officials told us that this council has become
the primary vehicle for communicating workforce planning information
among IRS's four operating divisions. At FWS, the human resources
office hosted a 3-day workforce planning conference in 2002 to draw on
the experience and expertise of agency personnel in identifying
critical workforce issues for the agency for the next 3 to 5 years.
Managers and program experts representing all eight major FWS program
offices, seven field regions, and headquarters offices participated.
Agencies also compared their performance and needed skills and
competencies with leading organizations through benchmarking. In August
2000, VHA commissioned an internal task force charged with developing a
well-defined, comprehensive succession plan for the agency. The ideas
garnered from benchmarking led to VHA establishing an expectation for
agency leaders to help identify and train their successors. The task
force's December 2001 report presented a comprehensive succession plan
for VHA, and implementing a comprehensive leadership development
program was one of the six major components of this plan.
Agency officials also analyzed workforce data to assess skills and help
identify training needs. Generally, they collected information on
employee demographics and retirement eligibility and used these data to
project attrition and retirement rates. OPM officials, for example,
collected and analyzed attrition and turnover data on the agency's
senior executives along with the distribution of current executives by
unit and projected retirements through 2010.[Footnote 4] They also
collected and analyzed data on hires, separations, and workforce
diversity across the agency. As part of this analysis, officials
assessed the agency's use of contractors and considered how sourcing
alternatives could affect OPM's plans for hiring, training, and
development.
Officials from the five agencies told us that they used a wide range of
resources and tools to assess skills and competencies as part of
identifying and designing needed training and development programs.
They used workforce planning models; assessed the workforce in view of
organizational, occupational, and unit-based competency standards;
conducted knowledge and skills inventories; and evaluated job
performance appraisals and information from individual development
plans (IDP).[Footnote 5] To identify needed executive competencies, for
example, OPM used a five-step workforce planning model that it had
developed in its role of providing human capital tools for use by other
federal agencies (see fig. 3). The resulting analysis called for
enhancing leadership development within OPM.
Figure 3: OPM's Five-step Workforce Planning Process:
1. Set strategic direction to drive agency operations and define how
the agency will know when and if it has succeeded;
2. Analyze the workforce, identify skill gaps, and conduct workforce
analysis;
3. Develop a workforce action plan that lays out specific tasks
and actions the agency needs to take in order to achieve the agency's
human resources goals and objectives;
4. Implement the workforce action plan by executing the schedule that
includes measurable workforce goals and milestones;
5. Monitor progress, evaluate success, and revise plan as needed.
Source: OPM.
[End of figure]
Officials said that their agencies have transitioned, or are in the
process of transitioning, to more comprehensive, consistent planning
approaches. This transition is coming about as agencies attempt to
institutionalize their workforce planning efforts as part of their
ongoing strategic planning and budgeting processes. In 2002, FWS
conducted its first formal, agencywide workforce planning process.
Although FWS initially employed a contractor to help develop a
permanent workforce planning process, it plans to continue to manage it
in-house. This workforce planning process is to be implemented on a 2-
year cycle that is integrated with the agency's strategic planning and
budgeting processes.
Agencies' Lessons Learned in Assessing Skill and Competency
Requirements and Identifying Training Needs:
By considering the viewpoints of a range of stakeholders and candidly
and openly assessing progress toward meeting their goals, agencies can
help ensure that their strategic and annual performance planning
processes adequately reflect current ideas, policies, and practices in
the field. Agencies continue to integrate workforce planning into these
other planning processes. It is important to note that a wide variety
of strategies other than training and development are also available to
agency leaders as they attempt to transform their cultures and
operations.[Footnote 6] Training and development is not always the best
solution--reengineering processes or other actions may be needed to
build an environment that effectively supports performance. In
addition, training and development strategies frequently need to be
implemented in conjunction with other initiatives, given that the day-
to-day environment and organizational culture may also need to change
to enable employees to successfully use new skills or competencies on
the job.
Our review identified four lessons learned by the agencies related to
assessing skills and competencies to identify, focus, and prioritize
training needs.
Lesson learned: Involve key stakeholders and benchmark with other
organizations when identifying skills and competencies to help ensure
that training and development programs are aligned with current and
emerging needs and business practices.
Organizations in the private and public sectors have increasingly
turned to developing competency models that outline behaviorally
defined skills and competencies employees should possess and that can
be tied directly to training and development plans and programs. We
have found that an effective performance management system uses
competencies to provide a fuller assessment of performance.[Footnote 7]
IRS involved key stakeholders and benchmarked with other organizations
in developing its leadership competency model:
In working to build its leadership development program in the wake of
the IRS Restructuring and Reform Act of 1998, IRS officials believed
they needed a leadership competency model that was based directly on
the work of IRS's business units. To identify the essential
characteristics that enable IRS employees to function as effective
leaders in the newly modernized agency, human capital specialists at
IRS conducted behavioral interviews with 35 top IRS leaders in 1999,
asking them to identify major successes and challenges during their
careers. Using information gathered from these interviews, officials
identified core management responsibilities and corresponding
competencies required for leaders in IRS. With the assistance of a
contractor, IRS validated the leadership competency model by comparing
it against leading practices in the public and private sector and
linking it to the mission and goals of the agency.[Footnote 8] The
resulting competency model now forms the basis for IRS's leadership
development efforts, as well as how IRS selects, evaluates, and
recognizes its leaders. Figure 4 shows a listing of IRS's five core
management responsibilities and the 21 corresponding leadership
competencies.
Figure 4: IRS's Core Management Responsibilities and Leadership
Competencies:
Leadership:
* Adaptability;
* Communication;
* Decisiveness;
* Integrity/honesty;
* Service motivation;
* Strategic thinking;
Employee satisfaction:
* Continual learning;
* Developing others;
* Diversity awareness;
* Group leadership;
* Teamwork;
Customer satisfaction:
* Customer focus;
* Entrepreneurship;
* External awareness;
* Influencing/ negotiating;
* Partnering;
Business results:
* Achievement orientation;
* Business acumen;
* Political savvy;
* Problem solving;
* Technical credibility;
Equal employment opportunity (EEO) and diversity:
* Supporting competencies[A].
Source: IRS.
[A] Italicized competencies support the "EEO and diversity"
responsibility.
[End of figure]
VHA benchmarked with other organizations and used a pilot test in
developing its high performance development model:
VHA conducted an extensive literature search and benchmarked with
several leading private sector firms (including Bell South, Coca-Cola,
and Motorola) to provide a foundation for its effort to create a new
competency model for VHA employees. VHA used this information in
developing its high performance development model, which was
implemented throughout the department in 2002. This model consists of
eight core competencies and related performance tools that represent
the major skills and competencies that employees need to fulfill VHA's
mission. The model was designed to serve as a framework for identifying
and developing future leaders, as well as to enhance development of
VHA's entire workforce. VHA said that using this model helped more
effectively align training and development programs with agency
priorities. According to VHA, the fact that the core competencies apply
to all levels and functions within the agency helps ensure alignment
within and between organizational units and is a key component in
motivating sustained and improved performance. VHA also uses the model
on an individual employee basis as a process for identifying specific
developmental needs. Figure 5 lists the eight core competencies in
VHA's high performance development model.
Figure 5: Competencies in VHA's High Performance Development Model:
* Personal mastery;
* Interpersonal effectiveness;
* Flexibility and adaptability;
* Systems thinking;
* Technical skills;
* Customer service;
* Creative thinking;
* Organizational stewardship.
Source: VHA.
[End of figure]
Lesson learned: Analyze existing agency data on individual employee's
skills and competencies and information from performance appraisals to
help identify skills and competencies that need to be addressed
throughout the agency as well as on an individual basis.
To obtain a unit-or agencywide perspective of skills and competencies,
some agencies such as USACE and IRS have explored new ways of
aggregating data from tools that are primarily focused on individual
employees, such as IDPs, performance assessments, and 360-degree
feedback instruments.[Footnote 9] Officials from these agencies told us
that this information helped them discern a clearer picture of the
overall strengths and weaknesses of their employees and offered
direction in planning and designing training and development programs
to help focus efforts to enhance skills and competencies throughout the
agency.
USACE's automated training management program provides a Web-enabled
integrated database:
According to USACE officials, using an automated training management
program has allowed managers to identify divisionwide gaps in workforce
skills and competencies. Using this system (currently in four of
USACE's eight divisions) employees prepare an IDP assessing their
knowledge, skills, and abilities in relation to a series of mission
essential tasks. With supervisory guidance, each task is identified as
critical, important, or beneficial and employees indicate whether they
have received adequate, partial, or no training in that area. With this
assessment as a guide, the supervisor and employee can consult the
system's built-in course catalog to select internal or external
training to enhance the employee's development. In addition, the system
also has the capability of aggregating data. USACE officials said that
this capability provides a simple method for division managers to
obtain a picture of the level of skills and competencies in their
workforce. This information informs decisions on training priorities
and helps managers determine the most efficient use of available
resources.
IRS aggregated data from 360-degree feedback instruments to help
identify training needs:
To assess the progress and developmental needs of leaders within IRS,
the agency's leadership development office recently aggregated and
analyzed multiyear data from the 360-degree performance assessments of
IRS managers. This analysis helped to show areas of strength and
weakness in skills and competencies across the agency's managerial
ranks. The director of leadership and organizational effectiveness at
IRS said that, in the past, the agency did not sufficiently assist
managers in effectively using 360-degree feedback they received.
However, he said that IRS now emphasizes the importance of using 360-
degree feedback data, both on an organizational and individual basis,
to focus on strengths in developing key leadership competencies. IRS's
Extraordinary Leader Program involves designing unique developmental
approaches to help managers become more effective leaders. Using
results of the manager's 360-degree assessment, IRS creates a
customized leadership development program focusing first on correcting
any "fatal flaw" weaknesses and then building on the manager's
demonstrated strengths in areas that IRS has identified as key to
providing effective leadership within its organizational culture and
operating environment.
Lesson learned: Link the agency's workforce planning efforts with
training needs assessments to ensure consistency and enhance strategic
alignment.
The agencies' training and development organizations had a range of
responsibilities, including designing training and development
programs based on strategic initiatives, soliciting input from
stakeholders, and prioritizing and scheduling training based on
strategic initiatives and stakeholder input. Generally, the training
organization and agency stakeholders can work together more effectively
when they better understand how each office or function within the
agency contributes to achieving business goals. In some cases, this
included efforts to link training needs assessments with the agency's
overall workforce planning efforts. Officials told us that this linkage
helped ensure that workforce plans developed by the agency's human
capital office were consistent with training needs assessments done by
the agency's training and development organization.
FWS involved key internal stakeholders in its planning processes:
When assessing workforce skills and competencies, FWS officials worked
to ensure that the agency's workforce plan was linked with a training
needs assessment done by its training center. The workforce planning
effort identified broad competencies needed across the agency's
workforce while the training needs assessment identified the types of
training courses to develop skills and competencies within agency units
and occupations. Officials said that they viewed the training needs
assessment as a tool that was useful in refining the agency's workforce
plan and in prioritizing and budgeting for the development and delivery
of training. Both efforts involved key stakeholders from the human
capital and training offices as well as other FWS units. In the future,
officials said they will rely on the results of the agency's workforce
planning efforts to directly serve as the agency's training needs
assessment.
USACE relies on its Learning Advisory Board and automated training
management program to effectively link planning efforts:
USACE relies chiefly on the coordination activities of its Learning
Advisory Board to ensure its workforce planning efforts and training
needs assessments are effectively linked. USACE in 2001 formed the
Learning Advisory Board, comprised of senior managers from across the
agency, to review the adequacy of USACE's training and development and
ensure that training is properly aligned with the agency's missions,
goals, and plans. In addition, the four divisions that use the
automated training management program can also rely on data from that
system to assess training needs. This system allows managers to compare
information on individuals' skills and competencies with workforce
planning results from within the division and across the agency.
According to USACE officials, this systematic comparison more closely
links workforce planning and training needs assessments to the
essential mission-related operations.
Lesson learned: Consider the training needs of staff from other
organizations that will likely use the agency's training programs or
facilities to effectively leverage training investments and meet
diverse needs.
When planning and designing training programs for its employees, FWS
and USACE officials told us that they gained insight into the
assessment of potential training solutions by considering the possible
involvement of trainees from other organizations. In some instances,
for example, it would not have been cost-effective to design, develop,
and deliver a training effort for a small number of employees or
occupations. However, officials' determinations that other agencies or
organizations also needed similar training provided the critical mass
needed to move forward. Officials said they found that partnering with
other organizations helped make training efforts more cost effective to
design, develop, and deliver.
FWS's training center assessed training needs of possible participants
from other Interior components:
Officials from FWS's training center said that although they focus
primarily on meeting the needs of FWS employees when planning and
designing training, they also look at the training needs of other
organizations, particularly other agencies within Interior. To aid in
communication and coordination, agencies within the department
designate employees to serve as liaisons between each of the Interior
agencies and FWS's training center. These liaisons facilitate efforts
to incorporate their agency's needs into the training center's plans
and the design and delivery of training and development programs.
According to agency officials, these liaisons help ensure that the
center's courses remain current because they facilitate a dialogue
between the various agencies. For example, a recent FWS course on
wetland plant identification involved participants from FWS; U.S.
Geological Survey, which is another Interior component; Natural
Resources Conservation Service, an agency of the U.S. Department of
Agriculture; USACE; and three private sector firms.
USACE's training center considers training needs of staff from other
federal and state organizations:
Officials from USACE's training center told us that during the agency's
annual training needs assessment, they assess the possible training
demand from other organizations' employees in addition to identifying
workforce development needs of units and offices within USACE.
According to USACE, approximately 2,500 people each year--about 25
percent of the participants in the agency's training center programs--
are from other federal and state organizations. Officials noted that
the training center offers courses needed to obtain certifications for
certain professional requirements. It offers courses accredited by
several professional associations, including the National Society for
Professional Engineers, the American Institute of Architects, and the
International Association for Continuing Education and Training. USACE
officials said that they would like to offer training to a greater
number of employees from private sector firms; however, current law
requires receipts for services provided to private individuals and
organizations to be deposited into the general treasury as
miscellaneous receipts.[Footnote 10] USACE officials said that this
requirement hampers the ability of USACE's training center to keep
funds it could generate to further invest in its training programs.
However, when agencies are required to return receipts for services to
the Treasury, the Congress is preserving its oversight and control over
the programs generating the fees.
Agencies Developed Strategies and Solutions for Their Training Needs:
Agencies carry out their training and development efforts on the basis
of estimated needs, priorities, and available resources and recognize
that adequate planning allows them to establish priorities and
determine the best ways to leverage investments to improve performance.
The five agencies we reviewed set priorities for training and
development on the basis of various factors, such as the results from
skill and competency assessments, the availability of resources, and
the interests of agency leaders. They usually relied on training
officials, agency managers, and subject matter experts to assist in
developing strategies and approaches for addressing training needs.
Although the agencies designed and delivered training using both
centralized and decentralized approaches, we found that leadership
development programs were more highly centralized and managed at
headquarters. Agency officials acknowledged that they found projecting
costs and benefits of proposed training and development programs to be
very challenging. Although they sometimes developed broad information
on anticipated benefits and expected costs, this often did not involve
tying anticipated benefits to specific performance improvements or
considering all related costs.
As outlined in figure 6, agencies can plan and establish priorities by
developing an annual training plan to target developmental areas of
greatest need and outline the most cost-effective training approaches
to address those areas. Considerations involved in assessing investment
opportunities for the training plan include balancing the competing
demands confronting the agency and the amount of resources available in
order to determine how those demands can best be met with available
resources. It is also important to consider how to effectively
integrate all of the strategies the agency plans to use to improve
performance and meet emerging demands. When training is identified as a
solution to improve performance, agencies can compare various training
strategies by weighing their estimated costs and anticipated benefits
to build a convincing business case that supports the selected training
strategy. Developing a business case that sets forth the expected costs
and benefits of the performance improvement investment provides
decision makers with essential information for allocating necessary
resources.
Figure 6: Steps for Developing Strategies and Solutions for Training
and Development Needs:
[See PDF for image]
[End of figure]
Agencies' Experiences in Developing Strategies and Solutions for
Training Needs:
The agencies we reviewed used a wide range of strategies and solutions
to improve performance through designing training and development
programs for their employees. Officials told us they considered a
mixture of both on-the-job and other developmental programs,
contemplated an assortment of mechanisms for delivering the training,
and assessed potential sources to meet their learning needs. For
example, USACE's leadership development program for midlevel engineers
and scientists involved formal classroom training, mentoring, and a 6-
month developmental assignment. VHA employees can access a wide variety
of informational and educational content through the VA Knowledge
Network, a satellite-based system of live and on-demand programming
delivered directly to employees' desktops. IRS's training unit
developed an automated ROI workbook tool that the agency's business
units can use to assess whether proposed training programs should be
delivered in a classroom or by an e-learning approach. OPM partnered
with an employee union to offer a midcareer development program that
provided an opportunity for current OPM employees to enhance existing
skills, explore new career fields, and gain practical experience.
Officials from FWS's training center cited courses in negotiation,
communication, and interpersonal skills as examples of vendor-provided
courses. They said the center decided not to invest in designing these
courses since they do not require field experience and expertise, which
are critical in designing other FWS courses.
Projecting costs and benefits of proposed training and development
programs was a challenge for the five agencies. They usually developed
broad information on anticipated benefits and expected costs, often
without tying anticipated benefits to specific performance improvements
or considering all costs related to the training program. For example,
VHA officials told us that the agency's assessments of anticipated
benefits and expected costs of proposed training are generally
unsystematic. One VHA office or field location may not have analyzed
the relative costs or benefits of proposed training while another
office or location may have considered anticipated benefits and
developed estimates of costs and savings using different training
approaches. At FWS, some proposed training programs, such as its
Advanced Leadership Development Program, involved detailed estimates of
costs, both for the training center and for participating FWS field
offices, as well as the identification of specific competencies to be
developed in the program. Other proposed training programs at FWS did
not have documented and detailed estimates of expected costs and
benefits. FWS officials said that they assess the anticipated costs and
benefits of all their proposed training and development programs but
that the extent of these assessments and the amount of documentation
supporting the assessments vary, depending on many factors, such as the
content, delivery mechanism, and uniqueness of the proposed training.
According to agency officials we interviewed, limited funding sometimes
affected agencies' abilities to design and deliver training and
development programs that officials believed were needed. USACE
officials said, for example, that in a recent survey, about two-thirds
of agency supervisors and one-half of agency executives believed that
the agency had less funding for training civilian employees than is
needed. Funding and resource limitations sometimes forced the agencies
to think of new and practical ways to ensure that their employees had
the knowledge and skills needed to carry out their work. For example,
because of an unexpected decrease in available travel funds, VHA
officials canceled plans for a large national conference on the use on
an automated managerial cost accounting system. This system was
designed to provide VHA managers with data important in making clinical
decisions, managing workload, and controlling medical care costs.
Instead, to enable employees to obtain the information that would have
been presented at the conference, VHA officials provided the content
via a satellite broadcast along with a series of audio conferences.
At FWS, the number of people that the training center can train onsite,
of course, is limited by the current capacity of the facility
classrooms and residences. According to FWS officials, incorporating a
blended learning approach into its curriculum has enabled trainees to
perform part of the course work outside class, thus allowing
instructors to focus on those topics that require special attention.
Training center officials also said that the agency has increased its
use of e-learning and other mechanisms to develop employees where they
work instead of coming to the training center. Officials told us that
FWS has reduced training costs by offering more training online and
using CD ROMs to provide field offices with course material that
previously had been offered only on location at the training center.
Agencies' Lessons Learned in Developing Strategies and Solutions for
Their Training Needs:
Agency officials have encountered a variety of challenges in their
efforts to design training programs to meet the developmental needs of
their employees. How agencies respond to these challenges can greatly
affect their success in aligning priorities with strategic direction,
assessing strategies, identifying alternative sources and methods,
weighing potential costs and anticipated benefits, and assessing how
other performance improvement initiatives might complement training
efforts. It is also important to consider evaluation feedback on an
ongoing basis. We identified eight lessons learned related to the five
agencies' efforts to develop strategies and solutions for their
training and development needs.
Lesson learned: Incorporate information on employees' various
competency levels and job needs into the design of training and
development programs to increase their relevancy and timeliness.
When designing effective training and development programs, the way the
work is actually to be done on the job and the developmental needs of
the expected trainees are key considerations. Analyzing the tasks of
specific jobs and occupations can help ensure that training accurately
reflects the way employees are expected to perform on a day-to-day
basis. To help ensure that the training effectively addresses
employees' developmental needs, agencies can determine the workforce's
level of proficiency in mission-critical skills and competencies by
conducting skills assessments and using information obtained through
interviews or surveys of employees and their supervisors. The increased
information and insight provided by these approaches can allow agencies
to incorporate information on employees' various competency levels and
job needs into the design of training and development programs--
increasing both the relevancy and timeliness of the learning.
FWS targeted training to employees to leverage specific knowledge and
experiences:
FWS's training center recognized employees' various competency levels
and job needs when developing courses on the use of geographic
information systems (GIS), which are becoming increasingly important
mapping and information analysis tools for natural resources agencies,
according to FWS officials.[Footnote 11] Rather than providing
instruction focused solely on software features and functions of GIS,
the training center designed its curriculum to teach the application of
GIS to employees based on their roles in natural resources management.
With a focus on these varied roles and related needs, the center
developed separate GIS training courses for "explorers" (natural
resource managers or others just wanting to know about GIS), "users"
(biologists and other personnel using GIS in their daily job) and
"developers" (those individuals designing and developing a natural
resources GIS for use by others). Officials at FWS's training center
said that with the training targeted to the specific background and
needs of employees, FWS can minimize the time spent teaching
participants information that they already know or do not need to know
in carrying out their job responsibilities. The training center's
curriculum consists of 15 separate GIS courses, with 3 additional
courses under consideration or development. According to FWS officials,
about 1,200 FWS employees at over 400 offices use GIS software in their
jobs.
IRS used skills assessments to focus on developmental needs:
To assist in identifying employees' competency levels and incorporating
job needs into training, IRS developed and used technical assessment
batteries for the agency's field assistance personnel and customer
service representatives. These multiple-choice instruments were
designed to assess the key technical knowledge that the employee needs
in order to carry out his or her job. On the basis of each employee's
test results, the agency will recommend specific training and other
appropriate interventions, such as mentoring, to improve performance.
IRS officials told us that in some cases where the assessments showed
that individuals were already knowledgeable in a particular area,
employees still wanted to take the related training because they viewed
training as a job benefit as well as a way to improve knowledge and
skills for their jobs. This provides an example of how important
agencies' considerations of the organizational culture and working
environment when designing training and development programs are in
preparing for and addressing issues that may arise during
implementation. As we point out in our training guide, employees need
to not only understand the goals of agencies' training and development
efforts, but also to accept responsibility for developing their
competencies and careers, as well as for improving their organization's
performance.[Footnote 12]
Lesson learned: Assess options for using other organizations' course
content, staff, services, or facilities when designing a new training
and development program in order to develop efficient and cost-
effective strategies.
When thinking about strategies and sources for the design of a new
training and development program, officials can potentially discover
more efficient and cost-effective approaches through the use of other
organizations' course content, staff, services, and/or facilities.
Adequate planning can help an agency in meeting the developmental needs
of trainees without overburdening the agency's training capacity or
creating excess capacity. Obtaining reasonable estimates of likely
costs and identifying potential obstacles of using others' training
resources can help agencies develop more informed perspectives on ways
to effectively leverage resources.
IRS considered several options in designing a course for senior
managers:
In evaluating options for designing and delivering a new training
course for its senior managers, IRS considered various sources for such
training, including internal resources, contractor support, and
partnerships. IRS officials said that they considered internal
resources to design the training but quickly realized that the agency
did not have sufficient expertise. The officials also considered using
a contractor but concluded that the costs would be too high. Instead,
IRS decided to partner with the Federal Executive Institute, an OPM-
sponsored training facility in Charlottesville, Virginia, that provides
training to senior employees from across the federal government. IRS
officials found that partnering IRS design and subject matter experts
with institute and other renowned leaders in the field delivered the
most cost-effective approach and yielded the best results. The design
team produced a course called "Learning Through Others," delivered on
the Charlottesville campus. According to IRS officials, this course
surpassed agency needs and expectations and was less expensive than a
direct contracting arrangement with an outside vendor. They added that
participants in the training program could learn public service values
through lessons and encounter competency-based experiential learning,
business-related challenges, and a capstone simulation. IRS officials
characterized the course as high quality and said the prestige
associated with studying at the Federal Executive Institute provided an
additional benefit for IRS participants.
Lesson learned: Establish mechanisms and controls to avoid unnecessary
duplication or inconsistency within and across agencies' training
efforts.
The agencies used both centralized and decentralized approaches by, for
example, centrally managing reporting and record keeping while allowing
some localized management of training content. Whatever mix of
centralized and decentralized approaches is used, agencies recognize
that it is important to limit overlap and duplication and ensure the
delivery of an integrated message when appropriate. VHA and FWS
officials found that establishing mechanisms and controls is important
to limit duplication or inconsistencies within an agency, across
component organizations within a department, or across the federal
government as a whole.
VHA's Employee Education System helped limit duplication of effort:
VHA's Employee Education System, which serves as an internal training
consulting team within VHA, assists the agency's 21 regional networks
in designing and implementing programs to develop general and specific
skills for VHA employees. Within each of the 21 regional networks, an
Education Service Representative acts as a liaison in coordinating
numerous developmental programs with VHA headquarters--sharing
information with their counterparts about effective practices and
identifying areas of possible duplication or inconsistency across VHA.
According to VHA officials, the coordination and communication achieved
through this organizational structure has helped ensure consistency in
implementing the agency's national training priorities. For example,
officials said that the consulting team assisted in implementing
changes to VHA's processes for collecting third-party insurance
reimbursements in the wake of legislation that required VA to make
greater efforts to collect unpaid debts from veterans. They told us
that these legislative changes enabled local VHA facilities to receive
these reimbursements, but also overwhelmed the local billing and debt
collection processes. To address the problem and help ensure
consistency across the agency, the training consulting team
participated in redesigning the processes for coding, billing, and debt
collection; trained the employees responsible for billing and debt
collection in the new processes; and created graphical representations
of the new processes and posted them throughout the agency's facilities
to aid employees in learning.
Interior's Training Directors Council facilitated communication across
departmental components:
Interior used its Training Directors Council to facilitate
communication across the department's different bureaus, thus helping
to minimize duplicative training and development efforts. This council
provides opportunities--through formal meetings and informal
communications--for training managers from Interior's various bureaus
to share curriculum and related training ideas with their colleagues.
The director of the training center at FWS, who chairs the council,
told us that on more than one occasion he has discovered through
council business that other Interior components had developed
strategies or solutions to address emerging or existing needs that
FWS's training center had also identified. For instance, the training
center at FWS was considering whether to add a new course on grants
management to its curriculum. Through its participation in the council,
FWS determined that another component agency, the Bureau of Land
Management, already offered grants management courses through its
National Training Center in Phoenix, Arizona. After reviewing the
content of these Bureau of Land Management courses, officials at FWS's
training center determined that they did not need to design and develop
a separate grants management course. Instead, FWS and the Bureau of
Land Management now jointly manage the delivery of this training
course.
Lesson learned: Develop and use criteria for determining the optimal
mix of delivery mechanisms to use in order to select the most effective
approaches given each learning situation.
In response to emerging demands and the increasing availability of new
technologies, agencies are faced with the challenge of choosing the
optimal mix of training delivery mechanisms to design training that is
as effective and efficient as possible. Agency officials consider a
wide variety of instructional approaches to achieve learning--in the
classroom, through distance learning, or through structured on-the-job
experiences. Officials also took other factors into account, including
whether to provide individualized instruction or team-based training
and when to use blended learning that combines different teaching
methods (e.g., Web-based and instructor-led) within the same training
program. USACE found that identifying and systematically using criteria
to help select effective delivery mechanisms assisted in building well-
supported justifications for the design of training and development
programs.
USACE used criteria to select media and method of instruction:
To select the appropriate media and method of instruction for its
training programs, USACE uses criteria contained in the Corps of
Engineers Systems Approach to Training, the agency's documented process
for developing training programs. The criteria include issues such as
the expected frequency of changes to the training content, the size and
diversity of the target population, and the degree of student
interaction required. USACE officials told us that, using these
criteria, course managers from USACE's training center coordinate with
relevant agency program offices and subject matter experts to decide on
the appropriate mode for training delivery. While most of the training
center's courses occur in a conventional classroom setting, agency
decision makers have focused on trying to identify courses (or modules
of courses) to convert from classroom training to more economical modes
of delivery, such as distance learning, computer-assisted instruction,
computer-based instruction, or a combination of such approaches. USACE
officials said that many of their courses now incorporate CD ROM and
Internet-based materials as prework assignments before attending
classroom training and for reference use during and after the training
events.
VHA used a profiling tool to help in selecting delivery approaches:
VHA's internal training consulting team used a training delivery
strategy tool that consists of a series of questions structured to
guide users through the process of selecting an appropriate delivery
approach for a proposed training effort. This team designed the
training delivery strategy tool to help staff plan, analyze, develop,
and deliver training and development activities. To aid in decision
making, the tool includes factors such as audience composition, course
goals and objectives, course modules, any prerequisites, participant
preparation, and course follow-up and evaluation. VHA officials said
that applying this tool and analyzing the resultant profile helps
maximize learner understanding, retention, and application.
Lesson learned: Ensure that employees have the needed equipment and
technologies so that they can take maximum advantage of learning
opportunities.
Many organizations are taking advantage of more flexible design and
delivery methods made possible by technology to, for example, deliver
training to the user's desktop, thereby making training more accessible
and cost effective. As agencies move forward in using new approaches,
it is important to ensure that employees have the needed equipment and
technology to take maximum advantage of learning opportunities.
IRS converted mandatory training courses to an online format:
IRS recently converted a series of mandatory training courses from
face-to-face group briefings to an online format in order to more
efficiently provide this training to its employees. Through these
online mandatory training programs, which include computer security
awareness, ethics issues, and prevention of sexual harassment, IRS
wanted to (1) reduce the burden of managers who previously had to
prepare for and deliver the training, (2) provide ready access of the
information to line employees when and where they need it, and (3)
lower the costs associated with the group briefings. IRS officials said
that online delivery lessened employee time in taking the training from
approximately 6 to 2 hours and in some cases eliminated the need for
travel. To ensure the accessibility and usability of these online
briefings, IRS worked to resolve various challenges in the conversion,
such as designing the online product for the lowest computing
capabilities of the bulk of the trainee population and providing
alternative delivery mechanisms for individuals who were without a
computer or Intranet access. The officials said that they also learned
it is important to design the online briefings based on a common
template and style guide to standardize their look and feel, provide
online text-only versions of the training for persons who are visually
impaired, and test the usability of the online briefings with end users
on a range of equipment and allow sufficient time for needed revisions.
OPM initiated a pilot program for its employees to use online
courseware:
To explore opportunities for increased use of e-learning approaches,
OPM established a pilot program that allowed about 250 of its employees
access to approximately 1,800 online courses through the Department of
Transportation's Transportation Virtual University. OPM's training
unit worked with offices throughout OPM to identify employees to
participate in the pilot program. OPM officials said that all employees
selected for the pilot program had access to the equipment and
technology needed to make use of the University's online training, by
using a computer either at the employee's desktop or at some central
location. As a result of this pilot, OPM enhanced its offerings of
online training for its employees by becoming a partner in the
GoLearn.com initiative, a governmentwide online training center for
federal employees. OPM officials said the pilot program also showed
that not all employees function effectively in an online training
environment and that some employees need a more structured format in
order to learn.
Lesson learned: Plan early when developing integrated solutions that
complement other planned and ongoing strategies to improve performance
so that when implemented the strategies work effectively and are
aligned to achieve agency goals.
When designing training and development programs, the agencies
sometimes considered how they could integrate them with other
strategies to improve performance and meet emerging demands. If the
work environment is not conducive to providing opportunities to use new
skills or work in different ways, no matter how good the training
program is it may not be effective or successful in terms of changing
on-the-job performance. In addition, training and development programs
represent a significant investment of resources (including time and
money) and may:
not always be part of an appropriate solution. The agencies developed
integrated solutions that included developing and using job aids,
performance support tools, and other approaches to enhance knowledge
management[Footnote 13] and to aid employees on the job as a complement
to training.
OPM planned for an electronic support tool to aid agency employees in
using a new computer system to process retirement claims:
As part of its effort to reengineer and modernize its processing and
support of federal employees' retirement claims, OPM is developing
plans for an integrated Electronic Performance Support System to aid
the agency's benefits specialists in using a new computer system.
Procedural and information job aids are to be built directly into the
software to provide documentation and guidance, "just in time"
assistance, and error detection. This is intended to be an integrated
system to permit coordination between different modes of training and
enhance the learning and performance of the OPM employees working with
reengineered business processes and the new computer system. OPM
officials said, for example, that this system would assist employees in
completing steps using actual data and circumstances of a particular
case they were working on rather than consulting a manual or using data
put together just for training. According to OPM officials, as the
focus under modernization shifts from processing claims to providing
customer service, this system will help employees working in OPM's
retirement program to interact more directly with program participants
to answer questions and solve problems about retirement issues.
USACE identified online solutions to help enhance and integrate
training efforts:
As a complement to the training and development programs it offers to
its employees, USACE recently entered into a joint project with the
Department of Labor to use an online knowledge management system called
Workforce Connections. This system, which resulted from a memorandum of
understanding promoting cooperative efforts between the departments of
Defense and Labor, will provide the USACE workforce with on-demand,
online access to job aids, performance support materials, and:
course content 7 days a week, 24 hours daily. The system will feature
development and maintenance of online communities of practice to
support knowledge management of USACE's Learning Network, which is
USACE's overall platform for delivering a wide variety of learning
resources to agency employees.[Footnote 14] Another part of the
learning network is USACE's Virtual Campus, a distance learning site
that allows employees access to Web-based courses and training events.
Another component of the learning network includes electronic
performance support tools, such as job aids and other information
resources. USACE officials said that they consider the systems in the
learning network to comprise a powerful solution that effectively
integrates the agency's training efforts.
Lesson learned: Plan for the direct participation of senior agency
leaders and experienced staff in the delivery of training and
development programs to increase buy-in and build support for
organizational change.
Internal resources, such as subject matter experts and high performers,
can often provide valuable insight into training design because of
their familiarity with the agency's policies, programs, and corporate
culture. To increase buy-in, help establish greater credibility, and
build support for organizational change, the agencies have learned the
value of planning for the direct involvement of senior managers in the
training program.
IRS and OPM involved executives and managers:
IRS officials told us that a key feature of the agency's frontline
managers course is that it was designed to use senior managers and
experienced frontline managers drawn from the agency's business units
to teach the course. In addition, executives participate in course
modules that focus on emerging issues facing the agency. The deputy
commissioner of IRS's Wage and Investment business unit served as the
executive sponsor of the training program and participated in course
modules featuring executives. The officials also told us that IRS
executives partner with outside vendors to serve as an instructor team
to deliver all courses designed for senior managers. Using business
unit executives and managers as course instructors helped ensure that
the course's content and emphasis related to the mission, goals, and
guiding principles of the agency.
At OPM, agency managers have been the first to take special initiative
training--such as courses on prohibited personnel practices,
whistleblower procedures, and information technology security--before
they are offered agencywide. OPM officials said that teaching the
material to agency managers in advance of line employees enables the
managers to model desired behaviors and learning for their employees
and convincingly convey how they personally benefited from the
training.
FWS and VHA relied on in-house experts:
FWS's training center brings in FWS field office personnel when
building a cadre of senior, in-house instructors. Training center
officials said that involving trainers from the field helps to build
trust with trainees and provides an added level of credibility that
neither academics nor other subject matter experts who lack field
experience can easily replicate. According to these officials, many
expert employees come from the field and stay to teach at the center
for 3 or 4 years. Some, however, teach only one or two courses or get
involved for a short duration before returning to their positions in
the field. The director of the training center said that he views this
passing-on of information from seasoned veterans to less experienced
employees as crucial for maintaining the unique knowledge base of the
agency.
VHA used "super users" to teach medical center personnel to use its
computerized patient record system, a computer interface that allows
hospital personnel to keep more comprehensive patient records and
enables clinicians, managers, and other staff to review and analyze
data gathered on any patient. The super users--VHA employees with other
job-related duties and responsibilities--were trained to be thoroughly
knowledgeable about the system so they could demonstrate its
capabilities and directly relate the training to employees' work. VHA
initiated this strategy when the agency began rolling out the
application in 1997. VHA officials said that planning to build on the
direct involvement of these super users was successful because they
served as first-line resources for employees' questions about the new
system and helped the agency to build organizational support for the
system. According to VHA, the agency developed a cadre of more than
2,500 super users, and about 180,000 VHA employees use the patient
record system.
Agencies are Considering More Sophisticated Evaluation Approaches As
Part of Designing their Training and Development Programs:
Without evaluation of training programs, participants may take
ineffective courses that do not provide the necessary learning
experience or that do not translate to improved performance on the job.
Overall, the five agencies in our review relied primarily on standard
end-of-course evaluations to obtain the participants' reaction to, and
satisfaction with, a specific training course or learning opportunity.
Although the agencies encountered challenges given some of the
difficulties associated with measuring the impact of training on
individual and organizational performance, they have begun or are
planning to use more comprehensive and sophisticated evaluation
techniques for assessing their training and development efforts. Such
techniques include the use of pre-and post-testing to determine the
extent of learning accomplished, tracking the performance or
advancement of individuals and work units before and after training is
completed to assess professional growth and improvements in
organizational performance, and limited use of ROI analyses to compare
the benefits (quantified in dollars) with the costs of a training and
development program.
To help determine whether the objectives of training and development
are achieved, agencies can begin by incorporating measures of
effectiveness into the design of training and development programs.
Defining objectives in a measurable way enables agency officials to
offer a more convincing business case and contributes to improving the
quality of feedback. Whenever possible, training goals should measure
the individual and organizational results achieved rather than the
training inputs or outputs (e.g., number of available courses or people
trained). Figure 7 depicts some of the steps involved in determining
the evaluation methods to use in designing training and development
programs.
Figure 7: Steps in Determining Methods for Evaluating Training
Programs:
[See PDF for image]
[End of figure]
Agencies' Experiences in Determining Methods for Evaluating Their
Training and Development Programs:
Agency officials recognized the importance of determining during design
how they planned to evaluate the effectiveness of their training and
development programs. To collect information on participants' reaction
to and satisfaction with the training program, for example, VHA uses a
standard evaluation form with questions related to program design,
delivery, outcomes, overall satisfaction, and logistics. In completing
this survey, training participants evaluate their success in completing
learning objectives and the performance of the faculty. Other agencies
also obtained participant feedback though interviews or focus groups.
OPM conducted exit interviews with individuals who participated in that
agency's Presidential Management Intern program, a 2-year developmental
program for individuals from a wide variety of academic disciplines who
wish to enter the federal service. These exit interviews were designed
to obtain feedback from the participants about their overall
satisfaction and reaction to the Presidential Management Intern
program, including their suggestions for enhancing the training
provided.
The five agencies have begun to use, or are planning to use, more
comprehensive and sophisticated evaluation techniques to assess the
extent to which training and development programs increased employees'
knowledge and skills or enhanced individual and organizational
performance. One of these more sophisticated evaluation techniques is
the use of pre-and post-testing to determine the extent of learning
during the training program. USACE's training center conducts pre-and
post-tests on over 90 percent of the courses it offers and is working
toward the goal of using such tests for all courses. The agencies also
tracked job performance and the advancement or movement of personnel to
assess the potential effectiveness of training. FWS officials told us
they track participants' career advancement to determine the extent to
which participation in the leadership development program for midlevel
employees contributed to increased mobility into more responsible
leadership positions in the agency. According to training center
officials, about 37 percent of the program graduates have taken either
promotions or new lateral assignments since the program's inception in
January 2002.
In addition, some agencies attempted to conduct ROI analyses to compare
the benefits (quantified in dollars) to the costs of a particular
training and development program. VHA officials pointed to concerted
efforts to conduct ROI analyses on several training and development
programs, including customer service, leadership development, and
computer-based training. IRS officials, on the other hand, have decided
that the challenges and difficulties in conducting such analyses are
not worth the effort for the resultant information--given the challenge
of isolating the performance improvements that might result from a
specific training activity and the difficulty in monetizing identified
benefits in order to calculate the ROI. IRS instead uses the concept of
"time to capability" to determine whether and to what extent a training
course, program, or other training intervention has improved the
organization's ability to perform its mission successfully. IRS defines
time to capability as the validated accumulation over time of employees
who have been trained in specific competencies deemed critical to the
success of an organizational unit. Under this approach, when IRS has
trained a predetermined number of employees, officials consider that
the agency has achieved the goal of training a critical mass within its
workforce and conclude that the agency has an organizationwide
capability in the specific competencies.
Agencies' Lessons Learned in Determining Methods for Evaluating Their
Training and Development Programs:
Agencies' training and development efforts involve a continuous effort
throughout planning, design, implementation, and evaluation.
Therefore, it is important to recognize that evaluation is not a static
requirement to be carried out after the fact. When undertaking design
and development of training, agencies can rely on evaluations and
benchmarking to determine what approaches work best given all the
related elements, such as the proposed audience for the training
program, the material to be covered, and possible delivery mechanisms
that could be employed. Determining methods for evaluating training
programs as part of their design can help identify and remove obstacles
to successful implementation. For example, an agency officials said
that catching potential problems early on saved valuable time and
resources that a major redesign of training later on likely would have
entailed. On the basis of our review at the five agencies, we
identified four lessons learned regarding the agencies' efforts to
determine methods for evaluating their training and development
programs.
Lesson learned: Incorporate appropriate aspects of the evaluation
approach when designing training and development programs by specifying
what results are expected to better ensure the availability and use of
quality performance data.
In assessing how and to what degree performance could be improved with
a specific training program, agencies should try to establish a
targeted level of improved performance as well as assess the possible
consequences if the training were not to occur. Determining a target
level for improved performance can aid agencies in assessing whether
the expected costs associated with the proposed training are worth the
anticipated benefits. Agreeing upon the planned evaluation approach in
the design clearly sets forth the results the agency expects to achieve
through the training. In addition, planning ahead helps ensure the
availability and use of sound and relevant performance data.
VHA decided to evaluate a customer service training program after it
was implemented:
In response to low scores on customer satisfaction surveys, a VHA
regional network office pursued various initiatives to improve customer
service, including the design, development, and implementation of a new
training program focused on creating a more customer-service-oriented
culture and improving employee morale and collaboration to better meet
customer needs. After delivery of this new training program, called
"The Customer," VHA selected a contractor to evaluate the effectiveness
of the training effort. In its report, the contractor stated that
because the opportunity to conduct the evaluation did not occur until
after the training program had been delivered, the use of preferred
evaluation methods was not possible. The contractor reported that this
lack of preprogram planning had also been experienced in some other VHA
network offices. Nevertheless, the contractor evaluated the customer
service training program by surveying participating employees and their
supervisors. From these survey results, the contractor concluded that
the training program was viewed as successful for those who needed it,
but that the training did not receive a sufficient level of support
from supervisors. The contractor recommended that VHA obtain additional
feedback from supervisors as well as from an individual hired to
telephone or visit the VHA facilities in the network office to observe
customer service activities.
Lesson learned: Consider new approaches for collecting and analyzing
performance data with the aim of increasing the quality and quantity of
training evaluation feedback.
The agencies faced various challenges in obtaining a high quality and
quantity of feedback needed to evaluate their training and development
programs. We previously reported that low participation on the part of
employees and managers in surveys and focus groups may limit an
agency's access to the data needed to complete valid and useful
evaluations of training programs.[Footnote 15] With strong agency
support and proper planning, stakeholders, including training
participants, supervisors, managers, and trainers, are more likely to
provide the information and feedback needed to successfully and
effectively evaluate agency training and development programs. USACE,
for example, recognized that it needed to ensure that it incorporated a
wider variety of stakeholder perspectives in assessing the impact of
training on employee and agency performance. Stakeholders' perspectives
can be obtained through surveys and questionnaires, individual or group
interviews, or communication with more formal multidisciplinary bodies
such as advisory or education councils. Valuable sources of information
include the training participants; training designers, developers, and
facilitators; agency leaders, managers, supervisors, subordinates, and
coworkers; employee organizations; internal and external customers; and
functional and subject matter experts.
USACE's training center altered its approach to obtaining supervisory
feedback on its training efforts:
USACE's training center has been attempting to obtain more
sophisticated evaluation feedback to determine if its training courses
affected employees' behavior on the job. Training center officials told
us that they originally had planned to evaluate all of the center's
training courses by obtaining feedback from employees and supervisors 6
months after the training course on the extent to which employee on-
the-job behavior had changed. In a test of this approach, the training
center sent out hard copies of two generic survey forms: one for the
employee and one for the supervisor. Because the training center had
locator information for trainees but not for their supervisors, both
surveys were mailed to the employee, who was then asked to forward one
to his or her supervisor. Training center officials said the response
to this survey effort was disappointing. For one 35-person class, for
example, the center received 3 employee surveys and 1 supervisor
survey. To increase the level of feedback, officials told us that the
center is transitioning to an electronic process whereby the survey
forms are sent via e-mail to the employee, who is then requested to
forward the survey via e-mail to his or her supervisor. Training center
officials said that in a recent test of this new approach on one
course, the center received a 67 percent response rate from employees
and a 36 percent response rate from supervisors. The officials told us
that they are working to develop unique evaluation forms for each
course in order to obtain feedback on specific learning objectives
rather than rely on a generic survey form for all courses.
Lesson learned: Plan for the use of multiple data types and sources to
provide a balanced approach in assessing the effectiveness of training
and development programs.
Successful organizations typically develop and implement human capital
approaches based on a thorough assessment of the organizations'
specific needs and capabilities. Valid and reliable data are the
starting point for such assessments. To assess the results achieved
through training and development, agencies can rely upon hard
(quantitative) data, such as productivity/output, quality, costs, and
time, or soft (qualitative) data, such as feedback on how well a
training program satisfied employees' expectations. By taking steps to
agree on measures of success up front, agency officials can decide on
the objectives for each training and development program. Using a
balanced approach that reflects feedback from customers and employees,
as well as organizational results, is particularly important as
agencies transform their cultures and operations. In addition, because
the work of federal employees can be complex and often cannot be
reduced to a single task, a balanced approach to both the types and
sources of data helps to strengthen the linkages between training and
development programs and improved performance.
USACE conducts evaluations of some training courses:
In addition to obtaining end-of-course participant feedback and
administering pre-and post-tests for many of its courses, USACE's
training center also conducts course evaluations of its offerings. To
conduct this evaluation, the designated course manager from USACE's
training center is responsible for observing the full course and
assessing various aspects of its design and delivery, including the
training content, materials, and instructors. Training center officials
said that various factors can trigger the decision to conduct an
overall course evaluation, including a significant decline in trainees'
overall satisfaction with the course, the introduction of new
instructors, the use of contractor assistance, or a specific
recommendation from an agency office or unit. According to training
center officials, the results of this course evaluation are assembled
with the end-of-course participant survey feedback and pre-and post-
testing results to present a comprehensive and balanced view of the
effectiveness of the training program.
Lesson learned: Take into account all relevant factors for determining
the costs of a training and development program to better ascertain
whether it is cost-effective in relation to benefits achieved.
Calculating the ROI for a training program involves identifying and
monetizing the program's benefits and then dividing this by a full
tabulation of the program's costs. These costs should usually include
the cost of program materials provided to each participant; the cost of
the facilities; the costs of the facilitator or instructor, including
time for both preparation and delivery; any travel-related expenses for
participants; salaries and benefits costs of the participants for the
time they attend the program; and an allocation of relevant
administrative and overhead costs.
VHA's evaluation of a leadership development program did not include
all costs:
In cooperation with VHA's internal training consulting team, one of
VHA's regional network offices designed, developed, and implemented a
networkwide leadership development program called "Competency
Development for Leaders in the 21st Century." According to information
we gathered during our review, the costs incurred for the consulting
team's efforts on this training program were not included in the ROI
calculation even though it contributed substantially toward developing
and implementing the program. VHA officials told us that these costs
were not included in the analysis because the course designers only
wanted to determine the return on the network's investment, not the
agency's overall investment. Agency officials said that one of the main
goals of involving the consulting team was to help the network develop
the capability to use the ROI process to evaluate training. Although
these costs were not included in the ROI cost tabulation, VHA did
include the salaries (plus employee benefits) of the participants for
the time they attended the developmental program. As we noted in our
recently issued assessment guide, agencies might overlook the costs of
participant attendance when calculating the total costs of a training
program.[Footnote 16]
Conclusions and Observations:
Federal agencies are faced with the need to invest resources wisely to
ensure that their employees possess the information, skills, and
competencies required to carry out their work successfully. The
examples provided in this report may help to address this need by
describing some of the experiences and lessons learned that other
agencies might find applicable or adaptable to their unique situations.
This information is intended to both provide a realistic perspective on
how agencies have approached designing their training and development
programs to date as well as to take a more detailed look at some of the
concepts explored in our recently issued assessment guide focused on
strategic training and development efforts in the federal government.
Our work reviewing the selected agencies' efforts to design training
and development programs reinforces the significance of good planning
and design of these programs to ensure their successful implementation
and evaluation. The experiences and lessons learned we identified also
demonstrate how effective design efforts--as part of a strategic
training and development process--rely on the eight core
characteristics that we identified in our earlier work: (1) strategic
alignment, (2) leadership commitment and communication, (3) stakeholder
involvement, (4) accountability and recognition, (5) effective resource
allocation, (6) partnerships and learning from others, (7) data quality
assurance, and (8) continuous performance improvement. Indeed, by
focusing on these eight core characteristics, agencies can improve not
only the design of their training and development efforts but also the
planning, implementation, and evaluation of their programs to better
ensure that their employees have the information, skills, and
competencies needed to carry out their work successfully.
Agency Comments and Our Evaluation:
We provided a draft of this report on December 12, 2003, to the
Secretary of Defense, the Secretary of the Interior, the Commissioner
of Internal Revenue, the Director of OPM, and the Secretary of VA.
Interior, IRS, and OPM provided written comments on the draft report.
In his written comments (see app. IV), Interior's Assistant Secretary
for Fish and Wildlife and Parks generally agreed with the report's
findings regarding the Department and the FWS. He said that the report
provides important examples that can help the Department continue to
move forward with additional confidence in its actions. In his written
comments (see app. V), the Commissioner of Internal Revenue said that
IRS was honored to share some of its lessons learned with us for
governmentwide dissemination. He said that our review also provides IRS
with practices from other agencies to assist IRS in its efforts to
continually improve its programs. In her written comments (see app.
VI), the Director of OPM said that she appreciated the opportunity for
OPM to be included in the report and to share information on OPM's
training and development activities and programs. Interior, IRS, and
OPM also provided technical comments to clarify specific points
regarding the information presented in the draft report, which we have
incorporated as appropriate. In comments by E-mail through its GAO
liaison, VA agreed with the information presented regarding the
Department and had no additional comments on the draft report. USACE
informed us that it had no comments on the draft report.
As agreed with your office we are sending copies of this report to the
Ranking Minority Member, Subcommittee on Oversight of Government
Management, the Federal Workforce and the District of Columbia, Senate
Committee on Governmental Affairs; the Chairmen and Ranking Minority
Members of the Senate Committee on Governmental Affairs and the House
Committee on Government Reform; and other interested congressional
parties. We are also providing copies to the Secretary of the Army, the
Secretary of the Interior, the Secretary of the Treasury, the Director
of OPM, and the Secretary of Veterans Affairs. This report is available
to others upon request. In addition, the report is available at no
charge on the GAO Web site at [Hyperlink, http://www.gao.gov] http://
www.gao.gov.
If you have any questions about this report, please contact me on (202)
512-6806. Key contributors to this report are listed in appendix VII.
Sincerely yours,
Signed by:
George H. Stalcup:
Director, Strategic Issues:
[End of section]
Appendixes:
Appendix I: Objective, Scope and Methodology:
The objective of this review was to provide information on selected
federal agencies' experiences and lessons learned in key aspects of
designing training and development programs for their employees.
Specifically, we focused on the agencies' experiences and lessons
learned related to:
* assessing current and future agency skill and competency requirements
and identifying related training and development needs,
* developing strategies and solutions for training and development
needs, and:
* determining methods to evaluate the effectiveness of training and
development programs.
For this review, lessons learned were defined as knowledge that could
be applied in the future that the agencies gained through either
positive or negative experiences.
To address this objective we focused on five agencies: the U.S. Army
Corps of Engineers (USACE), Department of Defense; the U.S. Fish &
Wildlife Service (FWS), Department of the Interior; the Internal
Revenue Service (IRS), Department of the Treasury; the U.S. Office of
Personnel Management (OPM); and the Veterans Health Administration
(VHA), Department of Veterans Affairs. We chose the five agencies for a
variety of reasons, including the diversity of employee occupations
within the agencies, reported innovative approaches for training and
developing their employees, and congressional requester interest. We
selected USACE to obtain a Department of Defense perspective in light
of that department's reputation as a leader in the area of training and
developing military personnel. We included FWS and VHA to obtain
information related to a broad mix of employee occupations. We selected
IRS because of reported innovative approaches to training and
development and included OPM because of its role as the federal
government's human capital agency.
To obtain information and related documentation, we visited the
following locations:
* USACE's headquarters human resources directorate in Washington, D.C.,
and its Professional Development Support Center in Huntsville, Alabama.
* FWS's headquarters human resources division in Arlington, Virginia,
and its National Conservation Training Center in Shepherdstown, West
Virginia.
* IRS's strategic human resources division in Arlington, Virginia, and
the small business and self-employed business line.
* OPM's headquarters human resources office and the career development
branch of the center for retirement and insurance service.
* VHA's Employee Education System headquarters in Washington, D.C., and
the medical center and local network offices in Durham, North Carolina.
It is important to note that our methodology was not designed to
identify examples that would be representative of all training and
development efforts at the five agencies in our review or of the
government as a whole. We did not verify the accuracy and reliability
of the data provided to us or the systems used to produce the
information. Further, in citing training and development programs as
examples in connection with lessons learned, we did not assess the
effectiveness of the training programs and practices. Rather, our
intent was to highlight and briefly describe some experiences and
lessons learned that agency officials believed helped each agency
improve or enhance its training and development programs.
To obtain information about the five agencies' experiences and lessons
learned related to designing training and development programs, we:
* Interviewed agency human capital and training officials and subject
matter experts responsible for agency training, performance, and other
initiatives; and:
* Reviewed and analyzed agency documents such as workforce plans,
analyses, and reports; strategic, performance, and succession plans and
reports; organizational, occupational, and unit-based competency
standards; knowledge and skills inventories; skills gaps assessments;
competency and skill assessments; surveys of agency employees; training
plans and proposals; workforce demographic data; budget data;
evaluation plans and reports; and performance measures.
The lessons learned we identified for inclusion in this report were
based on (1) their linkages with one or more of the eight core
characteristics of a strategic training and development process, which
we had identified in our previous work (see app. III) and (2)
sufficient evidence from the agency to support the experiences that
they relayed to us. At the exit conferences for the five agencies, we
presented agency officials with the list of lessons learned that we had
identified and wished to attribute to their experiences. At that time,
we also informed each agency of the specific examples from their
experiences that we would likely attribute to these lessons learned. In
these meetings, agency officials expressed no objections to the lessons
learned we had identified and in some cases provided additional
information to support specific examples from their experiences that we
proposed to use for this report.
We conducted our audit work between August 2002 and November 2003 in
accordance with generally accepted government auditing standards.
[End of section]
Appendix II: Background on Selected Agencies and Their Training and
Development Functions:
The following summarizes key information on the five agencies included
in this review. These summaries include information on the agencies'
missions, organizational structures, and training and development
functions.
U.S. Army Corps of Engineers:
The U.S. Army Corps of Engineers (USACE), part of the Department of the
Army within the Department of Defense, is comprised of approximately
34,600 civilian and 650 military men and women. USACE has a diverse
workforce consisting of military and civilian engineers, biologists,
geologists, hydrologists, natural resource managers, and other
specialists who work in engineering and environmental matters. USACE's
mission is to provide engineering services to the nation: (1) planning,
designing, building and operating water resources and other civil works
projects, (2) designing and managing the construction of military
facilities for the Army and Air Force, and (3) providing design and
construction management support for other Department of Defense and
federal agencies. USACE headquarters office is located in Washington,
D.C.
USACE's Professional Development Support Center, located in Huntsville,
Alabama, serves as the center of learning and training for the agency.
The training center manages and implements the Proponent-Sponsored
Engineer Corps Training program, which provides job-related training
through technical, professional, managerial, and leadership courses for
USACE and other government agencies. USACE's training center offers
more than 200 courses covering topics that support the agency's
mission.
U.S. Fish and Wildlife Service:
The mission of the U.S. Fish and Wildlife Service (FWS), a component of
the Department of the Interior, is working with others to conserve,
protect, and enhance fish, wildlife, and plants and their habitats for
the continuing benefit of the American people. FWS's headquarters is
located in Washington D.C., while its field units are located
throughout the United States. FWS employs more than 9,600 people and is
supported by a volunteer force of 29,000. Nearly 90 percent of FWS
employees work in field locations.
The National Conservation Training Center, located in Shepherdstown,
West Virginia, is FWS's training center and is responsible for training
a wide range of employees in the conservation community and serves as a
gathering place where conservation professionals from government,
nonprofit organizations, and corporations work toward common goals.
Training for FWS's law enforcement personnel is primarily conducted
through the Federal Law Enforcement Training Center, an interagency law
enforcement training organization headquartered in Glynco, Georgia.
Internal Revenue Service:
The Internal Revenue Service (IRS) is a branch of the U.S. Department
of the Treasury. IRS's mission is to provide America's taxpayers top
quality service by helping them understand and meet their tax
responsibilities and by applying the tax law with integrity and
fairness to all. IRS's organizational structure includes the following
business units: four operating divisions organized around four major
customer segments (Wage and Investment, Small Business/Self-Employed,
Large and Mid-Size Business, and Tax Exempt and Government Entities);
four functional divisions (National Taxpayer Advocate, Appeals,
Criminal Investigation, and Communications and Liaison); and two shared
services/support divisions (Agency-Wide Shared Services and
Modernization and Information Technology Services). As of March 2003,
IRS had about 116,300 employees.
IRS takes a decentralized approach to training and developing its
workforce. Each business unit has an embedded human resources component
that provides advice and analysis on related policies and issues and
formulates strategies, procedures, and practices to address the unit's
human capital needs. Learning and Education, one of eight major
divisions comprising IRS's Office of Strategic Human Resources,
provides guidance and sets policy and standards on training and
development for the agency's business units and headquarters offices.
U.S. Office of Personnel Management:
The U.S. Office of Personnel Management (OPM), the federal government's
human capital agency, provides human resources policy leadership,
technical advice and assistance, and products and services to federal
agencies, employees, annuitants, and job seekers. It also oversees
governmentwide compensation and performance management systems, and
provides retirement, health benefit, and other insurance services to
federal employees, annuitants, other beneficiaries, and agencies. In
March 2003, OPM completed a major restructuring process through which
it consolidated various agency functions. As of March 2003, OPM
employed approximately 3,500 people, many of them stationed in agency
headquarters in Washington, D.C. OPM has a field presence in 16 major
U.S. cities as well as operating centers in Pennsylvania and Georgia.
OPM's training and development efforts are largely decentralized to the
agency's various program and staff offices. The employee training and
development unit within the agency's human capital management office is
responsible for setting overall strategy and for planning and
implementing agencywide training such as leadership development
programs and various mandatory training programs. According to OPM, the
agency's newly established Chief Human Capital Officer plays a
significant role in advising the OPM Director on overall employee
training and development initiatives and programs, as well as the
establishment of the agency's training budget.
Veterans Health Administration:
The Veterans Health Administration (VHA), one of three major
administrations within the Department of Veterans Affairs (VA), is
responsible for providing primary care, specialized care, and related
medical and social support services to veterans through an integrated
health care system. VHA administers its functions through a group of 21
regional network offices located around the United States. As of March
2003, VHA employed about 203,500 people out of a total VA workforce of
about 225,000 employees.
VA takes a decentralized approach to training and development
operations. VA's human resources office provides advice and guidance on
training to VHA and the other departmental components but delegates
training and development operations to each component. VHA's
organization includes the Employee Education System, which is an
internal training consulting group that provides educational services
that support the workforce development and continuing education needs
for VHA employees. This internal consulting group of about 300
individuals primarily helps to assess agency training needs at the
national level, while VHA network offices and medical centers take lead
responsibility for assessing their own local needs. These internal
training consultants are available to assist VHA network offices and
medical centers in designing, developing, implementing, and evaluating
training and development programs to meet these local needs.
[End of section]
Appendix III: Core Characteristics of a Strategic Training and
Development Process:
The following summarizes the eight core characteristics that make a
training and development process effective and strategically focused on
achieving results. We identified these core characteristics as part of
our recent work in developing an assessment guide to assist federal
agencies in evaluating their training and development efforts.[Footnote
17]
* Strategic alignment. Clear linkages exist between the agency's
mission, goals, and culture and its training and development efforts.
The agency's mission and goals drive a strategic training and
development approach and help ensure that the agency takes full
advantage of an optimal mix of strategies to improve performance and
enhance capacity to meet new and emerging challenges.
* Leadership commitment and communication. Agency leaders and managers
consistently demonstrate that they support and value continuous
learning, are receptive to and use feedback from employees on
developmental needs and training results, and set the expectation that
fair and effective training and development practices will improve
individual and organizational performance.
* Stakeholder involvement. Agency stakeholders are involved throughout
the training and development process to help ensure that different
perspectives are taken into account and contribute to effective
training and development programs. Stakeholders' views are incorporated
in identifying needed performance enhancements, developing and
effectively implementing well-thought-out strategies, and helping to
conceptualize and use balanced measures that accurately reflect the
extent to which training and development efforts contribute toward
achieving results.
* Accountability and recognition. Appropriate accountability
mechanisms, such as performance management systems, are in place to
hold managers and employees responsible for learning and working in new
ways. Appropriate rewards and incentives exist and are used fairly and
equitably to encourage innovation, reinforce changed behaviors, and
enhance performance.
* Effective resource allocation. The agency provides an appropriate
level of funding and other tools and resources--along with external
expertise and assistance when needed--to ensure that its training and
development programs reflect the importance of its investment in human
capital to achieving its mission and goals.
* Partnerships and learning from others. Coordination within and among
agencies achieves economies of scale and limits duplication of efforts.
In addition to benchmarking high-performing organizations, these
efforts allow an agency to keep abreast of current practices, enhance
efficiency, and increase the effectiveness of its training and
development programs.
* Data quality assurance. The agency has established policies and
procedures that recognize and support the importance of quality data
and of evaluating the quality and effectiveness of training and
development efforts. It establishes valid measures and validated
systems to provide reliable and relevant information that is useful in
improving the agency's training and development efforts.
* Continuous performance improvement. Agency practices and policies
foster a culture of continuous improvement and optimal organizational
performance regarding training and other activities. Stakeholders rely
on and use program performance information and other data to assess and
refine ongoing training and development efforts; target new initiatives
to improve performance; and design, develop, and implement new
approaches to train and develop employees.
[End of section]
Appendix IV: Comments from the Department of the Interior:
United States Department of the Interior:
OFFICE OF THE SECRETARY WASHINGTON, D.C. 20240:
JAN 23 2004:
Mr. George Stalcup
Director, Strategic Issues
U.S. General Accounting Office
441 G Street, N.W.
Washington, D.C. 20548:
Dear Mr. Stalcup:
Thank you for providing the Department of the Interior the opportunity
to review and comment on the draft U.S. General Accounting Office
report entitled, "Human Capital: Selected Agencies' Experiences and
Lessons Learned in Designing Training and Development Programs" (GAO-
04-291) dated December 12, 2003. In general, we agree with the report's
findings that pertain to the Department. The report provides important
examples that can help us continue to move forward with additional
confidence in our actions.
The Department has one specific comment. On page 37, the U.S. Fish and
Wildlife Service's mission description should be changed to: "The
mission of the U.S. Fish and Wildlife Service is, working with others,
to conserve, protect and enhance fish, wildlife, and plants and their
habitats for the continuing benefit of the American people.":
We look forward to receiving the final report.
Sincerely,
Signed by:
Craig Manson:
Assistant Secretary for Fish and Wildlife and Parks:
The following is GAO's comment on the Department of the Interior's
letter dated January 23, 2004.
GAO Comment:
We have clarified the mission statement of the Fish and Wildlife
Service to note its collaboration with others to accomplish its
mission.
[End of section]
Appendix V: Comments from the Internal Revenue Service:
DEPARTMENT OF THE TREASURY:
INTERNAL REVENUE SERVICE:
WASHINGTON, D.C. 20224:
COMMISSIONER:
January 12, 2004:
Mr. George Stalcup:
Director, Strategic Issues
United States General Accounting Office
Washington, D.C. 20548:
Dear Mr. Stalcup:
I have reviewed your draft report entitled, "Human Capital: Selected
Agencies' Experiences and Lessons Learned in Designing Training and
Development Programs" (GAO-04-291). Thank you for selecting the
Internal Revenue Service to be a part of this important study. We are
honored to share some of our lessons learned with you for government-
wide dissemination. I am particularly pleased that we could share our
work in leadership development, strategic planning, skills assessments
and use of technology as best practices. The study not only recognizes
some of our accomplishments but also provides us with practices from
other agencies to assist us in our efforts to continually improve our
programs.
Strategic management of human capital is a vital part of moving the IRS
forward. As a result, I have created a new Human Capital Office. Our
Chief Human Capital Officer, Beverly Ortega Babers, will provide the
leadership necessary to focus our resources on critical training and
development programs. We will build on the good planning and design
practices cited in your report to ensure that our employees have the
skills and competencies they need to provide top quality service to
America's taxpayers.
If you have any questions, please contact me or Beverly Ortega Babers,
Chief Human Capital Officer, at (202) 622-7902.
Sincerely,
Mark W. Everson:
Signed for Mark W. Everson:
[End of section]
Appendix VI: Comments from the Office of Personnel Management:
OFFICE OF THE DIRECTOR:
UNITED STATES OFFICE OF PERSONNEL MANAGEMENT
WASHINGTON, DC 20415-0001:
JAN 20 2004:
The Honorable David M. Walker
Comptroller General:
General Accounting Office
Washington, DC 20548:
Dear Mr. Walker:
Thank you for the opportunity to respond to GAO's draft report entitled
Selected Agencies' Experiences and Lessons Learned in Designing
Training and Development Programs. As you pointed out, it is critical
that agencies invest training and development resources wisely to
ensure that their employees have the necessary skills and competencies
to carry out their agency's mission. Within the Office of Personnel
Management (OPM), I am very committed to developing our employees and
making the kinds of strategic investments that will enhance their
ability to contribute to achieving OPM's important mission. This
commitment is evident in OPM's current restructuring whereby OPM is now
organized around mission goals and priorities. The restructuring better
positions OPM to meet the President's plan to reform the Federal
Government by recognizing the importance of talented employees to
deliver a citizen-centered, results-oriented, and market-based
Government.
Furthermore, OPM's Plan for the Strategic Management of OPM's Human
Capital: FY 2004-2007, lays out an integrated and goal specific
approach to ensuring OPM has the talent and capacity to meet our new
and expanded responsibilities.
Since your staff conducted its review of our training and development
programs, we have made enhancements to our programs. These enhancements
are enclosed and are also included in our Plan for the Strategic
Management of OPM's Human Capital: FY 2004-2007, which is also
enclosed.
I appreciate the opportunity for OPM to be highlighted in your report,
and to share the innovative training and development activities and
programs we have created.
Sincerely,
Signed by:
Kay Coles James:
Director:
Enclosures:
Office of Personnel Management Comments on GAO draft report: Selected
Agencies' Experiences and Lessons Learned in Designing Training and
Development Programs:
Page 8 - regarding OPM's 2001 skills assessment...
In addition to the 2001 skills assessment described in this Section, in
late FY 2003, OPM conducted another skills assessment to reflect our
recent restructuring and new strategic priorities. We identified
mission critical occupations and competencies needed in our current and
future workforce, along with initiatives to address any gaps. This
business forecasting was combined with techniques to recruit, select,
place and train employees and ensure an organizational climate that
fosters commitment. This workforce planning initiative was a major
aspect of our recently published Plan for the Strategic Management of
OPM's Human Capital, 2004-2007. The Plan identifies our mission
critical occupations and the most critical competencies which enhance
our ability to create strategic training and development plans to help
us carry out our important mission.
Page 38 - regarding setting of overall training and development
strategy...
OPM's training and development activities were accurately addressed in
this Section, however, since the time that the draft report was
completed, OPM has increased the role of its Chief Human Capital
Officer (CHCO). The CHCO now plays a more significant role in advising
the Director on overall employee training and development initiatives
and programs, as well as the establishment of the agency's training
budget. Under the direction of OPM's CHCO, program offices, with the
assistance from the internal Human Capital Management Office, create
specific training and development strategies and activities that
address mission critical competencies identified in our FY 2003 skills
assessment. This strategic approach, based upon recent workforce data,
better positions OPM to prioritize its training needs, and forecast
funds to support those needs.
[End of section]
Appendix VII: GAO Contacts and Staff Acknowledgments:
GAO Contacts:
George H. Stalcup or Susan Ragland, (202) 512-6806:
Acknowledgments:
In addition to the persons named above, K. Scott Derrick, Gerard Burke,
T.J. Thomson, and Thomas Davies, Jr. made key contributions to this
report.
(450147):
FOOTNOTES
[1] U.S. General Accounting Office, Human Capital: A Guide for
Assessing Strategic Training and Development Efforts in the Federal
Government - Exposure Draft, GAO-03-893G (Washington, D.C.: July 2003).
[2] U.S. General Accounting Office, Human Capital: Key Principles for
Effective Strategic Workforce Planning, GAO-04-039 (Washington, D.C.:
Dec. 11, 2003).
[3] In providing technical comments on a draft of this report, OPM
noted that it conducted another skills assessment in late fiscal year
2003 to reflect its recent restructuring and new strategic priorities.
[4] In previous work, we identified practices used by agencies in other
countries to manage the succession of senior executives and other
employees with critical skills. See Human Capital: Insights for U.S.
Agencies from Other Countries' Succession Planning and Management
Initiatives, GAO-03-914 (Washington, D.C.: Sept. 15, 2003).
[5] An IDP is a written plan, cooperatively prepared by the employee
and his or her supervisor, that outlines the steps the employee will
take to develop knowledge, skills, and abilities in building on
strengths and addressing weaknesses as he or she seeks to improve job
performance and pursue career goals. These individual development plans
are also known as personal development plans, personal training plans,
and individual training plans.
[6] For more information on key practices and implementation steps that
can help agencies transform their cultures, see U.S. General Accounting
Office, Results-Oriented Cultures: Implementation Steps to Assist
Mergers and Organizational Transformations, GAO-03-669 (Washington,
D.C.: July 2, 2003).
[7] U.S. General Accounting Office, Results-Oriented Cultures: Creating
a Clear Linkage between Individual Performance and Organizational
Success, GAO-03-488 (Washington, D.C.: Mar. 14, 2003).
[8] For more information on performance management of senior executives
at IRS, see U.S. General Accounting Office, Results-Oriented Cultures:
Using Balanced Expectations to Manage Senior Executive Performance,
GAO-02-966 (Washington, D.C.: Sept. 27, 2002).
[9] The 360-degree feedback process is designed to provide a manager
direct input from various sources--supervisor, peers, customers, and
subordinates--and to compare those results to a self-evaluation. With
this feedback, managers can identify action items and incorporate them
into their individual performance plans.
[10] 31 U.S.C. 3302.
[11] GIS is a system of computer software, hardware, and data used to
manipulate, analyze, and graphically present a potentially wide array
of information associated with geographic locations.
[12] GAO-03-893G, p. 7.
[13] Knowledge management is an approach to capturing, understanding,
and using the collective body of information and intellect within an
organization to accomplish its mission.
[14] Communities of practice provide an on-line resource for peers to
ask and respond to questions and share knowledge.
[15] U.S. General Accounting Office, Human Capital: Design,
Implementation, and Evaluation of Training at Selected Agencies, GAO/T-
GGD-00-131 (Washington, D.C: May 18, 2000).
[16] GAO-03-893G, p. 69.
[17] GAO-03-893G, p. 75.
GAO's Mission:
The General Accounting Office, the investigative arm of Congress,
exists to support Congress in meeting its constitutional
responsibilities and to help improve the performance and accountability
of the federal government for the American people. GAO examines the use
of public funds; evaluates federal programs and policies; and provides
analyses, recommendations, and other assistance to help Congress make
informed oversight, policy, and funding decisions. GAO's commitment to
good government is reflected in its core values of accountability,
integrity, and reliability.
Obtaining Copies of GAO Reports and Testimony:
The fastest and easiest way to obtain copies of GAO documents at no
cost is through the Internet. GAO's Web site ( www.gao.gov ) contains
abstracts and full-text files of current reports and testimony and an
expanding archive of older products. The Web site features a search
engine to help you locate documents using key words and phrases. You
can print these documents in their entirety, including charts and other
graphics.
Each day, GAO issues a list of newly released reports, testimony, and
correspondence. GAO posts this list, known as "Today's Reports," on its
Web site daily. The list contains links to the full-text document
files. To have GAO e-mail this list to you every afternoon, go to
www.gao.gov and select "Subscribe to e-mail alerts" under the "Order
GAO Products" heading.
Order by Mail or Phone:
The first copy of each printed report is free. Additional copies are $2
each. A check or money order should be made out to the Superintendent
of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or
more copies mailed to a single address are discounted 25 percent.
Orders should be sent to:
U.S. General Accounting Office
441 G Street NW,
Room LM Washington,
D.C. 20548:
To order by Phone:
Voice: (202) 512-6000:
TDD: (202) 512-2537:
Fax: (202) 512-6061:
To Report Fraud, Waste, and Abuse in Federal Programs:
Contact:
Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov
Automated answering system: (800) 424-5454 or (202) 512-7470:
Public Affairs:
Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S.
General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C.
20548: