Human Capital

Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs Gao ID: GAO-04-291 January 30, 2004

Effective training and development programs are an integral part of a learning environment, helping improve federal workforce performance in achieving agency results. Therefore, in this report GAO was asked to identify examples of selected federal agencies' experiences and some of the key lessons they have learned in designing their training and development programs. This work focused on ways that these agencies (1) assessed agency skills gaps and identified training needs, (2) developed strategies and solutions for these training and development needs, and (3) determined methods to evaluate the effectiveness of training and development programs. GAO worked with five agencies to identify their experiences and lessons learned: the U.S. Army Corps of Engineers (USACE), Department of Defense; Fish and Wildlife Service (FWS), Department of the Interior (Interior); Internal Revenue Service (IRS), Department of the Treasury; the Office of Personnel Management (OPM); and Veterans Health Administration (VHA), Department of Veterans Affairs (VA). Agency officials provided information during interviews and furnished supporting documentation for analysis and review.

GAO identified important lessons learned from five federal agencies' experiences in designing training and development programs for their employees that could be useful to other agencies facing similar challenges. These lessons learned are related to the following three areas. Assessing Agency Skill Requirements and Identifying Training Needs: The agencies used a variety of approaches to assess current and future skill requirements, such as implementing workforce planning models and conducting knowledge and skills inventories. Generally, the agencies are transitioning to more comprehensive approaches. One of the lessons learned was to involve stakeholders and benchmark with others to identify critical skills and competencies and related training needs. For example, IRS officials believed they needed a leadership competency model directly based on the work of their agency's business units. To develop a comprehensive model, they interviewed top IRS leaders and benchmarked with leading practices in the public and private sector. Developing Strategies and Solutions for Training Needs: The agencies considered a mixture of delivery mechanisms, as well as potential sources for training and development opportunities. However, projecting costs and benefits of proposed training and development programs presented challenges for them. The agencies usually developed broad information on anticipated benefits and expected costs of potential investments, although often without tying benefits to specific performance improvements or considering all costs. One of the lessons learned was to establish mechanisms to avoid duplication or inconsistencies. Education Service Representatives in each regional VHA network, for example, coordinate training and development programs with headquarters--sharing information about successful practices and identifying areas where coordination is needed. Determining Methods for Evaluating Training Programs: Overall, the agencies relied primarily on participants' end-of-course evaluations, but they are beginning to use more comprehensive evaluation approaches, including limited use of return-on-investment analysis techniques. One of the lessons learned is to plan for the use of multiple data types and sources in order to attain a balanced assessment once the course is implemented. For example, USACE's training center incorporated pre- and post-tests on over 90 percent of its courses, as well as approaches to collect participants' and course managers' feedback, as part of the design. Four of the five agencies provided comments on a draft of this report. Interior and VA said that they generally agreed with the report's findings regarding their respective agencies. IRS and OPM said that they appreciated the opportunity to be included in the report and to share information on training activities. USACE provided no comments on the draft report.



GAO-04-291, Human Capital: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs This is the accessible text file for GAO report number GAO-04-291 entitled 'Human Capital: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs' which was released on January 30, 2004. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Chairman, Subcommittee on Oversight of Government Management, the Federal Workforce, and the District of Columbia, Committee on Governmental Affairs, U.S. Senate: January 2004: HUMAN CAPITAL: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs: GAO-04-291: GAO Highlights: Highlights of GAO-04-291, a report to the Chairman, Subcommittee on Oversight of Government Management, the Federal Workforce, and the District of Columbia, Committee on Governmental Affairs, U.S. Senate Why GAO Did This Study: Effective training and development programs are an integral part of a learning environment, helping improve federal workforce performance in achieving agency results. Therefore, in this report GAO was asked to identify examples of selected federal agencies‘ experiences and some of the key lessons they have learned in designing their training and development programs. This work focused on ways that these agencies (1) assessed agency skills gaps and identified training needs, (2) developed strategies and solutions for these training and development needs, and (3) determined methods to evaluate the effectiveness of training and development programs. GAO worked with five agencies to identify their experiences and lessons learned: the U.S. Army Corps of Engineers (USACE), Department of Defense; Fish and Wildlife Service (FWS), Department of the Interior (Interior); Internal Revenue Service (IRS), Department of the Treasury; the Office of Personnel Management (OPM); and Veterans Health Administration (VHA), Department of Veterans Affairs (VA). Agency officials provided information during interviews and furnished supporting documentation for analysis and review. What GAO Found: GAO identified important lessons learned from five federal agencies‘ experiences in designing training and development programs for their employees that could be useful to other agencies facing similar challenges. These lessons learned are related to the following three areas. Assessing Agency Skill Requirements and Identifying Training Needs: The agencies used a variety of approaches to assess current and future skill requirements, such as implementing workforce planning models and conducting knowledge and skills inventories. Generally, the agencies are transitioning to more comprehensive approaches. One of the lessons learned was to involve stakeholders and benchmark with others to identify critical skills and competencies and related training needs. For example, IRS officials believed they needed a leadership competency model directly based on the work of their agency‘s business units. To develop a comprehensive model, they interviewed top IRS leaders and benchmarked with leading practices in the public and private sector. Developing Strategies and Solutions for Training Needs: The agencies considered a mixture of delivery mechanisms, as well as potential sources for training and development opportunities. However, projecting costs and benefits of proposed training and development programs presented challenges for them. The agencies usually developed broad information on anticipated benefits and expected costs of potential investments, although often without tying benefits to specific performance improvements or considering all costs. One of the lessons learned was to establish mechanisms to avoid duplication or inconsistencies. Education Service Representatives in each regional VHA network, for example, coordinate training and development programs with headquarters”sharing information about successful practices and identifying areas where coordination is needed. Determining Methods for Evaluating Training Programs: Overall, the agencies relied primarily on participants‘ end-of-course evaluations, but they are beginning to use more comprehensive evaluation approaches, including limited use of return-on-investment analysis techniques. One of the lessons learned is to plan for the use of multiple data types and sources in order to attain a balanced assessment once the course is implemented. For example, USACE‘s training center incorporated pre- and post-tests on over 90 percent of its courses, as well as approaches to collect participants‘ and course managers‘ feedback, as part of the design. Four of the five agencies provided comments on a draft of this report. Interior and VA said that they generally agreed with the report‘s findings regarding their respective agencies. IRS and OPM said that they appreciated the opportunity to be included in the report and to share information on training activities. USACE provided no comments on the draft report. What GAO Recommends: www.gao.gov/cgi-bin/getrpt?GAO-04-291. To view the full product, including the scope and methodology, click on the link above. For more information, contact George Stalcup at (202) 512-6806 or stalcupg@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: Agencies Used Varied Approaches in Assessing Skills and Competencies and Identifying Related Training Needs: Agencies Developed Strategies and Solutions for Their Training Needs: Agencies are Considering More Sophisticated Evaluation Approaches As Part of Designing their Training and Development Programs: Conclusions and Observations: Agency Comments and Our Evaluation: Appendixes: Appendix I: Objective, Scope and Methodology: Appendix II: Background on Selected Agencies and Their Training and Development Functions: Appendix III: Core Characteristics of a Strategic Training and Development Process: Appendix IV: Comments from the Department of the Interior: GAO Comment: Appendix V: Comments from the Internal Revenue Service: Appendix VI: Comments from the Office of Personnel Management: Appendix VII: GAO Contacts and Staff Acknowledgments: GAO Contacts: Acknowledgments: Figures: Figure 1: Four Components of a Strategic Training and Development Process: Figure 2: Strategic Workforce Planning Process: Figure 3: OPM's Five-step Workforce Planning Process: Figure 4: IRS's Core Management Responsibilities and Leadership Competencies: Figure 5: Competencies in VHA's High Performance Development Model: Figure 6: Steps for Developing Strategies and Solutions for Training and Development Needs: Figure 7: Steps in Determining Methods for Evaluating Training Programs: Letter January 30, 2004: The Honorable George V. Voinovich: Chairman: Subcommittee on Oversight of Government Management, the Federal Workforce, and the District of Columbia: Committee on Governmental Affairs: United States Senate: Dear Mr. Chairman: To effectively address the nation's most urgent priorities and take advantage of emerging opportunities, federal agencies need to continue to build their fundamental capabilities to carry our their work in an environment that is increasingly complex and rapidly changing. To build their capacity, agencies should invest resources wisely to ensure that their employees have the information, skills, and competencies they need to succeed. As you are well aware, these investments must include training and development efforts to continuously enhance the skills and competencies of the federal workforce and improve the quality of agencies' results. As agreed with your office, this report provides information on selected federal agencies' experiences and lessons learned in key aspects of designing training and development programs for their employees. Specifically, we focused on the agencies' experiences and lessons learned related to: * assessing current and future agency skill and competency requirements and identifying related training and development needs, * developing strategies and solutions for training and development needs, and: * determining methods to evaluate the effectiveness of training and development programs. For this review, lessons learned were defined as knowledge that could be applied in the future that the agencies gained through either positive or negative experiences. The experiences and lessons learned from the five agencies we reviewed may well provide some valuable ideas and useful approaches that could be adopted by other federal agencies as they attempt to address ongoing training and development challenges- -particularly those related to the elements within the training process that relate to design and development. To address these issues and as agreed with your office, we focused our review on five federal agencies: U.S. Army Corps of Engineers (USACE), Department of Defense; Fish and Wildlife Service (FWS), Department of the Interior (Interior); Internal Revenue Service (IRS), Department of the Treasury; Office of Personnel Management (OPM); and Veterans Health Administration (VHA), Department of Veterans Affairs (VA). At the headquarters of the five agencies--and in some field locations--we collected and reviewed documents on the agencies' training and development efforts and interviewed officials from the agencies' human capital and training organizations, as well as assorted program offices. We selected the five agencies for various reasons, including your office's interests, the diversity of employee occupations within the agency, and reported innovative approaches for training and developing their employees. This agency selection process was not designed to identify examples that could be considered representative of all training and development efforts at the five agencies or the federal government as a whole. Furthermore, in citing examples that relate to the lessons learned on the design of training, we did not assess the effectiveness of these training programs and practices. Rather, we attempted to highlight some of the experiences and lessons that the agencies found helped them move forward in improving their training and development programs. Federal agencies' training and development strategies, and how they are designed to operate in conjunction with other strategies to improve individual and organizational performance, continue to change and evolve. We conducted our review from August 2002 through November 2003 in accordance with generally accepted government auditing standards. See appendix I for additional information on our objective, scope, and methodology. Results in Brief: Officials from the five agencies in our review recognized the importance of assessing current and future agency skill and competency requirements to identify related workforce training needs. The agencies used several approaches to design training and development programs that focused on the skills and competencies their assessments indicated needed enhanced attention. The officials emphasized that agencies are transitioning to more formal and comprehensive planning approaches to assess skill and competency requirements and identify related training and development needs--primarily as part of broader efforts to incorporate workforce planning into ongoing strategic planning and budgeting processes focused on achieving results. The following are lessons learned identified during this review related to assessing skill and competency requirements and identifying training needs. * Involve key stakeholders and benchmark with other organizations when identifying skills and competencies to help ensure that training and development programs are aligned with current and emerging needs and business practices. * Analyze existing agency data on employees' skills and competencies and information from performance appraisals to help identify skills and competencies that need to be addressed throughout the agency as well as on an individual basis. * Link the agency's workforce planning efforts with training needs assessments to ensure consistency and enhance strategic alignment. * Consider the training needs of staff from other organizations that will likely use the agency's training programs or facilities to effectively leverage training investments and meet diverse needs. These agencies developed a wide range of strategies and solutions to improve performance through designing training and development programs for their employees. Officials told us they considered a mixture of both on-the-job and other developmental programs and contemplated an assortment of mechanisms for delivering the training as well as potential sources for the learning need. Agency officials have found projecting costs and benefits of proposed training and development programs to be very challenging. Although they sometimes developed broad information on anticipated benefits and expected costs, this often did not involve tying anticipated benefits to specific performance improvements or considering all related costs. Our review identified the following lessons learned in developing strategies and solutions when designing agency training and development programs. * Incorporate information on employees' various competency levels and job needs into the design of training and development programs to increase their relevancy and timeliness. * Assess options for using other organizations' course content, staff, services, or facilities when designing a new training and development program in order to develop efficient and cost-effective strategies. * Establish mechanisms and controls to avoid unnecessary duplication or inconsistency within and across agencies' training efforts. * Develop and use criteria for determining the optimal mix of delivery mechanisms to use in order to select the most effective approaches given each learning situation. * Ensure that employees have the needed equipment and technologies so that they can take maximum advantage of learning opportunities. * Plan early when developing integrated solutions that complement other planned and ongoing strategies to improve performance so that when implemented the strategies work effectively and are aligned to help achieve agency goals. * Plan for the direct participation of senior agency leaders and experienced staff in the delivery of training and development programs to increase buy-in and build support for organizational change. Evaluating training programs is key to ensuring that training and development programs are effective. Overall, the five agencies relied primarily on standard end-of-course evaluations to obtain the participants' reaction to, and satisfaction with, a specific training course or learning opportunity. However, officials said that they have begun or are planning to use more comprehensive and sophisticated techniques for assessing the extent to which training and development programs increased employees' knowledge and skills and enhanced individual and organizational performance. These techniques included pre-and post-testing, tracking changes in individual and program performance, and some limited use of return-on-investment (ROI) analyses. The lessons agencies learned in designing methods to evaluate training and development programs included the following. * Incorporate appropriate aspects of the evaluation approach when designing training and development programs by specifying what results are expected to better ensure the availability and use of quality performance data. * Consider new approaches for collecting and analyzing performance data with the aim of increasing the quality and quantity of training evaluation feedback. * Plan for the use of multiple data types and sources to provide a balanced approach in assessing the effectiveness of training and development programs. * Take into account all relevant factors for determining the costs of a training and development program to better ascertain whether it is cost-effective in relation to benefits achieved. Four of the five selected agencies provided comments on a draft of this report. Interior and VA said that they generally agreed with the report's findings relating to their respective agencies. IRS said that it was honored to share some of its lessons learned with us for governmentwide dissemination. OPM said that it appreciated the opportunity to be included in the report and to share information on its training and development activities and programs. USACE informed us that it had no comments on our draft report. Background: We recently issued an exposure draft of an assessment guide that introduces a framework for evaluating a federal agency's training and development efforts.[Footnote 1] This assessment guide consists of a set of principles and key questions that federal agencies can use to ensure that their training and development investments are targeted strategically and are not wasted on efforts that are irrelevant, duplicative, or ineffective. As detailed in our assessment guide, the training and development process can loosely be segmented into four broad, interrelated components: (1) planning/front-end analysis, (2) design/development, (3) implementation, and (4) evaluation. Figure 1 depicts an overview of this process along with the general relationships between the four components that help to produce a strategic approach to federal agencies' training and development efforts. Although these components can be discussed separately, they are not mutually exclusive and encompass subcomponents that may blend with one another. For instance, evaluation is an integral part of the planning/front-end analysis as agencies strive to reach agreement up front on how the success of various strategies to improve performance, including training and development efforts, will be assessed. As noted in the assessment guide, agencies can build on lessons learned and performance data and feedback from previous experiences. This report can provide a starting point for agencies to use to build on the experiences and lessons learned by the five agencies we reviewed as part of their efforts to design and develop training and development programs. (See app. II for a description of the five agencies included in this study.): Figure 1: Four Components of a Strategic Training and Development Process: [See PDF for image] [End of figure] Our assessment guide also summarizes our observations on the core characteristics that make a training and development process effective and strategically focused on achieving results. These eight core characteristics are described in more detail in appendix III, and include: * strategic alignment, * leadership commitment and communication, * stakeholder involvement, * accountability and recognition, * effective resource allocation, * partnerships and learning from others, * data quality assurance, and: * continuous performance improvement. A concerted effort to integrate these core characteristics can further an agency's efforts to continually improve its training and development process. Agencies Used Varied Approaches in Assessing Skills and Competencies and Identifying Related Training Needs: Federal agencies face diverse challenges in their efforts to identify and measure the skills and competencies that their employees must possess to support missions and goals. Officials from the five agencies in our review recognized the importance of assessing the need for specific skills and competencies now and in the future in order to identify related workforce training needs. These agencies generally focused on the desired performance of the agency and its employees, determined the difference between the desired and actual skill levels, and attempted to identify the key factors contributing to performance, including the need for enhanced workforce competencies. Officials used a variety of approaches and tools to assist in determining the human capital skills and competencies that are critical to achieving their long-term goals. An agency's ultimate goal in undertaking training and development efforts is, of course, to optimize employee and organizational performance. To help ensure that each training program is linked to improving individual and agency performance, agencies first need to analyze their strategic and performance goals so that they can determine where training and development can most effectively enhance goal achievement. Organizations can evaluate the extent to which human capital approaches support their accomplishment of current, emerging, and future strategic goals through the use of workforce planning. Workforce planning focuses on determining the skills and competencies needed now and in the future to meet the agency's goals; identifying the current and projected level of the skills and competencies of the workforce; and crafting strategies for acquiring, developing, and retaining people to address any identified needs. These needs include the knowledge, skills, and abilities needed for the agency to pursue its current and future mission as well as the size of the workforce and its deployment across the organization. After identifying the skills and competencies that employees need now or in the future, agencies must tackle the challenge of determining what combination of strategies to use, such as hiring new employees with needed skills and competencies, relying on outsourcing, and/or enhancing employees' skills and competencies through training and development. While agencies' approaches to workforce planning will vary, we have identified the need for a strategic workforce planning process to ensure that each agency's human capital program capitalizes on its workforce's strengths and addresses related challenges in a manner that is clearly linked to achieving the agency's missions and goals.[Footnote 2] Figure 2 presents a model of this strategic workforce planning process. Figure 2: Strategic Workforce Planning Process: [See PDF for image] [End of figure] The focus in this report is on the workforce strategies that involve the design of training and development programs. Agencies' Experiences in Assessing Skill and Competency Requirements and Identifying Training Needs: The five agencies used several approaches to design training and development programs that focused on skills and competencies their assessments indicated needed enhanced attention. One common approach that officials used to help identify training needs was interviewing or surveying managers, supervisors, and employees. Agencies also established councils and held conferences, made comparisons with leading organizations through benchmarking, and analyzed workforce data and trends. OPM's 2001 skills assessment, for example, relied on a survey of agency managers and supervisors.[Footnote 3] Officials used the survey results to identify the most important occupational competencies needed to achieve OPM's mission, the level at which employees possessed those competencies, and the level at which they would be needed in the future. Their analysis identified minor gaps in the level of competencies needed for both current and future work in mission critical occupations. In addition, the analysis pointed out more serious developmental needs for OPM's retirement and insurance benefits specialists. These needs were related to the changing role of these specialists, who increasingly need to work more closely with clients in responding to complex issues. Also in 2001, IRS established a workforce planning council consisting of senior management representatives from each of the agency's operating divisions. IRS officials told us that this council has become the primary vehicle for communicating workforce planning information among IRS's four operating divisions. At FWS, the human resources office hosted a 3-day workforce planning conference in 2002 to draw on the experience and expertise of agency personnel in identifying critical workforce issues for the agency for the next 3 to 5 years. Managers and program experts representing all eight major FWS program offices, seven field regions, and headquarters offices participated. Agencies also compared their performance and needed skills and competencies with leading organizations through benchmarking. In August 2000, VHA commissioned an internal task force charged with developing a well-defined, comprehensive succession plan for the agency. The ideas garnered from benchmarking led to VHA establishing an expectation for agency leaders to help identify and train their successors. The task force's December 2001 report presented a comprehensive succession plan for VHA, and implementing a comprehensive leadership development program was one of the six major components of this plan. Agency officials also analyzed workforce data to assess skills and help identify training needs. Generally, they collected information on employee demographics and retirement eligibility and used these data to project attrition and retirement rates. OPM officials, for example, collected and analyzed attrition and turnover data on the agency's senior executives along with the distribution of current executives by unit and projected retirements through 2010.[Footnote 4] They also collected and analyzed data on hires, separations, and workforce diversity across the agency. As part of this analysis, officials assessed the agency's use of contractors and considered how sourcing alternatives could affect OPM's plans for hiring, training, and development. Officials from the five agencies told us that they used a wide range of resources and tools to assess skills and competencies as part of identifying and designing needed training and development programs. They used workforce planning models; assessed the workforce in view of organizational, occupational, and unit-based competency standards; conducted knowledge and skills inventories; and evaluated job performance appraisals and information from individual development plans (IDP).[Footnote 5] To identify needed executive competencies, for example, OPM used a five-step workforce planning model that it had developed in its role of providing human capital tools for use by other federal agencies (see fig. 3). The resulting analysis called for enhancing leadership development within OPM. Figure 3: OPM's Five-step Workforce Planning Process: 1. Set strategic direction to drive agency operations and define how the agency will know when and if it has succeeded; 2. Analyze the workforce, identify skill gaps, and conduct workforce analysis; 3. Develop a workforce action plan that lays out specific tasks and actions the agency needs to take in order to achieve the agency's human resources goals and objectives; 4. Implement the workforce action plan by executing the schedule that includes measurable workforce goals and milestones; 5. Monitor progress, evaluate success, and revise plan as needed. Source: OPM. [End of figure] Officials said that their agencies have transitioned, or are in the process of transitioning, to more comprehensive, consistent planning approaches. This transition is coming about as agencies attempt to institutionalize their workforce planning efforts as part of their ongoing strategic planning and budgeting processes. In 2002, FWS conducted its first formal, agencywide workforce planning process. Although FWS initially employed a contractor to help develop a permanent workforce planning process, it plans to continue to manage it in-house. This workforce planning process is to be implemented on a 2- year cycle that is integrated with the agency's strategic planning and budgeting processes. Agencies' Lessons Learned in Assessing Skill and Competency Requirements and Identifying Training Needs: By considering the viewpoints of a range of stakeholders and candidly and openly assessing progress toward meeting their goals, agencies can help ensure that their strategic and annual performance planning processes adequately reflect current ideas, policies, and practices in the field. Agencies continue to integrate workforce planning into these other planning processes. It is important to note that a wide variety of strategies other than training and development are also available to agency leaders as they attempt to transform their cultures and operations.[Footnote 6] Training and development is not always the best solution--reengineering processes or other actions may be needed to build an environment that effectively supports performance. In addition, training and development strategies frequently need to be implemented in conjunction with other initiatives, given that the day- to-day environment and organizational culture may also need to change to enable employees to successfully use new skills or competencies on the job. Our review identified four lessons learned by the agencies related to assessing skills and competencies to identify, focus, and prioritize training needs. Lesson learned: Involve key stakeholders and benchmark with other organizations when identifying skills and competencies to help ensure that training and development programs are aligned with current and emerging needs and business practices. Organizations in the private and public sectors have increasingly turned to developing competency models that outline behaviorally defined skills and competencies employees should possess and that can be tied directly to training and development plans and programs. We have found that an effective performance management system uses competencies to provide a fuller assessment of performance.[Footnote 7] IRS involved key stakeholders and benchmarked with other organizations in developing its leadership competency model: In working to build its leadership development program in the wake of the IRS Restructuring and Reform Act of 1998, IRS officials believed they needed a leadership competency model that was based directly on the work of IRS's business units. To identify the essential characteristics that enable IRS employees to function as effective leaders in the newly modernized agency, human capital specialists at IRS conducted behavioral interviews with 35 top IRS leaders in 1999, asking them to identify major successes and challenges during their careers. Using information gathered from these interviews, officials identified core management responsibilities and corresponding competencies required for leaders in IRS. With the assistance of a contractor, IRS validated the leadership competency model by comparing it against leading practices in the public and private sector and linking it to the mission and goals of the agency.[Footnote 8] The resulting competency model now forms the basis for IRS's leadership development efforts, as well as how IRS selects, evaluates, and recognizes its leaders. Figure 4 shows a listing of IRS's five core management responsibilities and the 21 corresponding leadership competencies. Figure 4: IRS's Core Management Responsibilities and Leadership Competencies: Leadership: * Adaptability; * Communication; * Decisiveness; * Integrity/honesty; * Service motivation; * Strategic thinking; Employee satisfaction: * Continual learning; * Developing others; * Diversity awareness; * Group leadership; * Teamwork; Customer satisfaction: * Customer focus; * Entrepreneurship; * External awareness; * Influencing/ negotiating; * Partnering; Business results: * Achievement orientation; * Business acumen; * Political savvy; * Problem solving; * Technical credibility; Equal employment opportunity (EEO) and diversity: * Supporting competencies[A]. Source: IRS. [A] Italicized competencies support the "EEO and diversity" responsibility. [End of figure] VHA benchmarked with other organizations and used a pilot test in developing its high performance development model: VHA conducted an extensive literature search and benchmarked with several leading private sector firms (including Bell South, Coca-Cola, and Motorola) to provide a foundation for its effort to create a new competency model for VHA employees. VHA used this information in developing its high performance development model, which was implemented throughout the department in 2002. This model consists of eight core competencies and related performance tools that represent the major skills and competencies that employees need to fulfill VHA's mission. The model was designed to serve as a framework for identifying and developing future leaders, as well as to enhance development of VHA's entire workforce. VHA said that using this model helped more effectively align training and development programs with agency priorities. According to VHA, the fact that the core competencies apply to all levels and functions within the agency helps ensure alignment within and between organizational units and is a key component in motivating sustained and improved performance. VHA also uses the model on an individual employee basis as a process for identifying specific developmental needs. Figure 5 lists the eight core competencies in VHA's high performance development model. Figure 5: Competencies in VHA's High Performance Development Model: * Personal mastery; * Interpersonal effectiveness; * Flexibility and adaptability; * Systems thinking; * Technical skills; * Customer service; * Creative thinking; * Organizational stewardship. Source: VHA. [End of figure] Lesson learned: Analyze existing agency data on individual employee's skills and competencies and information from performance appraisals to help identify skills and competencies that need to be addressed throughout the agency as well as on an individual basis. To obtain a unit-or agencywide perspective of skills and competencies, some agencies such as USACE and IRS have explored new ways of aggregating data from tools that are primarily focused on individual employees, such as IDPs, performance assessments, and 360-degree feedback instruments.[Footnote 9] Officials from these agencies told us that this information helped them discern a clearer picture of the overall strengths and weaknesses of their employees and offered direction in planning and designing training and development programs to help focus efforts to enhance skills and competencies throughout the agency. USACE's automated training management program provides a Web-enabled integrated database: According to USACE officials, using an automated training management program has allowed managers to identify divisionwide gaps in workforce skills and competencies. Using this system (currently in four of USACE's eight divisions) employees prepare an IDP assessing their knowledge, skills, and abilities in relation to a series of mission essential tasks. With supervisory guidance, each task is identified as critical, important, or beneficial and employees indicate whether they have received adequate, partial, or no training in that area. With this assessment as a guide, the supervisor and employee can consult the system's built-in course catalog to select internal or external training to enhance the employee's development. In addition, the system also has the capability of aggregating data. USACE officials said that this capability provides a simple method for division managers to obtain a picture of the level of skills and competencies in their workforce. This information informs decisions on training priorities and helps managers determine the most efficient use of available resources. IRS aggregated data from 360-degree feedback instruments to help identify training needs: To assess the progress and developmental needs of leaders within IRS, the agency's leadership development office recently aggregated and analyzed multiyear data from the 360-degree performance assessments of IRS managers. This analysis helped to show areas of strength and weakness in skills and competencies across the agency's managerial ranks. The director of leadership and organizational effectiveness at IRS said that, in the past, the agency did not sufficiently assist managers in effectively using 360-degree feedback they received. However, he said that IRS now emphasizes the importance of using 360- degree feedback data, both on an organizational and individual basis, to focus on strengths in developing key leadership competencies. IRS's Extraordinary Leader Program involves designing unique developmental approaches to help managers become more effective leaders. Using results of the manager's 360-degree assessment, IRS creates a customized leadership development program focusing first on correcting any "fatal flaw" weaknesses and then building on the manager's demonstrated strengths in areas that IRS has identified as key to providing effective leadership within its organizational culture and operating environment. Lesson learned: Link the agency's workforce planning efforts with training needs assessments to ensure consistency and enhance strategic alignment. The agencies' training and development organizations had a range of responsibilities, including designing training and development programs based on strategic initiatives, soliciting input from stakeholders, and prioritizing and scheduling training based on strategic initiatives and stakeholder input. Generally, the training organization and agency stakeholders can work together more effectively when they better understand how each office or function within the agency contributes to achieving business goals. In some cases, this included efforts to link training needs assessments with the agency's overall workforce planning efforts. Officials told us that this linkage helped ensure that workforce plans developed by the agency's human capital office were consistent with training needs assessments done by the agency's training and development organization. FWS involved key internal stakeholders in its planning processes: When assessing workforce skills and competencies, FWS officials worked to ensure that the agency's workforce plan was linked with a training needs assessment done by its training center. The workforce planning effort identified broad competencies needed across the agency's workforce while the training needs assessment identified the types of training courses to develop skills and competencies within agency units and occupations. Officials said that they viewed the training needs assessment as a tool that was useful in refining the agency's workforce plan and in prioritizing and budgeting for the development and delivery of training. Both efforts involved key stakeholders from the human capital and training offices as well as other FWS units. In the future, officials said they will rely on the results of the agency's workforce planning efforts to directly serve as the agency's training needs assessment. USACE relies on its Learning Advisory Board and automated training management program to effectively link planning efforts: USACE relies chiefly on the coordination activities of its Learning Advisory Board to ensure its workforce planning efforts and training needs assessments are effectively linked. USACE in 2001 formed the Learning Advisory Board, comprised of senior managers from across the agency, to review the adequacy of USACE's training and development and ensure that training is properly aligned with the agency's missions, goals, and plans. In addition, the four divisions that use the automated training management program can also rely on data from that system to assess training needs. This system allows managers to compare information on individuals' skills and competencies with workforce planning results from within the division and across the agency. According to USACE officials, this systematic comparison more closely links workforce planning and training needs assessments to the essential mission-related operations. Lesson learned: Consider the training needs of staff from other organizations that will likely use the agency's training programs or facilities to effectively leverage training investments and meet diverse needs. When planning and designing training programs for its employees, FWS and USACE officials told us that they gained insight into the assessment of potential training solutions by considering the possible involvement of trainees from other organizations. In some instances, for example, it would not have been cost-effective to design, develop, and deliver a training effort for a small number of employees or occupations. However, officials' determinations that other agencies or organizations also needed similar training provided the critical mass needed to move forward. Officials said they found that partnering with other organizations helped make training efforts more cost effective to design, develop, and deliver. FWS's training center assessed training needs of possible participants from other Interior components: Officials from FWS's training center said that although they focus primarily on meeting the needs of FWS employees when planning and designing training, they also look at the training needs of other organizations, particularly other agencies within Interior. To aid in communication and coordination, agencies within the department designate employees to serve as liaisons between each of the Interior agencies and FWS's training center. These liaisons facilitate efforts to incorporate their agency's needs into the training center's plans and the design and delivery of training and development programs. According to agency officials, these liaisons help ensure that the center's courses remain current because they facilitate a dialogue between the various agencies. For example, a recent FWS course on wetland plant identification involved participants from FWS; U.S. Geological Survey, which is another Interior component; Natural Resources Conservation Service, an agency of the U.S. Department of Agriculture; USACE; and three private sector firms. USACE's training center considers training needs of staff from other federal and state organizations: Officials from USACE's training center told us that during the agency's annual training needs assessment, they assess the possible training demand from other organizations' employees in addition to identifying workforce development needs of units and offices within USACE. According to USACE, approximately 2,500 people each year--about 25 percent of the participants in the agency's training center programs-- are from other federal and state organizations. Officials noted that the training center offers courses needed to obtain certifications for certain professional requirements. It offers courses accredited by several professional associations, including the National Society for Professional Engineers, the American Institute of Architects, and the International Association for Continuing Education and Training. USACE officials said that they would like to offer training to a greater number of employees from private sector firms; however, current law requires receipts for services provided to private individuals and organizations to be deposited into the general treasury as miscellaneous receipts.[Footnote 10] USACE officials said that this requirement hampers the ability of USACE's training center to keep funds it could generate to further invest in its training programs. However, when agencies are required to return receipts for services to the Treasury, the Congress is preserving its oversight and control over the programs generating the fees. Agencies Developed Strategies and Solutions for Their Training Needs: Agencies carry out their training and development efforts on the basis of estimated needs, priorities, and available resources and recognize that adequate planning allows them to establish priorities and determine the best ways to leverage investments to improve performance. The five agencies we reviewed set priorities for training and development on the basis of various factors, such as the results from skill and competency assessments, the availability of resources, and the interests of agency leaders. They usually relied on training officials, agency managers, and subject matter experts to assist in developing strategies and approaches for addressing training needs. Although the agencies designed and delivered training using both centralized and decentralized approaches, we found that leadership development programs were more highly centralized and managed at headquarters. Agency officials acknowledged that they found projecting costs and benefits of proposed training and development programs to be very challenging. Although they sometimes developed broad information on anticipated benefits and expected costs, this often did not involve tying anticipated benefits to specific performance improvements or considering all related costs. As outlined in figure 6, agencies can plan and establish priorities by developing an annual training plan to target developmental areas of greatest need and outline the most cost-effective training approaches to address those areas. Considerations involved in assessing investment opportunities for the training plan include balancing the competing demands confronting the agency and the amount of resources available in order to determine how those demands can best be met with available resources. It is also important to consider how to effectively integrate all of the strategies the agency plans to use to improve performance and meet emerging demands. When training is identified as a solution to improve performance, agencies can compare various training strategies by weighing their estimated costs and anticipated benefits to build a convincing business case that supports the selected training strategy. Developing a business case that sets forth the expected costs and benefits of the performance improvement investment provides decision makers with essential information for allocating necessary resources. Figure 6: Steps for Developing Strategies and Solutions for Training and Development Needs: [See PDF for image] [End of figure] Agencies' Experiences in Developing Strategies and Solutions for Training Needs: The agencies we reviewed used a wide range of strategies and solutions to improve performance through designing training and development programs for their employees. Officials told us they considered a mixture of both on-the-job and other developmental programs, contemplated an assortment of mechanisms for delivering the training, and assessed potential sources to meet their learning needs. For example, USACE's leadership development program for midlevel engineers and scientists involved formal classroom training, mentoring, and a 6- month developmental assignment. VHA employees can access a wide variety of informational and educational content through the VA Knowledge Network, a satellite-based system of live and on-demand programming delivered directly to employees' desktops. IRS's training unit developed an automated ROI workbook tool that the agency's business units can use to assess whether proposed training programs should be delivered in a classroom or by an e-learning approach. OPM partnered with an employee union to offer a midcareer development program that provided an opportunity for current OPM employees to enhance existing skills, explore new career fields, and gain practical experience. Officials from FWS's training center cited courses in negotiation, communication, and interpersonal skills as examples of vendor-provided courses. They said the center decided not to invest in designing these courses since they do not require field experience and expertise, which are critical in designing other FWS courses. Projecting costs and benefits of proposed training and development programs was a challenge for the five agencies. They usually developed broad information on anticipated benefits and expected costs, often without tying anticipated benefits to specific performance improvements or considering all costs related to the training program. For example, VHA officials told us that the agency's assessments of anticipated benefits and expected costs of proposed training are generally unsystematic. One VHA office or field location may not have analyzed the relative costs or benefits of proposed training while another office or location may have considered anticipated benefits and developed estimates of costs and savings using different training approaches. At FWS, some proposed training programs, such as its Advanced Leadership Development Program, involved detailed estimates of costs, both for the training center and for participating FWS field offices, as well as the identification of specific competencies to be developed in the program. Other proposed training programs at FWS did not have documented and detailed estimates of expected costs and benefits. FWS officials said that they assess the anticipated costs and benefits of all their proposed training and development programs but that the extent of these assessments and the amount of documentation supporting the assessments vary, depending on many factors, such as the content, delivery mechanism, and uniqueness of the proposed training. According to agency officials we interviewed, limited funding sometimes affected agencies' abilities to design and deliver training and development programs that officials believed were needed. USACE officials said, for example, that in a recent survey, about two-thirds of agency supervisors and one-half of agency executives believed that the agency had less funding for training civilian employees than is needed. Funding and resource limitations sometimes forced the agencies to think of new and practical ways to ensure that their employees had the knowledge and skills needed to carry out their work. For example, because of an unexpected decrease in available travel funds, VHA officials canceled plans for a large national conference on the use on an automated managerial cost accounting system. This system was designed to provide VHA managers with data important in making clinical decisions, managing workload, and controlling medical care costs. Instead, to enable employees to obtain the information that would have been presented at the conference, VHA officials provided the content via a satellite broadcast along with a series of audio conferences. At FWS, the number of people that the training center can train onsite, of course, is limited by the current capacity of the facility classrooms and residences. According to FWS officials, incorporating a blended learning approach into its curriculum has enabled trainees to perform part of the course work outside class, thus allowing instructors to focus on those topics that require special attention. Training center officials also said that the agency has increased its use of e-learning and other mechanisms to develop employees where they work instead of coming to the training center. Officials told us that FWS has reduced training costs by offering more training online and using CD ROMs to provide field offices with course material that previously had been offered only on location at the training center. Agencies' Lessons Learned in Developing Strategies and Solutions for Their Training Needs: Agency officials have encountered a variety of challenges in their efforts to design training programs to meet the developmental needs of their employees. How agencies respond to these challenges can greatly affect their success in aligning priorities with strategic direction, assessing strategies, identifying alternative sources and methods, weighing potential costs and anticipated benefits, and assessing how other performance improvement initiatives might complement training efforts. It is also important to consider evaluation feedback on an ongoing basis. We identified eight lessons learned related to the five agencies' efforts to develop strategies and solutions for their training and development needs. Lesson learned: Incorporate information on employees' various competency levels and job needs into the design of training and development programs to increase their relevancy and timeliness. When designing effective training and development programs, the way the work is actually to be done on the job and the developmental needs of the expected trainees are key considerations. Analyzing the tasks of specific jobs and occupations can help ensure that training accurately reflects the way employees are expected to perform on a day-to-day basis. To help ensure that the training effectively addresses employees' developmental needs, agencies can determine the workforce's level of proficiency in mission-critical skills and competencies by conducting skills assessments and using information obtained through interviews or surveys of employees and their supervisors. The increased information and insight provided by these approaches can allow agencies to incorporate information on employees' various competency levels and job needs into the design of training and development programs-- increasing both the relevancy and timeliness of the learning. FWS targeted training to employees to leverage specific knowledge and experiences: FWS's training center recognized employees' various competency levels and job needs when developing courses on the use of geographic information systems (GIS), which are becoming increasingly important mapping and information analysis tools for natural resources agencies, according to FWS officials.[Footnote 11] Rather than providing instruction focused solely on software features and functions of GIS, the training center designed its curriculum to teach the application of GIS to employees based on their roles in natural resources management. With a focus on these varied roles and related needs, the center developed separate GIS training courses for "explorers" (natural resource managers or others just wanting to know about GIS), "users" (biologists and other personnel using GIS in their daily job) and "developers" (those individuals designing and developing a natural resources GIS for use by others). Officials at FWS's training center said that with the training targeted to the specific background and needs of employees, FWS can minimize the time spent teaching participants information that they already know or do not need to know in carrying out their job responsibilities. The training center's curriculum consists of 15 separate GIS courses, with 3 additional courses under consideration or development. According to FWS officials, about 1,200 FWS employees at over 400 offices use GIS software in their jobs. IRS used skills assessments to focus on developmental needs: To assist in identifying employees' competency levels and incorporating job needs into training, IRS developed and used technical assessment batteries for the agency's field assistance personnel and customer service representatives. These multiple-choice instruments were designed to assess the key technical knowledge that the employee needs in order to carry out his or her job. On the basis of each employee's test results, the agency will recommend specific training and other appropriate interventions, such as mentoring, to improve performance. IRS officials told us that in some cases where the assessments showed that individuals were already knowledgeable in a particular area, employees still wanted to take the related training because they viewed training as a job benefit as well as a way to improve knowledge and skills for their jobs. This provides an example of how important agencies' considerations of the organizational culture and working environment when designing training and development programs are in preparing for and addressing issues that may arise during implementation. As we point out in our training guide, employees need to not only understand the goals of agencies' training and development efforts, but also to accept responsibility for developing their competencies and careers, as well as for improving their organization's performance.[Footnote 12] Lesson learned: Assess options for using other organizations' course content, staff, services, or facilities when designing a new training and development program in order to develop efficient and cost- effective strategies. When thinking about strategies and sources for the design of a new training and development program, officials can potentially discover more efficient and cost-effective approaches through the use of other organizations' course content, staff, services, and/or facilities. Adequate planning can help an agency in meeting the developmental needs of trainees without overburdening the agency's training capacity or creating excess capacity. Obtaining reasonable estimates of likely costs and identifying potential obstacles of using others' training resources can help agencies develop more informed perspectives on ways to effectively leverage resources. IRS considered several options in designing a course for senior managers: In evaluating options for designing and delivering a new training course for its senior managers, IRS considered various sources for such training, including internal resources, contractor support, and partnerships. IRS officials said that they considered internal resources to design the training but quickly realized that the agency did not have sufficient expertise. The officials also considered using a contractor but concluded that the costs would be too high. Instead, IRS decided to partner with the Federal Executive Institute, an OPM- sponsored training facility in Charlottesville, Virginia, that provides training to senior employees from across the federal government. IRS officials found that partnering IRS design and subject matter experts with institute and other renowned leaders in the field delivered the most cost-effective approach and yielded the best results. The design team produced a course called "Learning Through Others," delivered on the Charlottesville campus. According to IRS officials, this course surpassed agency needs and expectations and was less expensive than a direct contracting arrangement with an outside vendor. They added that participants in the training program could learn public service values through lessons and encounter competency-based experiential learning, business-related challenges, and a capstone simulation. IRS officials characterized the course as high quality and said the prestige associated with studying at the Federal Executive Institute provided an additional benefit for IRS participants. Lesson learned: Establish mechanisms and controls to avoid unnecessary duplication or inconsistency within and across agencies' training efforts. The agencies used both centralized and decentralized approaches by, for example, centrally managing reporting and record keeping while allowing some localized management of training content. Whatever mix of centralized and decentralized approaches is used, agencies recognize that it is important to limit overlap and duplication and ensure the delivery of an integrated message when appropriate. VHA and FWS officials found that establishing mechanisms and controls is important to limit duplication or inconsistencies within an agency, across component organizations within a department, or across the federal government as a whole. VHA's Employee Education System helped limit duplication of effort: VHA's Employee Education System, which serves as an internal training consulting team within VHA, assists the agency's 21 regional networks in designing and implementing programs to develop general and specific skills for VHA employees. Within each of the 21 regional networks, an Education Service Representative acts as a liaison in coordinating numerous developmental programs with VHA headquarters--sharing information with their counterparts about effective practices and identifying areas of possible duplication or inconsistency across VHA. According to VHA officials, the coordination and communication achieved through this organizational structure has helped ensure consistency in implementing the agency's national training priorities. For example, officials said that the consulting team assisted in implementing changes to VHA's processes for collecting third-party insurance reimbursements in the wake of legislation that required VA to make greater efforts to collect unpaid debts from veterans. They told us that these legislative changes enabled local VHA facilities to receive these reimbursements, but also overwhelmed the local billing and debt collection processes. To address the problem and help ensure consistency across the agency, the training consulting team participated in redesigning the processes for coding, billing, and debt collection; trained the employees responsible for billing and debt collection in the new processes; and created graphical representations of the new processes and posted them throughout the agency's facilities to aid employees in learning. Interior's Training Directors Council facilitated communication across departmental components: Interior used its Training Directors Council to facilitate communication across the department's different bureaus, thus helping to minimize duplicative training and development efforts. This council provides opportunities--through formal meetings and informal communications--for training managers from Interior's various bureaus to share curriculum and related training ideas with their colleagues. The director of the training center at FWS, who chairs the council, told us that on more than one occasion he has discovered through council business that other Interior components had developed strategies or solutions to address emerging or existing needs that FWS's training center had also identified. For instance, the training center at FWS was considering whether to add a new course on grants management to its curriculum. Through its participation in the council, FWS determined that another component agency, the Bureau of Land Management, already offered grants management courses through its National Training Center in Phoenix, Arizona. After reviewing the content of these Bureau of Land Management courses, officials at FWS's training center determined that they did not need to design and develop a separate grants management course. Instead, FWS and the Bureau of Land Management now jointly manage the delivery of this training course. Lesson learned: Develop and use criteria for determining the optimal mix of delivery mechanisms to use in order to select the most effective approaches given each learning situation. In response to emerging demands and the increasing availability of new technologies, agencies are faced with the challenge of choosing the optimal mix of training delivery mechanisms to design training that is as effective and efficient as possible. Agency officials consider a wide variety of instructional approaches to achieve learning--in the classroom, through distance learning, or through structured on-the-job experiences. Officials also took other factors into account, including whether to provide individualized instruction or team-based training and when to use blended learning that combines different teaching methods (e.g., Web-based and instructor-led) within the same training program. USACE found that identifying and systematically using criteria to help select effective delivery mechanisms assisted in building well- supported justifications for the design of training and development programs. USACE used criteria to select media and method of instruction: To select the appropriate media and method of instruction for its training programs, USACE uses criteria contained in the Corps of Engineers Systems Approach to Training, the agency's documented process for developing training programs. The criteria include issues such as the expected frequency of changes to the training content, the size and diversity of the target population, and the degree of student interaction required. USACE officials told us that, using these criteria, course managers from USACE's training center coordinate with relevant agency program offices and subject matter experts to decide on the appropriate mode for training delivery. While most of the training center's courses occur in a conventional classroom setting, agency decision makers have focused on trying to identify courses (or modules of courses) to convert from classroom training to more economical modes of delivery, such as distance learning, computer-assisted instruction, computer-based instruction, or a combination of such approaches. USACE officials said that many of their courses now incorporate CD ROM and Internet-based materials as prework assignments before attending classroom training and for reference use during and after the training events. VHA used a profiling tool to help in selecting delivery approaches: VHA's internal training consulting team used a training delivery strategy tool that consists of a series of questions structured to guide users through the process of selecting an appropriate delivery approach for a proposed training effort. This team designed the training delivery strategy tool to help staff plan, analyze, develop, and deliver training and development activities. To aid in decision making, the tool includes factors such as audience composition, course goals and objectives, course modules, any prerequisites, participant preparation, and course follow-up and evaluation. VHA officials said that applying this tool and analyzing the resultant profile helps maximize learner understanding, retention, and application. Lesson learned: Ensure that employees have the needed equipment and technologies so that they can take maximum advantage of learning opportunities. Many organizations are taking advantage of more flexible design and delivery methods made possible by technology to, for example, deliver training to the user's desktop, thereby making training more accessible and cost effective. As agencies move forward in using new approaches, it is important to ensure that employees have the needed equipment and technology to take maximum advantage of learning opportunities. IRS converted mandatory training courses to an online format: IRS recently converted a series of mandatory training courses from face-to-face group briefings to an online format in order to more efficiently provide this training to its employees. Through these online mandatory training programs, which include computer security awareness, ethics issues, and prevention of sexual harassment, IRS wanted to (1) reduce the burden of managers who previously had to prepare for and deliver the training, (2) provide ready access of the information to line employees when and where they need it, and (3) lower the costs associated with the group briefings. IRS officials said that online delivery lessened employee time in taking the training from approximately 6 to 2 hours and in some cases eliminated the need for travel. To ensure the accessibility and usability of these online briefings, IRS worked to resolve various challenges in the conversion, such as designing the online product for the lowest computing capabilities of the bulk of the trainee population and providing alternative delivery mechanisms for individuals who were without a computer or Intranet access. The officials said that they also learned it is important to design the online briefings based on a common template and style guide to standardize their look and feel, provide online text-only versions of the training for persons who are visually impaired, and test the usability of the online briefings with end users on a range of equipment and allow sufficient time for needed revisions. OPM initiated a pilot program for its employees to use online courseware: To explore opportunities for increased use of e-learning approaches, OPM established a pilot program that allowed about 250 of its employees access to approximately 1,800 online courses through the Department of Transportation's Transportation Virtual University. OPM's training unit worked with offices throughout OPM to identify employees to participate in the pilot program. OPM officials said that all employees selected for the pilot program had access to the equipment and technology needed to make use of the University's online training, by using a computer either at the employee's desktop or at some central location. As a result of this pilot, OPM enhanced its offerings of online training for its employees by becoming a partner in the GoLearn.com initiative, a governmentwide online training center for federal employees. OPM officials said the pilot program also showed that not all employees function effectively in an online training environment and that some employees need a more structured format in order to learn. Lesson learned: Plan early when developing integrated solutions that complement other planned and ongoing strategies to improve performance so that when implemented the strategies work effectively and are aligned to achieve agency goals. When designing training and development programs, the agencies sometimes considered how they could integrate them with other strategies to improve performance and meet emerging demands. If the work environment is not conducive to providing opportunities to use new skills or work in different ways, no matter how good the training program is it may not be effective or successful in terms of changing on-the-job performance. In addition, training and development programs represent a significant investment of resources (including time and money) and may: not always be part of an appropriate solution. The agencies developed integrated solutions that included developing and using job aids, performance support tools, and other approaches to enhance knowledge management[Footnote 13] and to aid employees on the job as a complement to training. OPM planned for an electronic support tool to aid agency employees in using a new computer system to process retirement claims: As part of its effort to reengineer and modernize its processing and support of federal employees' retirement claims, OPM is developing plans for an integrated Electronic Performance Support System to aid the agency's benefits specialists in using a new computer system. Procedural and information job aids are to be built directly into the software to provide documentation and guidance, "just in time" assistance, and error detection. This is intended to be an integrated system to permit coordination between different modes of training and enhance the learning and performance of the OPM employees working with reengineered business processes and the new computer system. OPM officials said, for example, that this system would assist employees in completing steps using actual data and circumstances of a particular case they were working on rather than consulting a manual or using data put together just for training. According to OPM officials, as the focus under modernization shifts from processing claims to providing customer service, this system will help employees working in OPM's retirement program to interact more directly with program participants to answer questions and solve problems about retirement issues. USACE identified online solutions to help enhance and integrate training efforts: As a complement to the training and development programs it offers to its employees, USACE recently entered into a joint project with the Department of Labor to use an online knowledge management system called Workforce Connections. This system, which resulted from a memorandum of understanding promoting cooperative efforts between the departments of Defense and Labor, will provide the USACE workforce with on-demand, online access to job aids, performance support materials, and: course content 7 days a week, 24 hours daily. The system will feature development and maintenance of online communities of practice to support knowledge management of USACE's Learning Network, which is USACE's overall platform for delivering a wide variety of learning resources to agency employees.[Footnote 14] Another part of the learning network is USACE's Virtual Campus, a distance learning site that allows employees access to Web-based courses and training events. Another component of the learning network includes electronic performance support tools, such as job aids and other information resources. USACE officials said that they consider the systems in the learning network to comprise a powerful solution that effectively integrates the agency's training efforts. Lesson learned: Plan for the direct participation of senior agency leaders and experienced staff in the delivery of training and development programs to increase buy-in and build support for organizational change. Internal resources, such as subject matter experts and high performers, can often provide valuable insight into training design because of their familiarity with the agency's policies, programs, and corporate culture. To increase buy-in, help establish greater credibility, and build support for organizational change, the agencies have learned the value of planning for the direct involvement of senior managers in the training program. IRS and OPM involved executives and managers: IRS officials told us that a key feature of the agency's frontline managers course is that it was designed to use senior managers and experienced frontline managers drawn from the agency's business units to teach the course. In addition, executives participate in course modules that focus on emerging issues facing the agency. The deputy commissioner of IRS's Wage and Investment business unit served as the executive sponsor of the training program and participated in course modules featuring executives. The officials also told us that IRS executives partner with outside vendors to serve as an instructor team to deliver all courses designed for senior managers. Using business unit executives and managers as course instructors helped ensure that the course's content and emphasis related to the mission, goals, and guiding principles of the agency. At OPM, agency managers have been the first to take special initiative training--such as courses on prohibited personnel practices, whistleblower procedures, and information technology security--before they are offered agencywide. OPM officials said that teaching the material to agency managers in advance of line employees enables the managers to model desired behaviors and learning for their employees and convincingly convey how they personally benefited from the training. FWS and VHA relied on in-house experts: FWS's training center brings in FWS field office personnel when building a cadre of senior, in-house instructors. Training center officials said that involving trainers from the field helps to build trust with trainees and provides an added level of credibility that neither academics nor other subject matter experts who lack field experience can easily replicate. According to these officials, many expert employees come from the field and stay to teach at the center for 3 or 4 years. Some, however, teach only one or two courses or get involved for a short duration before returning to their positions in the field. The director of the training center said that he views this passing-on of information from seasoned veterans to less experienced employees as crucial for maintaining the unique knowledge base of the agency. VHA used "super users" to teach medical center personnel to use its computerized patient record system, a computer interface that allows hospital personnel to keep more comprehensive patient records and enables clinicians, managers, and other staff to review and analyze data gathered on any patient. The super users--VHA employees with other job-related duties and responsibilities--were trained to be thoroughly knowledgeable about the system so they could demonstrate its capabilities and directly relate the training to employees' work. VHA initiated this strategy when the agency began rolling out the application in 1997. VHA officials said that planning to build on the direct involvement of these super users was successful because they served as first-line resources for employees' questions about the new system and helped the agency to build organizational support for the system. According to VHA, the agency developed a cadre of more than 2,500 super users, and about 180,000 VHA employees use the patient record system. Agencies are Considering More Sophisticated Evaluation Approaches As Part of Designing their Training and Development Programs: Without evaluation of training programs, participants may take ineffective courses that do not provide the necessary learning experience or that do not translate to improved performance on the job. Overall, the five agencies in our review relied primarily on standard end-of-course evaluations to obtain the participants' reaction to, and satisfaction with, a specific training course or learning opportunity. Although the agencies encountered challenges given some of the difficulties associated with measuring the impact of training on individual and organizational performance, they have begun or are planning to use more comprehensive and sophisticated evaluation techniques for assessing their training and development efforts. Such techniques include the use of pre-and post-testing to determine the extent of learning accomplished, tracking the performance or advancement of individuals and work units before and after training is completed to assess professional growth and improvements in organizational performance, and limited use of ROI analyses to compare the benefits (quantified in dollars) with the costs of a training and development program. To help determine whether the objectives of training and development are achieved, agencies can begin by incorporating measures of effectiveness into the design of training and development programs. Defining objectives in a measurable way enables agency officials to offer a more convincing business case and contributes to improving the quality of feedback. Whenever possible, training goals should measure the individual and organizational results achieved rather than the training inputs or outputs (e.g., number of available courses or people trained). Figure 7 depicts some of the steps involved in determining the evaluation methods to use in designing training and development programs. Figure 7: Steps in Determining Methods for Evaluating Training Programs: [See PDF for image] [End of figure] Agencies' Experiences in Determining Methods for Evaluating Their Training and Development Programs: Agency officials recognized the importance of determining during design how they planned to evaluate the effectiveness of their training and development programs. To collect information on participants' reaction to and satisfaction with the training program, for example, VHA uses a standard evaluation form with questions related to program design, delivery, outcomes, overall satisfaction, and logistics. In completing this survey, training participants evaluate their success in completing learning objectives and the performance of the faculty. Other agencies also obtained participant feedback though interviews or focus groups. OPM conducted exit interviews with individuals who participated in that agency's Presidential Management Intern program, a 2-year developmental program for individuals from a wide variety of academic disciplines who wish to enter the federal service. These exit interviews were designed to obtain feedback from the participants about their overall satisfaction and reaction to the Presidential Management Intern program, including their suggestions for enhancing the training provided. The five agencies have begun to use, or are planning to use, more comprehensive and sophisticated evaluation techniques to assess the extent to which training and development programs increased employees' knowledge and skills or enhanced individual and organizational performance. One of these more sophisticated evaluation techniques is the use of pre-and post-testing to determine the extent of learning during the training program. USACE's training center conducts pre-and post-tests on over 90 percent of the courses it offers and is working toward the goal of using such tests for all courses. The agencies also tracked job performance and the advancement or movement of personnel to assess the potential effectiveness of training. FWS officials told us they track participants' career advancement to determine the extent to which participation in the leadership development program for midlevel employees contributed to increased mobility into more responsible leadership positions in the agency. According to training center officials, about 37 percent of the program graduates have taken either promotions or new lateral assignments since the program's inception in January 2002. In addition, some agencies attempted to conduct ROI analyses to compare the benefits (quantified in dollars) to the costs of a particular training and development program. VHA officials pointed to concerted efforts to conduct ROI analyses on several training and development programs, including customer service, leadership development, and computer-based training. IRS officials, on the other hand, have decided that the challenges and difficulties in conducting such analyses are not worth the effort for the resultant information--given the challenge of isolating the performance improvements that might result from a specific training activity and the difficulty in monetizing identified benefits in order to calculate the ROI. IRS instead uses the concept of "time to capability" to determine whether and to what extent a training course, program, or other training intervention has improved the organization's ability to perform its mission successfully. IRS defines time to capability as the validated accumulation over time of employees who have been trained in specific competencies deemed critical to the success of an organizational unit. Under this approach, when IRS has trained a predetermined number of employees, officials consider that the agency has achieved the goal of training a critical mass within its workforce and conclude that the agency has an organizationwide capability in the specific competencies. Agencies' Lessons Learned in Determining Methods for Evaluating Their Training and Development Programs: Agencies' training and development efforts involve a continuous effort throughout planning, design, implementation, and evaluation. Therefore, it is important to recognize that evaluation is not a static requirement to be carried out after the fact. When undertaking design and development of training, agencies can rely on evaluations and benchmarking to determine what approaches work best given all the related elements, such as the proposed audience for the training program, the material to be covered, and possible delivery mechanisms that could be employed. Determining methods for evaluating training programs as part of their design can help identify and remove obstacles to successful implementation. For example, an agency officials said that catching potential problems early on saved valuable time and resources that a major redesign of training later on likely would have entailed. On the basis of our review at the five agencies, we identified four lessons learned regarding the agencies' efforts to determine methods for evaluating their training and development programs. Lesson learned: Incorporate appropriate aspects of the evaluation approach when designing training and development programs by specifying what results are expected to better ensure the availability and use of quality performance data. In assessing how and to what degree performance could be improved with a specific training program, agencies should try to establish a targeted level of improved performance as well as assess the possible consequences if the training were not to occur. Determining a target level for improved performance can aid agencies in assessing whether the expected costs associated with the proposed training are worth the anticipated benefits. Agreeing upon the planned evaluation approach in the design clearly sets forth the results the agency expects to achieve through the training. In addition, planning ahead helps ensure the availability and use of sound and relevant performance data. VHA decided to evaluate a customer service training program after it was implemented: In response to low scores on customer satisfaction surveys, a VHA regional network office pursued various initiatives to improve customer service, including the design, development, and implementation of a new training program focused on creating a more customer-service-oriented culture and improving employee morale and collaboration to better meet customer needs. After delivery of this new training program, called "The Customer," VHA selected a contractor to evaluate the effectiveness of the training effort. In its report, the contractor stated that because the opportunity to conduct the evaluation did not occur until after the training program had been delivered, the use of preferred evaluation methods was not possible. The contractor reported that this lack of preprogram planning had also been experienced in some other VHA network offices. Nevertheless, the contractor evaluated the customer service training program by surveying participating employees and their supervisors. From these survey results, the contractor concluded that the training program was viewed as successful for those who needed it, but that the training did not receive a sufficient level of support from supervisors. The contractor recommended that VHA obtain additional feedback from supervisors as well as from an individual hired to telephone or visit the VHA facilities in the network office to observe customer service activities. Lesson learned: Consider new approaches for collecting and analyzing performance data with the aim of increasing the quality and quantity of training evaluation feedback. The agencies faced various challenges in obtaining a high quality and quantity of feedback needed to evaluate their training and development programs. We previously reported that low participation on the part of employees and managers in surveys and focus groups may limit an agency's access to the data needed to complete valid and useful evaluations of training programs.[Footnote 15] With strong agency support and proper planning, stakeholders, including training participants, supervisors, managers, and trainers, are more likely to provide the information and feedback needed to successfully and effectively evaluate agency training and development programs. USACE, for example, recognized that it needed to ensure that it incorporated a wider variety of stakeholder perspectives in assessing the impact of training on employee and agency performance. Stakeholders' perspectives can be obtained through surveys and questionnaires, individual or group interviews, or communication with more formal multidisciplinary bodies such as advisory or education councils. Valuable sources of information include the training participants; training designers, developers, and facilitators; agency leaders, managers, supervisors, subordinates, and coworkers; employee organizations; internal and external customers; and functional and subject matter experts. USACE's training center altered its approach to obtaining supervisory feedback on its training efforts: USACE's training center has been attempting to obtain more sophisticated evaluation feedback to determine if its training courses affected employees' behavior on the job. Training center officials told us that they originally had planned to evaluate all of the center's training courses by obtaining feedback from employees and supervisors 6 months after the training course on the extent to which employee on- the-job behavior had changed. In a test of this approach, the training center sent out hard copies of two generic survey forms: one for the employee and one for the supervisor. Because the training center had locator information for trainees but not for their supervisors, both surveys were mailed to the employee, who was then asked to forward one to his or her supervisor. Training center officials said the response to this survey effort was disappointing. For one 35-person class, for example, the center received 3 employee surveys and 1 supervisor survey. To increase the level of feedback, officials told us that the center is transitioning to an electronic process whereby the survey forms are sent via e-mail to the employee, who is then requested to forward the survey via e-mail to his or her supervisor. Training center officials said that in a recent test of this new approach on one course, the center received a 67 percent response rate from employees and a 36 percent response rate from supervisors. The officials told us that they are working to develop unique evaluation forms for each course in order to obtain feedback on specific learning objectives rather than rely on a generic survey form for all courses. Lesson learned: Plan for the use of multiple data types and sources to provide a balanced approach in assessing the effectiveness of training and development programs. Successful organizations typically develop and implement human capital approaches based on a thorough assessment of the organizations' specific needs and capabilities. Valid and reliable data are the starting point for such assessments. To assess the results achieved through training and development, agencies can rely upon hard (quantitative) data, such as productivity/output, quality, costs, and time, or soft (qualitative) data, such as feedback on how well a training program satisfied employees' expectations. By taking steps to agree on measures of success up front, agency officials can decide on the objectives for each training and development program. Using a balanced approach that reflects feedback from customers and employees, as well as organizational results, is particularly important as agencies transform their cultures and operations. In addition, because the work of federal employees can be complex and often cannot be reduced to a single task, a balanced approach to both the types and sources of data helps to strengthen the linkages between training and development programs and improved performance. USACE conducts evaluations of some training courses: In addition to obtaining end-of-course participant feedback and administering pre-and post-tests for many of its courses, USACE's training center also conducts course evaluations of its offerings. To conduct this evaluation, the designated course manager from USACE's training center is responsible for observing the full course and assessing various aspects of its design and delivery, including the training content, materials, and instructors. Training center officials said that various factors can trigger the decision to conduct an overall course evaluation, including a significant decline in trainees' overall satisfaction with the course, the introduction of new instructors, the use of contractor assistance, or a specific recommendation from an agency office or unit. According to training center officials, the results of this course evaluation are assembled with the end-of-course participant survey feedback and pre-and post- testing results to present a comprehensive and balanced view of the effectiveness of the training program. Lesson learned: Take into account all relevant factors for determining the costs of a training and development program to better ascertain whether it is cost-effective in relation to benefits achieved. Calculating the ROI for a training program involves identifying and monetizing the program's benefits and then dividing this by a full tabulation of the program's costs. These costs should usually include the cost of program materials provided to each participant; the cost of the facilities; the costs of the facilitator or instructor, including time for both preparation and delivery; any travel-related expenses for participants; salaries and benefits costs of the participants for the time they attend the program; and an allocation of relevant administrative and overhead costs. VHA's evaluation of a leadership development program did not include all costs: In cooperation with VHA's internal training consulting team, one of VHA's regional network offices designed, developed, and implemented a networkwide leadership development program called "Competency Development for Leaders in the 21st Century." According to information we gathered during our review, the costs incurred for the consulting team's efforts on this training program were not included in the ROI calculation even though it contributed substantially toward developing and implementing the program. VHA officials told us that these costs were not included in the analysis because the course designers only wanted to determine the return on the network's investment, not the agency's overall investment. Agency officials said that one of the main goals of involving the consulting team was to help the network develop the capability to use the ROI process to evaluate training. Although these costs were not included in the ROI cost tabulation, VHA did include the salaries (plus employee benefits) of the participants for the time they attended the developmental program. As we noted in our recently issued assessment guide, agencies might overlook the costs of participant attendance when calculating the total costs of a training program.[Footnote 16] Conclusions and Observations: Federal agencies are faced with the need to invest resources wisely to ensure that their employees possess the information, skills, and competencies required to carry out their work successfully. The examples provided in this report may help to address this need by describing some of the experiences and lessons learned that other agencies might find applicable or adaptable to their unique situations. This information is intended to both provide a realistic perspective on how agencies have approached designing their training and development programs to date as well as to take a more detailed look at some of the concepts explored in our recently issued assessment guide focused on strategic training and development efforts in the federal government. Our work reviewing the selected agencies' efforts to design training and development programs reinforces the significance of good planning and design of these programs to ensure their successful implementation and evaluation. The experiences and lessons learned we identified also demonstrate how effective design efforts--as part of a strategic training and development process--rely on the eight core characteristics that we identified in our earlier work: (1) strategic alignment, (2) leadership commitment and communication, (3) stakeholder involvement, (4) accountability and recognition, (5) effective resource allocation, (6) partnerships and learning from others, (7) data quality assurance, and (8) continuous performance improvement. Indeed, by focusing on these eight core characteristics, agencies can improve not only the design of their training and development efforts but also the planning, implementation, and evaluation of their programs to better ensure that their employees have the information, skills, and competencies needed to carry out their work successfully. Agency Comments and Our Evaluation: We provided a draft of this report on December 12, 2003, to the Secretary of Defense, the Secretary of the Interior, the Commissioner of Internal Revenue, the Director of OPM, and the Secretary of VA. Interior, IRS, and OPM provided written comments on the draft report. In his written comments (see app. IV), Interior's Assistant Secretary for Fish and Wildlife and Parks generally agreed with the report's findings regarding the Department and the FWS. He said that the report provides important examples that can help the Department continue to move forward with additional confidence in its actions. In his written comments (see app. V), the Commissioner of Internal Revenue said that IRS was honored to share some of its lessons learned with us for governmentwide dissemination. He said that our review also provides IRS with practices from other agencies to assist IRS in its efforts to continually improve its programs. In her written comments (see app. VI), the Director of OPM said that she appreciated the opportunity for OPM to be included in the report and to share information on OPM's training and development activities and programs. Interior, IRS, and OPM also provided technical comments to clarify specific points regarding the information presented in the draft report, which we have incorporated as appropriate. In comments by E-mail through its GAO liaison, VA agreed with the information presented regarding the Department and had no additional comments on the draft report. USACE informed us that it had no comments on the draft report. As agreed with your office we are sending copies of this report to the Ranking Minority Member, Subcommittee on Oversight of Government Management, the Federal Workforce and the District of Columbia, Senate Committee on Governmental Affairs; the Chairmen and Ranking Minority Members of the Senate Committee on Governmental Affairs and the House Committee on Government Reform; and other interested congressional parties. We are also providing copies to the Secretary of the Army, the Secretary of the Interior, the Secretary of the Treasury, the Director of OPM, and the Secretary of Veterans Affairs. This report is available to others upon request. In addition, the report is available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov] http:// www.gao.gov. If you have any questions about this report, please contact me on (202) 512-6806. Key contributors to this report are listed in appendix VII. Sincerely yours, Signed by: George H. Stalcup: Director, Strategic Issues: [End of section] Appendixes: Appendix I: Objective, Scope and Methodology: The objective of this review was to provide information on selected federal agencies' experiences and lessons learned in key aspects of designing training and development programs for their employees. Specifically, we focused on the agencies' experiences and lessons learned related to: * assessing current and future agency skill and competency requirements and identifying related training and development needs, * developing strategies and solutions for training and development needs, and: * determining methods to evaluate the effectiveness of training and development programs. For this review, lessons learned were defined as knowledge that could be applied in the future that the agencies gained through either positive or negative experiences. To address this objective we focused on five agencies: the U.S. Army Corps of Engineers (USACE), Department of Defense; the U.S. Fish & Wildlife Service (FWS), Department of the Interior; the Internal Revenue Service (IRS), Department of the Treasury; the U.S. Office of Personnel Management (OPM); and the Veterans Health Administration (VHA), Department of Veterans Affairs. We chose the five agencies for a variety of reasons, including the diversity of employee occupations within the agencies, reported innovative approaches for training and developing their employees, and congressional requester interest. We selected USACE to obtain a Department of Defense perspective in light of that department's reputation as a leader in the area of training and developing military personnel. We included FWS and VHA to obtain information related to a broad mix of employee occupations. We selected IRS because of reported innovative approaches to training and development and included OPM because of its role as the federal government's human capital agency. To obtain information and related documentation, we visited the following locations: * USACE's headquarters human resources directorate in Washington, D.C., and its Professional Development Support Center in Huntsville, Alabama. * FWS's headquarters human resources division in Arlington, Virginia, and its National Conservation Training Center in Shepherdstown, West Virginia. * IRS's strategic human resources division in Arlington, Virginia, and the small business and self-employed business line. * OPM's headquarters human resources office and the career development branch of the center for retirement and insurance service. * VHA's Employee Education System headquarters in Washington, D.C., and the medical center and local network offices in Durham, North Carolina. It is important to note that our methodology was not designed to identify examples that would be representative of all training and development efforts at the five agencies in our review or of the government as a whole. We did not verify the accuracy and reliability of the data provided to us or the systems used to produce the information. Further, in citing training and development programs as examples in connection with lessons learned, we did not assess the effectiveness of the training programs and practices. Rather, our intent was to highlight and briefly describe some experiences and lessons learned that agency officials believed helped each agency improve or enhance its training and development programs. To obtain information about the five agencies' experiences and lessons learned related to designing training and development programs, we: * Interviewed agency human capital and training officials and subject matter experts responsible for agency training, performance, and other initiatives; and: * Reviewed and analyzed agency documents such as workforce plans, analyses, and reports; strategic, performance, and succession plans and reports; organizational, occupational, and unit-based competency standards; knowledge and skills inventories; skills gaps assessments; competency and skill assessments; surveys of agency employees; training plans and proposals; workforce demographic data; budget data; evaluation plans and reports; and performance measures. The lessons learned we identified for inclusion in this report were based on (1) their linkages with one or more of the eight core characteristics of a strategic training and development process, which we had identified in our previous work (see app. III) and (2) sufficient evidence from the agency to support the experiences that they relayed to us. At the exit conferences for the five agencies, we presented agency officials with the list of lessons learned that we had identified and wished to attribute to their experiences. At that time, we also informed each agency of the specific examples from their experiences that we would likely attribute to these lessons learned. In these meetings, agency officials expressed no objections to the lessons learned we had identified and in some cases provided additional information to support specific examples from their experiences that we proposed to use for this report. We conducted our audit work between August 2002 and November 2003 in accordance with generally accepted government auditing standards. [End of section] Appendix II: Background on Selected Agencies and Their Training and Development Functions: The following summarizes key information on the five agencies included in this review. These summaries include information on the agencies' missions, organizational structures, and training and development functions. U.S. Army Corps of Engineers: The U.S. Army Corps of Engineers (USACE), part of the Department of the Army within the Department of Defense, is comprised of approximately 34,600 civilian and 650 military men and women. USACE has a diverse workforce consisting of military and civilian engineers, biologists, geologists, hydrologists, natural resource managers, and other specialists who work in engineering and environmental matters. USACE's mission is to provide engineering services to the nation: (1) planning, designing, building and operating water resources and other civil works projects, (2) designing and managing the construction of military facilities for the Army and Air Force, and (3) providing design and construction management support for other Department of Defense and federal agencies. USACE headquarters office is located in Washington, D.C. USACE's Professional Development Support Center, located in Huntsville, Alabama, serves as the center of learning and training for the agency. The training center manages and implements the Proponent-Sponsored Engineer Corps Training program, which provides job-related training through technical, professional, managerial, and leadership courses for USACE and other government agencies. USACE's training center offers more than 200 courses covering topics that support the agency's mission. U.S. Fish and Wildlife Service: The mission of the U.S. Fish and Wildlife Service (FWS), a component of the Department of the Interior, is working with others to conserve, protect, and enhance fish, wildlife, and plants and their habitats for the continuing benefit of the American people. FWS's headquarters is located in Washington D.C., while its field units are located throughout the United States. FWS employs more than 9,600 people and is supported by a volunteer force of 29,000. Nearly 90 percent of FWS employees work in field locations. The National Conservation Training Center, located in Shepherdstown, West Virginia, is FWS's training center and is responsible for training a wide range of employees in the conservation community and serves as a gathering place where conservation professionals from government, nonprofit organizations, and corporations work toward common goals. Training for FWS's law enforcement personnel is primarily conducted through the Federal Law Enforcement Training Center, an interagency law enforcement training organization headquartered in Glynco, Georgia. Internal Revenue Service: The Internal Revenue Service (IRS) is a branch of the U.S. Department of the Treasury. IRS's mission is to provide America's taxpayers top quality service by helping them understand and meet their tax responsibilities and by applying the tax law with integrity and fairness to all. IRS's organizational structure includes the following business units: four operating divisions organized around four major customer segments (Wage and Investment, Small Business/Self-Employed, Large and Mid-Size Business, and Tax Exempt and Government Entities); four functional divisions (National Taxpayer Advocate, Appeals, Criminal Investigation, and Communications and Liaison); and two shared services/support divisions (Agency-Wide Shared Services and Modernization and Information Technology Services). As of March 2003, IRS had about 116,300 employees. IRS takes a decentralized approach to training and developing its workforce. Each business unit has an embedded human resources component that provides advice and analysis on related policies and issues and formulates strategies, procedures, and practices to address the unit's human capital needs. Learning and Education, one of eight major divisions comprising IRS's Office of Strategic Human Resources, provides guidance and sets policy and standards on training and development for the agency's business units and headquarters offices. U.S. Office of Personnel Management: The U.S. Office of Personnel Management (OPM), the federal government's human capital agency, provides human resources policy leadership, technical advice and assistance, and products and services to federal agencies, employees, annuitants, and job seekers. It also oversees governmentwide compensation and performance management systems, and provides retirement, health benefit, and other insurance services to federal employees, annuitants, other beneficiaries, and agencies. In March 2003, OPM completed a major restructuring process through which it consolidated various agency functions. As of March 2003, OPM employed approximately 3,500 people, many of them stationed in agency headquarters in Washington, D.C. OPM has a field presence in 16 major U.S. cities as well as operating centers in Pennsylvania and Georgia. OPM's training and development efforts are largely decentralized to the agency's various program and staff offices. The employee training and development unit within the agency's human capital management office is responsible for setting overall strategy and for planning and implementing agencywide training such as leadership development programs and various mandatory training programs. According to OPM, the agency's newly established Chief Human Capital Officer plays a significant role in advising the OPM Director on overall employee training and development initiatives and programs, as well as the establishment of the agency's training budget. Veterans Health Administration: The Veterans Health Administration (VHA), one of three major administrations within the Department of Veterans Affairs (VA), is responsible for providing primary care, specialized care, and related medical and social support services to veterans through an integrated health care system. VHA administers its functions through a group of 21 regional network offices located around the United States. As of March 2003, VHA employed about 203,500 people out of a total VA workforce of about 225,000 employees. VA takes a decentralized approach to training and development operations. VA's human resources office provides advice and guidance on training to VHA and the other departmental components but delegates training and development operations to each component. VHA's organization includes the Employee Education System, which is an internal training consulting group that provides educational services that support the workforce development and continuing education needs for VHA employees. This internal consulting group of about 300 individuals primarily helps to assess agency training needs at the national level, while VHA network offices and medical centers take lead responsibility for assessing their own local needs. These internal training consultants are available to assist VHA network offices and medical centers in designing, developing, implementing, and evaluating training and development programs to meet these local needs. [End of section] Appendix III: Core Characteristics of a Strategic Training and Development Process: The following summarizes the eight core characteristics that make a training and development process effective and strategically focused on achieving results. We identified these core characteristics as part of our recent work in developing an assessment guide to assist federal agencies in evaluating their training and development efforts.[Footnote 17] * Strategic alignment. Clear linkages exist between the agency's mission, goals, and culture and its training and development efforts. The agency's mission and goals drive a strategic training and development approach and help ensure that the agency takes full advantage of an optimal mix of strategies to improve performance and enhance capacity to meet new and emerging challenges. * Leadership commitment and communication. Agency leaders and managers consistently demonstrate that they support and value continuous learning, are receptive to and use feedback from employees on developmental needs and training results, and set the expectation that fair and effective training and development practices will improve individual and organizational performance. * Stakeholder involvement. Agency stakeholders are involved throughout the training and development process to help ensure that different perspectives are taken into account and contribute to effective training and development programs. Stakeholders' views are incorporated in identifying needed performance enhancements, developing and effectively implementing well-thought-out strategies, and helping to conceptualize and use balanced measures that accurately reflect the extent to which training and development efforts contribute toward achieving results. * Accountability and recognition. Appropriate accountability mechanisms, such as performance management systems, are in place to hold managers and employees responsible for learning and working in new ways. Appropriate rewards and incentives exist and are used fairly and equitably to encourage innovation, reinforce changed behaviors, and enhance performance. * Effective resource allocation. The agency provides an appropriate level of funding and other tools and resources--along with external expertise and assistance when needed--to ensure that its training and development programs reflect the importance of its investment in human capital to achieving its mission and goals. * Partnerships and learning from others. Coordination within and among agencies achieves economies of scale and limits duplication of efforts. In addition to benchmarking high-performing organizations, these efforts allow an agency to keep abreast of current practices, enhance efficiency, and increase the effectiveness of its training and development programs. * Data quality assurance. The agency has established policies and procedures that recognize and support the importance of quality data and of evaluating the quality and effectiveness of training and development efforts. It establishes valid measures and validated systems to provide reliable and relevant information that is useful in improving the agency's training and development efforts. * Continuous performance improvement. Agency practices and policies foster a culture of continuous improvement and optimal organizational performance regarding training and other activities. Stakeholders rely on and use program performance information and other data to assess and refine ongoing training and development efforts; target new initiatives to improve performance; and design, develop, and implement new approaches to train and develop employees. [End of section] Appendix IV: Comments from the Department of the Interior: United States Department of the Interior: OFFICE OF THE SECRETARY WASHINGTON, D.C. 20240: JAN 23 2004: Mr. George Stalcup Director, Strategic Issues U.S. General Accounting Office 441 G Street, N.W. Washington, D.C. 20548: Dear Mr. Stalcup: Thank you for providing the Department of the Interior the opportunity to review and comment on the draft U.S. General Accounting Office report entitled, "Human Capital: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs" (GAO- 04-291) dated December 12, 2003. In general, we agree with the report's findings that pertain to the Department. The report provides important examples that can help us continue to move forward with additional confidence in our actions. The Department has one specific comment. On page 37, the U.S. Fish and Wildlife Service's mission description should be changed to: "The mission of the U.S. Fish and Wildlife Service is, working with others, to conserve, protect and enhance fish, wildlife, and plants and their habitats for the continuing benefit of the American people.": We look forward to receiving the final report. Sincerely, Signed by: Craig Manson: Assistant Secretary for Fish and Wildlife and Parks: The following is GAO's comment on the Department of the Interior's letter dated January 23, 2004. GAO Comment: We have clarified the mission statement of the Fish and Wildlife Service to note its collaboration with others to accomplish its mission. [End of section] Appendix V: Comments from the Internal Revenue Service: DEPARTMENT OF THE TREASURY: INTERNAL REVENUE SERVICE: WASHINGTON, D.C. 20224: COMMISSIONER: January 12, 2004: Mr. George Stalcup: Director, Strategic Issues United States General Accounting Office Washington, D.C. 20548: Dear Mr. Stalcup: I have reviewed your draft report entitled, "Human Capital: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs" (GAO-04-291). Thank you for selecting the Internal Revenue Service to be a part of this important study. We are honored to share some of our lessons learned with you for government- wide dissemination. I am particularly pleased that we could share our work in leadership development, strategic planning, skills assessments and use of technology as best practices. The study not only recognizes some of our accomplishments but also provides us with practices from other agencies to assist us in our efforts to continually improve our programs. Strategic management of human capital is a vital part of moving the IRS forward. As a result, I have created a new Human Capital Office. Our Chief Human Capital Officer, Beverly Ortega Babers, will provide the leadership necessary to focus our resources on critical training and development programs. We will build on the good planning and design practices cited in your report to ensure that our employees have the skills and competencies they need to provide top quality service to America's taxpayers. If you have any questions, please contact me or Beverly Ortega Babers, Chief Human Capital Officer, at (202) 622-7902. Sincerely, Mark W. Everson: Signed for Mark W. Everson: [End of section] Appendix VI: Comments from the Office of Personnel Management: OFFICE OF THE DIRECTOR: UNITED STATES OFFICE OF PERSONNEL MANAGEMENT WASHINGTON, DC 20415-0001: JAN 20 2004: The Honorable David M. Walker Comptroller General: General Accounting Office Washington, DC 20548: Dear Mr. Walker: Thank you for the opportunity to respond to GAO's draft report entitled Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs. As you pointed out, it is critical that agencies invest training and development resources wisely to ensure that their employees have the necessary skills and competencies to carry out their agency's mission. Within the Office of Personnel Management (OPM), I am very committed to developing our employees and making the kinds of strategic investments that will enhance their ability to contribute to achieving OPM's important mission. This commitment is evident in OPM's current restructuring whereby OPM is now organized around mission goals and priorities. The restructuring better positions OPM to meet the President's plan to reform the Federal Government by recognizing the importance of talented employees to deliver a citizen-centered, results-oriented, and market-based Government. Furthermore, OPM's Plan for the Strategic Management of OPM's Human Capital: FY 2004-2007, lays out an integrated and goal specific approach to ensuring OPM has the talent and capacity to meet our new and expanded responsibilities. Since your staff conducted its review of our training and development programs, we have made enhancements to our programs. These enhancements are enclosed and are also included in our Plan for the Strategic Management of OPM's Human Capital: FY 2004-2007, which is also enclosed. I appreciate the opportunity for OPM to be highlighted in your report, and to share the innovative training and development activities and programs we have created. Sincerely, Signed by: Kay Coles James: Director: Enclosures: Office of Personnel Management Comments on GAO draft report: Selected Agencies' Experiences and Lessons Learned in Designing Training and Development Programs: Page 8 - regarding OPM's 2001 skills assessment... In addition to the 2001 skills assessment described in this Section, in late FY 2003, OPM conducted another skills assessment to reflect our recent restructuring and new strategic priorities. We identified mission critical occupations and competencies needed in our current and future workforce, along with initiatives to address any gaps. This business forecasting was combined with techniques to recruit, select, place and train employees and ensure an organizational climate that fosters commitment. This workforce planning initiative was a major aspect of our recently published Plan for the Strategic Management of OPM's Human Capital, 2004-2007. The Plan identifies our mission critical occupations and the most critical competencies which enhance our ability to create strategic training and development plans to help us carry out our important mission. Page 38 - regarding setting of overall training and development strategy... OPM's training and development activities were accurately addressed in this Section, however, since the time that the draft report was completed, OPM has increased the role of its Chief Human Capital Officer (CHCO). The CHCO now plays a more significant role in advising the Director on overall employee training and development initiatives and programs, as well as the establishment of the agency's training budget. Under the direction of OPM's CHCO, program offices, with the assistance from the internal Human Capital Management Office, create specific training and development strategies and activities that address mission critical competencies identified in our FY 2003 skills assessment. This strategic approach, based upon recent workforce data, better positions OPM to prioritize its training needs, and forecast funds to support those needs. [End of section] Appendix VII: GAO Contacts and Staff Acknowledgments: GAO Contacts: George H. Stalcup or Susan Ragland, (202) 512-6806: Acknowledgments: In addition to the persons named above, K. Scott Derrick, Gerard Burke, T.J. Thomson, and Thomas Davies, Jr. made key contributions to this report. (450147): FOOTNOTES [1] U.S. General Accounting Office, Human Capital: A Guide for Assessing Strategic Training and Development Efforts in the Federal Government - Exposure Draft, GAO-03-893G (Washington, D.C.: July 2003). [2] U.S. General Accounting Office, Human Capital: Key Principles for Effective Strategic Workforce Planning, GAO-04-039 (Washington, D.C.: Dec. 11, 2003). [3] In providing technical comments on a draft of this report, OPM noted that it conducted another skills assessment in late fiscal year 2003 to reflect its recent restructuring and new strategic priorities. [4] In previous work, we identified practices used by agencies in other countries to manage the succession of senior executives and other employees with critical skills. See Human Capital: Insights for U.S. Agencies from Other Countries' Succession Planning and Management Initiatives, GAO-03-914 (Washington, D.C.: Sept. 15, 2003). [5] An IDP is a written plan, cooperatively prepared by the employee and his or her supervisor, that outlines the steps the employee will take to develop knowledge, skills, and abilities in building on strengths and addressing weaknesses as he or she seeks to improve job performance and pursue career goals. These individual development plans are also known as personal development plans, personal training plans, and individual training plans. [6] For more information on key practices and implementation steps that can help agencies transform their cultures, see U.S. General Accounting Office, Results-Oriented Cultures: Implementation Steps to Assist Mergers and Organizational Transformations, GAO-03-669 (Washington, D.C.: July 2, 2003). [7] U.S. General Accounting Office, Results-Oriented Cultures: Creating a Clear Linkage between Individual Performance and Organizational Success, GAO-03-488 (Washington, D.C.: Mar. 14, 2003). [8] For more information on performance management of senior executives at IRS, see U.S. General Accounting Office, Results-Oriented Cultures: Using Balanced Expectations to Manage Senior Executive Performance, GAO-02-966 (Washington, D.C.: Sept. 27, 2002). [9] The 360-degree feedback process is designed to provide a manager direct input from various sources--supervisor, peers, customers, and subordinates--and to compare those results to a self-evaluation. With this feedback, managers can identify action items and incorporate them into their individual performance plans. [10] 31 U.S.C. 3302. [11] GIS is a system of computer software, hardware, and data used to manipulate, analyze, and graphically present a potentially wide array of information associated with geographic locations. [12] GAO-03-893G, p. 7. [13] Knowledge management is an approach to capturing, understanding, and using the collective body of information and intellect within an organization to accomplish its mission. [14] Communities of practice provide an on-line resource for peers to ask and respond to questions and share knowledge. [15] U.S. General Accounting Office, Human Capital: Design, Implementation, and Evaluation of Training at Selected Agencies, GAO/T- GGD-00-131 (Washington, D.C: May 18, 2000). [16] GAO-03-893G, p. 69. [17] GAO-03-893G, p. 75. GAO's Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.