Employment and Training Administration

More Actions Needed to Improve Transparency and Accountability of Its Research Program Gao ID: GAO-11-285 March 15, 2011

To help guide the nation's workforce development system, the Department of Labor's (Labor) Employment and Training Administration (ETA) conducts research in areas related to job training and employment. Building upon our earlier work, GAO examined the following: (1) To what extent do ETA's research priorities reflect key national employment and training issues and how useful were the studies funded under them? (2) What steps has ETA taken to improve its research program? (3) How has ETA improved the availability of its research since our last review in January 2010? To answer these questions, GAO reviewed ETA's research reports disseminated between January 2008 and March 2010 costing $1 million or more, as well as ongoing studies costing $2 million or more. GAO also convened a virtual expert panel, interviewed Labor and ETA officials, and reviewed relevant documents.

ETA's 2007 to 2012 research plan generally addressed key employment and training issues, but some studies were limited in their usefulness. Most experts on our panel reported that the areas in ETA's plan reflected key national employment and training issues at least to a moderate extent. ETA invested most of its research and evaluation resources in the areas of Unemployment Insurance and increasing labor market participation of underutilized groups. Of the $96 million that ETA invested in the 58 research reports we reviewed, more than half--or about $56 million--funded studies in these two areas. The methodological approaches and statistical procedures researchers used in the studies we reviewed were generally consistent with the questions posed, but the studies were not always useful for informing policy and practice. For example, in one study, shortcomings in the data collection phase limited the strength of the findings. Experts suggested that ETA include more varied and rigorous methodologies in its future research projects. They also reported that future research should address additional areas, including a focus on employment and training approaches that work and for whom. Labor and ETA have taken steps to improve the way research is conducted, but additional actions are needed. In acknowledging the need for more rigorous evaluations to inform its policies, Labor recently established the Chief Evaluation Office to oversee departmentwide research and evaluation efforts. In addition, ETA made changes to some of its research practices. For example, ETA has begun involving outside experts in developing its research plan. Despite these improvements, ETA's process lacks critical elements needed to ensure that current improvements become part of its routine practices. For example, ETA's process lacks a formal provision to consult with the newly established Chief Evaluation Officer at important points in the research process. Moreover, ETA's current research practices fall short of ensuring research transparency and accountability--essential elements of a sound research program. For example, its research and evaluation center lacks safeguards to protect it from undue outside influence. ETA has recently begun efforts to increase the rigor of its research designs, but has faced design and implementation challenges. For example, some of ETA's ongoing research studies face challenges in recruiting large enough sample sizes to meet the studies' objectives. ETA has improved the availability of its research findings, but it lacks a plan for assessing the usability of its Web-based search page--the primary tool for making ETA's research publicly available. ETA recently improved the timeliness with which it disseminates its research reports, decreasing the average number of days to release its reports to the public from 804 days in 2008 to 76 days in 2009. ETA has taken steps to update its online, Web-based search page. However, the agency has not developed a formal plan for assessing the overall effectiveness of its Web-based search page, including user satisfaction. In addition to its research database, ETA uses a variety of other methods to disseminate its research, including providing its research reports at conferences and internal briefings. Experts suggested that ETA consider other effective dissemination methods, such as publishing a one-page summary of research findings. GAO recommends that ETA formally include the Chief Evaluation Officer in its research process, create a mechanism to enhance the transparency and accountability of its research program, and develop a plan to ensure that research reports are accessible through its Web-based search page. Labor agreed with our recommendations and noted its ongoing efforts to improve its research program. While these efforts are important, GAO stresses the need for additional actions to fully address the recommendations.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: George A. Scott Team: Government Accountability Office: Education, Workforce, and Income Security Phone: (202) 512-5932


GAO-11-285, Employment and Training Administration: More Actions Needed to Improve Transparency and Accountability of Its Research Program This is the accessible text file for GAO report number GAO-11-285 entitled 'Employment And Training Administration: More Actions Needed to Improve Transparency and Accountability of Its Research Program' which was released on April 14, 2011. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. United States Government Accountability Office: GAO: Report to Congressional Committees: March 2011: Employment And Training Administration: More Actions Needed to Improve Transparency and Accountability of Its Research Program: GAO-11-285: GAO Highlights: Highlights of GAO-11-285, a report to congressional committees. Why GAO Did This Study: To help guide the nation‘s workforce development system, the Department of Labor‘s (Labor) Employment and Training Administration (ETA) conducts research in areas related to job training and employment. Building upon our earlier work, GAO examined the following: (1) To what extent do ETA‘s research priorities reflect key national employment and training issues and how useful were the studies funded under them? (2) What steps has ETA taken to improve its research program? (3) How has ETA improved the availability of its research since our last review in January 2010? To answer these questions, GAO reviewed ETA‘s research reports disseminated between January 2008 and March 2010 costing $1 million or more, as well as ongoing studies costing $2 million or more. GAO also convened a virtual expert panel, interviewed Labor and ETA officials, and reviewed relevant documents. What GAO Found: ETA‘s 2007 to 2012 research plan generally addressed key employment and training issues, but some studies were limited in their usefulness. Most experts on our panel reported that the areas in ETA‘s plan reflected key national employment and training issues at least to a moderate extent. ETA invested most of its research and evaluation resources in the areas of Unemployment Insurance and increasing labor market participation of underutilized groups. Of the $96 million that ETA invested in the 58 research reports we reviewed, more than half”or about $56 million”funded studies in these two areas. The methodological approaches and statistical procedures researchers used in the studies we reviewed were generally consistent with the questions posed, but the studies were not always useful for informing policy and practice. For example, in one study, shortcomings in the data collection phase limited the strength of the findings. Experts suggested that ETA include more varied and rigorous methodologies in its future research projects. They also reported that future research should address additional areas, including a focus on employment and training approaches that work and for whom. Labor and ETA have taken steps to improve the way research is conducted, but additional actions are needed. In acknowledging the need for more rigorous evaluations to inform its policies, Labor recently established the Chief Evaluation Office to oversee departmentwide research and evaluation efforts. In addition, ETA made changes to some of its research practices. For example, ETA has begun involving outside experts in developing its research plan. Despite these improvements, ETA‘s process lacks critical elements needed to ensure that current improvements become part of its routine practices. For example, ETA‘s process lacks a formal provision to consult with the newly established Chief Evaluation Officer at important points in the research process. Moreover, ETA‘s current research practices fall short of ensuring research transparency and accountability”essential elements of a sound research program. For example, its research and evaluation center lacks safeguards to protect it from undue outside influence. ETA has recently begun efforts to increase the rigor of its research designs, but has faced design and implementation challenges. For example, some of ETA‘s ongoing research studies face challenges in recruiting large enough sample sizes to meet the studies‘ objectives. ETA has improved the availability of its research findings, but it lacks a plan for assessing the usability of its Web-based search page” the primary tool for making ETA‘s research publicly available. ETA recently improved the timeliness with which it disseminates its research reports, decreasing the average number of days to release its reports to the public from 804 days in 2008 to 76 days in 2009. ETA has taken steps to update its online, Web-based search page. However, the agency has not developed a formal plan for assessing the overall effectiveness of its Web-based search page, including user satisfaction. In addition to its research database, ETA uses a variety of other methods to disseminate its research, including providing its research reports at conferences and internal briefings. Experts suggested that ETA consider other effective dissemination methods, such as publishing a one-page summary of research findings. What GAO Recommends: GAO recommends that ETA formally include the Chief Evaluation Officer in its research process, create a mechanism to enhance the transparency and accountability of its research program, and develop a plan to ensure that research reports are accessible through its Web- based search page. Labor agreed with our recommendations and noted its ongoing efforts to improve its research program. While these efforts are important, GAO stresses the need for additional actions to fully address the recommendations. View [hyperlink, http://www.gao.gov/products/GAO-11-285] or key components. For more information, contact George Scott at (202) 512- 7215 or scottg@gao.gov. [End of section] Contents: Letter: Background: ETA's Research Areas Generally Reflect Key Issues, but Some Studies Are of Limited Usefulness: ETA Has Taken Steps to Improve Its Research Program, but Additional Actions Are Needed: ETA Has Improved the Availability of Its Research but There Are Opportunities to Improve Its Search Page and Dissemination Methods: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Status of Prior GAO Recommendations to the Department of Labor, as of January 2011: Appendix II: Scope and Methodology: Appendix III: The Panel's Ratings of Key Employment and Training Issues, Populations, and Programs That ETA Should Address in Its Future Research: Appendix IV: Characteristics of Research Studies Disseminated between January 2009 and March 2010 That Cost $1 Million or More: Appendix V: Delphi Phase I and Phase II Questionnaires: Appendix VI: Experts Who Agreed to Participate in GAO's Delphi Panel: Appendix VII: Comments from the Department of Labor: Appendix VIII: GAO Contact and Staff Acknowledgments: GAO Related Products: Tables: Table 1: Research and Evaluation Terms and Definitions: Table 2: ETA's Research Areas, with Examples of Topics Covered in the 2007 to 2012 Research Plan: Table 3: Estimated Funding by Research Area for Research Studies Disseminated between January 2008 and March 2010: Table 4: Characteristics of Three Ongoing Studies: Table 5: Descriptive Statistics on Key Employment and Training Issues ETA Should Address in Its Future Research: Table 6: Descriptive Statistics on Key Populations ETA Should Address in Its Future Research: Table 7: Descriptive Statistics on Key Employment and Training Programs ETA Should Address in Its Future Research: Figures: Figure 1: ETA's Combined Evaluations and Pilots, Demonstrations, and Research Budgets from Fiscal Years 2007 to 2010: Figure 2: Key Phases of the Research Process: Figure 3: Issue Areas Ranked among the Top Three by Experts for ETA to Address in Future Research: Figure 4: Unemployment Issues Frequently Cited by Experts as Important for ETA to Address in Future Research: Figure 5: Expert Panel Views on the Extent to Which ETA's Research Has Informed Policy and Practice: Figure 6: Methodological Approaches Frequently Cited by Experts as Very Important for ETA to Address in Future Research: Figure 7: ETA's 8-Step Research Process: Figure 8: Experts' Opinion on the Effectiveness of Various Research Dissemination Methods: Abbreviations: ETA: Employment and Training Administration: Project GATE II: Project Growing America Through Entrepreneurship II: Labor: Department of Labor: OMB: Office of Management and Budget: OPDR: Office of Policy Development and Research: TEN: Training and Employment Notice: UI: Unemployment Insurance: WIA:Workforce Investment Act of 1998: WIRED:Workforce Innovation in Regional Economic Development: [End of section] United States Government Accountability Office: Washington, DC 20548: March 15, 2011: The Honorable Tom Harkin: Chairman: Subcommittee on Labor, Health and Human Services, Education, and Related Agencies: Committee on Appropriations: United States Senate: The Honorable Michael B. Enzi: Ranking Member: Committee on Health, Education, Labor, and Pensions: United States Senate: The Honorable Patty Murray: Chairman: The Honorable Johnny Isakson: Ranking Member: Subcommittee on Employment and Workplace Safety: Committee on Health, Education, Labor, and Pensions: United States Senate: Across the country, the workforce development systems' one-stop centers serve as the key access point for services that are crucial in today's economy--services that include Unemployment Insurance (UI) benefits, job training, and employment assistance. The Department of Labor's (Labor) Employment and Training Administration (ETA) is responsible for guiding the nearly $13 billion public workforce development system. Its mission is to help make the U.S. labor market function more efficiently by developing policies that lead to high- quality job training, employment, labor market information, and income maintenance services. To help shape its policies, ETA conducts evaluations and research studies on a range of employment-related topics. Congress appropriated about $103 million to ETA's research and evaluation budget line items for 2010.[Footnote 1] Over the last decade, however, ETA's research and evaluation program has fallen short in its efforts to conduct research that can help answer urgent workforce policy questions. In 2008 and again in 2009, we faulted ETA for failing to conduct research and evaluations that would lead to an understanding of what works and what doesn't.[Footnote 2] For example, in January 2010, we reported on shortcomings in ETA's research structure and processes.[Footnote 3] We found that ETA's research and evaluation center, the Office of Policy Development and Research (OPDR), * lacked independent authority to make key decisions about its research; * maintained processes that were unclear and that lacked transparency and accountability; * lacked a standard process for ensuring stakeholder involvement or other strategies to ensure that research addressed key national priorities; and: * had been slow to distribute its research findings and slow to respond to its statutory mandate to evaluate the Workforce Investment Act of 1998 (WIA). Based on these findings, we made several recommendations to Labor to improve ETA's research program. (For information on the status of those recommendations, see appendix I.) As ETA's leadership moves forward to help the nation meet its current employment challenges, questions remain about how well ETA's research has prepared the workforce development system for the challenges of today. Against this backdrop, you asked us to build upon our January 2010 review and further examine ETA's research program. Specifically, we answered the following questions: 1. To what extent do ETA's research priorities reflect key national employment and training issues and how useful were the studies funded under them? 2. What steps has ETA taken to improve its research program? 3. How has ETA improved, if at all, the availability of its research since our last review in January 2010 and what other steps could ETA take to further ensure its research findings are readily available? To address our objectives, we reviewed the 58 research and evaluation reports that ETA disseminated between January 2008 and March 2010 and assessed the methodological soundness of the 11 completed studies that cost $1 million or more. In addition, we reviewed the 10 ongoing studies costing $2 million or more to determine if research practices or the soundness of research designs had changed over time. In addition, we convened a virtual (Delphi) expert panel of academics, researchers, and consultants to obtain their opinions on ETA's research priorities and dissemination methods. To learn how ETA determines what research to conduct, we interviewed Labor officials and reviewed relevant agency and budget documents. In addition, we conducted site visits to two local workforce agencies in Pennsylvania and Virginia that are implementing ETA's ongoing research studies to obtain information about challenges and issues associated with participating in studies. To evaluate the availability of ETA's research, we analyzed dissemination time frames for all publications released between January 2008 and March 2010 and we tested the ability of ETA's research database to support searches generally available to users of research databases. (See appendix II for more details on our objectives, scope, and methodology.) We conducted this performance audit from March 2009 through March 2011 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Background: WIA sets forth various requirements for the Secretary of Labor relating to research and evaluation of federally funded employment- related programs and activities. The law calls upon the Secretary of Labor to publish in the Federal Register every 2 years a plan that describes its pilot, demonstration, and research priorities for the next 5 years regarding employment and training. Specifically, WIA requires the Secretary to: * develop the research plan after consulting with states, localities, and other interested parties; * send the plan to the appropriate committees of Congress; and: * take into account such factors as the likelihood that the results of the projects will be useful to policymakers and stakeholders in addressing employment and training problems.[Footnote 4] Within ETA, OPDR's Division of Research and Evaluation plans, conducts, and disseminates employment and training-related research and evaluations. Nearly all of the agency's research and evaluation studies are conducted under contract; these contractors represent a range of research organizations and academic institutions. Furthermore, OPDR plans and conducts its research and evaluation activities in consultation with ETA's program offices, such as the Office of Workforce Investment and the Office of Trade Adjustment Assistance. ETA's research and evaluation funding is divided into two separate budget line items: Pilots, demonstrations, and research. Efforts in this category are focused on developing and testing new ways to approach problems and to deliver services. Under WIA, pilots and demonstrations shall be carried out "for the purpose of developing and implementing techniques and approaches, and demonstrating the effectiveness of specialized methods, in addressing employment and training needs." WIA also states that the Secretary shall "carry out research projects that will contribute to the solution of employment and training problems in the United States." Evaluations. Efforts in this category are focused on continuing evaluations of certain programs and activities carried out under WIA. These evaluations must address: * the effectiveness of these programs and activities carried out under WIA in relation to their cost; * the effectiveness of the performance measures relating to these programs and activities; * the effectiveness of the structure and mechanisms for delivery of services through these programs and activities; * the impact of the programs and activities on the community and participants involved, and on related programs and activities; * the extent to which such programs and activities meet the needs of various demographic groups; and: * such other factors as may be appropriate. In program year 2010, ETA's combined budget appropriation for conducting evaluations and pilots, demonstrations, and research was about $103 million--or nearly $34 million above what the agency requested.[Footnote 5] (See figure 1.) About $84 million of the 2010 funds were designated by the Congress for specific projects, including $30 million for Transitional Jobs activities for ex-offenders, and another $5.5 million for competitive grants addressing the employment and training needs of young parents. According to agency documents, in 2008 and 2009, the Congress similarly increased ETA's requested budget for pilots, demonstrations, and research, at the same time specifically designating how the majority of those funds would be used, including $4.9 million in 2008 and $5 million in 2009 for the young parents' demonstration. Figure 1: ETA's Combined Evaluations and Pilots, Demonstrations, and Research Budgets from 2007 to 2010: [Refer to PDF for image: stacked vertical bar graph] Year: 2007; Funds requested: $22.6 million; Funds designated for congressionally mandated projects: $0; Funds not designated for congressionally mandated projects: $19.6 million; ETA's combined budgets: $19.6 million. Year: 2008; Funds requested: $20.0 million; Funds designated for congressionally mandated projects: $48.5 million; Funds not designated for congressionally mandated projects: $4.8 million; ETA's combined budgets: $53.3 million. Year: 2009; Funds requested: $25.0 million; Funds designated for congressionally mandated projects: $46.3 million; Funds not designated for congressionally mandated projects: $9.4 million; ETA's combined budgets: $55.7 million. Year: 2010; Funds requested: $69.1 million; Funds designated for congressionally mandated projects: $84.4 million; Funds not designated for congressionally mandated projects: $18.7 million; ETA's combined budgets: $103.1 million. Source: GAO analysis of ETA budget documents. [End of figure] Key Elements of Sound Research and Evaluation Programs: While there is no single or ideal way for government agencies to conduct research, several leading national organizations have developed guidelines that identify key elements that promote a sound research program.[Footnote 6] These guidelines identify five elements as key: agency resources, professional competence, independence, evaluation policies and procedures, and evaluation plans. Resources. Research should be supported through stable, continuous funding sources and through special one-time funds for evaluation projects of interest to executive branch and congressional policymakers. Professional competence. Research should be performed by professionals with appropriate training and experience for the evaluation activity (such as performing a study, planning an evaluation agenda, reviewing evaluation results, or performing a statistical analysis). Independence. Although the heads of federal agencies and their component organizations should participate in establishing evaluation agendas, budgets, schedules, and priorities, the independence of evaluators must be maintained with respect to the design, conduct, and results of their evaluation studies. Evaluation policy and procedures. Each federal agency and its evaluation centers should publish policies and procedures and adopt quality standards to guide evaluations within its purview. Such policies and procedures should identify the kinds of evaluations to be performed and the criteria and administrative steps for developing evaluation plans and setting priorities, including selecting evaluation approaches to use, consulting experts, ensuring evaluation product quality, and publishing reports. Evaluation plans. Each federal agency should require its major program components to prepare annual and multiyear evaluation plans and to update these plans annually. The planning should take into account the need for evaluation results to inform program budgeting, reauthorization, agency strategic plans, program management, and responses to critical issues concerning program effectiveness. These plans should include an appropriate mix of short-and long-term studies to produce results for short-or long-term policy or management decisions. To the extent practical, the plans should be developed in consultation with program stakeholders. Furthermore, leading organizations, including the American Evaluation Association and the National Academy of Sciences, emphasize the need for research programs to establish specific policies and procedures to guide research activities. Based on several key elements identified by these organizations, we developed a framework comprised of five phases--agenda setting, selecting research, designing research, conducting research, and disseminating research results. (See figure 2.) Figure 2: Key Phases of the Research Process: [Refer to PDF for image: illustration] Agenda setting: Identify general research priority areas where the agency will focus its attention. Selecting research: Identify the criteria agencies use to select the research studies that they will fund. Designing research: Develop research questions and appropriate methodologies for research projects. Conducting research: Perform and monitor the implementation of research projects. Disseminating research: Make available sound research to policymakers, researchers, government officials, and the general public. Source: GAO analysis of key elements of a sound research process identified by leading national organizations,including the American Evaluation Association and the National Academy of Sciences. [End of figure] Agenda setting. Agencies should establish a structured process for developing their research priorities. The process should identify how agencies set research priority areas and provide for updating the areas on a regular basis. The process should also allow for the consideration of critical issues and state how internal and external stakeholders will be included in developing the plan. Selecting research. At this phase, the process should identify how the research program's staff identifies and selects studies to fund, including the criteria it uses to make those decisions. Steps might describe how the staff assembles a list of potential studies, works with internal program offices, and makes final decisions. Designing research. During the design phase, the process should identify steps taken to select appropriate research approaches and methods and the safeguards in place to ensure appropriate tradeoffs are made between what is desirable and what is practical and between the relative strengths and weaknesses of different methods. Conducting research. At this stage, the process should include policies and procedures to guide the conduct of research. The process should ensure that key events, activities, and time frames are specified and that knowledgeable staff in the sponsoring agency monitor the implementation of the research. Disseminating research. This process should describe how research findings are made available to the public and disseminated to all potential users. These dissemination methods should include safeguards to ensure research findings are disseminated in a timely manner and are accessible through the Internet with user-friendly search and retrieval technologies. Research Terminology in This Report: In this report, we use several technical terms in describing ETA's research designs and study characteristics. (See table 1.) Table 1: Research and Evaluation Terms and Definitions: Term: Bias; Definition: The extent to which a measurement or a sampling or analytic method systematically underestimates or overestimates a value. Term: Cross-sectional data; Definition: Observations collected on subjects or events at a single point in time. Term: Descriptive study; Definition: Provides descriptive information about specific conditions or events, such as the number of individuals who received a particular type of job training. Term: Experimental design; Definition: Units of study are randomly assigned to a treatment group or to one or more comparison groups. The program's effects are estimated by comparing outcomes for the treatment group with outcomes for each comparison group. Term: External validity/generalizability (used interchangeably with generalizability); Definition: The extent to which a finding applies (or can be generalized) to persons, objects, settings, or times other than those that were the subject of study. Term: Longitudinal data; Definition: Sometimes called "time series data," observations collected over a period of time; the sample (instances or cases) may or may not be the same each time but the population remains constant. Term: Meta-analysis; Definition: A systematic approach to summarizing or analyzing evaluation findings across studies, usually quantitative, allowing analysis of variation in program effect sizes by study methods or program characteristics. Term: Random assignment; Definition: A method for assigning subjects to two or more groups by chance, as in the flip of a coin. Term: Secondary analysis; Definition: The reanalysis of data collected in previous studies or surveys to address new questions or use methods not previously employed. Term: Quasi-experimental design; Definition: Units of study are assigned to a treatment and a comparison group without random assignment. Term: Representative sample; Definition: A sample that has approximately the same distribution of characteristics as the population from which it was drawn. Term: Treatment group; Definition: The subjects of the intervention being studied. Source: GAO analysis of program evaluation and methodology documents. [End of table] ETA's Research Areas Generally Reflect Key Issues, but Some Studies Are of Limited Usefulness: Experts Thought ETA's 2007 to 2012 Research Plan Reflected Key Areas, but They Also Suggested New Ones for Future Research: Our expert panel generally considered ETA's research areas to be the right ones for the period the research plan covered. About three- fourths of the panel members reported that ETA's 2007 to 2012 research agenda reflected key employment and training issues to at least a moderate extent. However, a few experts commented that some of ETA's research areas may be too broad and lack specificity. The areas in ETA's 2007 to 2012 research plan covered a range of issues, from job training to postsecondary education. Table 2 illustrates the scope of ETA's research areas. Table 2: ETA's Research Areas, with Examples of Topics Covered in the 2007 to 2012 Research Plan: Research area: Integration of workforce and regional economic development; Example topic areas[A]: * Job training initiatives to produce high-skill, high-wage jobs; * Strategic partnerships between private business sector and public entities; * Regional and economic development. Research area: Methods of expanding U.S. workforce skills; Example topic areas[A]: * Rapid response services for dislocated workers; * Costs and benefits of apprenticeship; * Trade adjustment assistance. Research area: Increasing the Labor market participation of underutilized populations; Example topic areas[A]: * Enhanced services for the hard-to-employ; * Employment-centered programs for ex-offenders; * Current strategies to employ and retain older workers. Research area: Using state-level administrative data to measure progress and outcomes; Example topic areas[A]: * Examination of local workforce investment areas; * WIA and employment outcomes. Research area: Postsecondary education and job training; Example topic areas[A]: * The role of community colleges in workforce development and training; * Career advancement accounts; * Growing America through entrepreneurship. Research area: Unemployment Insurance; Example topic areas[A]: * UI benefits; * Characteristics of unemployed workers. Source: GAO analysis of ETA's research agenda. [A] This is not a complete or comprehensive list of examples of the types of research conducted by ETA under each research area. [End of table] With regard to the specific studies within these research areas, ETA invested most of its research and evaluation resources in work that focused on increasing the labor market participation of underutilized workers and on UI. Of the estimated $96 million that supported the 58 research reports published between January 2008 and March 2010, more than half--about $56 million--funded research that addressed these two research areas. Other areas received far less funding. For example, funding for studies addressing the methods of expanding U.S. workforce skills and using state-level administrative data to measure progress and outcomes accounted for about $6.5 million, or about 6.7 percent of the cost of studies published during the period we examined. (See table 3.) Overall, the individual studies that ETA funded addressed a wide variety of issues and ranged in cost from about $15,000 to a high of about $22 million. Table 3: Estimated Funding by Research Area for Research Studies Disseminated between January 2008 and March 2010: Research area: Increasing the Labor market participation of underutilized populations; Total amount: $28.9 million. Research area: Unemployment Insurance; Total amount: $26.8 million. Research area: Postsecondary education and job training; Total amount: $12.4 million. Research area: Integration of workforce and regional economic development; Total amount: $9.9 million. Research area: Methods of expanding U.S. workforce skills; Total amount: $4.0 million. Research area: Using state-level administrative data to measure progress and outcomes; Total amount: $2.5 million. Research area: Other[A]; Total amount: $11.6 million. Research area: Total; Total amount: $96.1 million. Source: GAO analysis of study cost information provided by ETA. [A] "Other" includes studies that were not part of ETA's 2007 to 2012 priorities. [End of table] In addition to the research areas covered in ETA's 2007 to 2012 research plan, experts from our virtual panel suggested that ETA incorporate additional research areas in its future research agenda. Of the research areas identified, over half of our experts (28 of 39) ranked the identification of employment and training approaches that work, and for whom, as one of the top areas that ETA's future research should address. (See figure 3.) Without such focus, experts commented that it will be difficult to know how to improve the nation's workforce system. Other issues ranked at the top by experts included research on job creation strategies and the impact of long-term and short-term training. (See appendix III for more information on issue- area rankings.) Figure 3: Issue Areas Ranked among the Top Three by Experts for ETA to Address in Future Research: [Refer to PDF for image: horizontal bar graph] Which areas would you rank as the top 3 areas that ETA should address in future research? Employment and training approaches that work and for whom: Number of respondents answering: 28. Job creation strategies: Number of respondents answering: 18. Impact of long-term and/or short-term training: Number of respondents answering: 18. Long-term outcomes of employment and training programs: Number of respondents answering: 15. Effective performance measurement systems: Number of respondents answering: 9. Issues related to Unemployment Insurance: Number of respondents answering: 8. Linkages between the public workforce system and economic development entities: Number of respondents answering: 6. Short-term outcomes of employment and training programs: Number of respondents answering: 4. Needs of different labor markets or industries: Number of respondents answering: 4. One-stop center management and operations: Number of respondents answering: 3. Value of various credentials: Number of respondents answering: 3. Employment and training strategies for various economic conditions: Number of respondents answering: 1. Source: GAO‘s survey of ETA‘s research priorities and dissemination methods. [End of figure] In addition to identifying overall employment and training areas, including issues related to UI, experts also identified specific aspects of the UI system that could be examined in ETA's future research. In particular, most experts (34 of 39 respondents) reported that it would be at least moderately important, in the future, for ETA to research the linkage between UI and employment and safety net programs, such as Temporary Assistance for Needy Families or the Supplemental Nutrition Assistance Program. (See figure 4.) This area of research may be particularly important given the role that these programs play in supporting individuals during economic downturns. In addition, many experts (24 of 39 respondents) mentioned that ETA should make the examination of the incentives and disincentives in the UI system a research priority, given the challenge of supporting unemployed workers during difficult economic times, while promoting self-sufficiency through employment. Figure 4: Unemployment Issues Frequently Cited by Experts as Important for ETA to Address in Future Research: [Refer to PDF for image: stacked horizontal bar graph] Linkage between UI programs and safety net programs: Number of respondents answering: Not at all important: 0; Somewhat important: 8; Moderately important: 8; Very important: 11; Extremely important: 15. Examination of incentives and disincentives of UI system: Number of respondents answering: Not at all important: 0; Somewhat important: 5; Moderately important: 9; Very important: 17; Extremely important: 7. Modernization of UI system: Number of respondents answering: Not at all important: 0; Somewhat important: 3; Moderately important: 13; Very important: 15; Extremely important: 6. Costs and benefits of UI system in different business cycles: Number of respondents answering: Not at all important: 1; Somewhat important: 6; Moderately important: 8; Very important: 18; Extremely important: 3. Accountability mechanisms in UI system: Number of respondents answering: Not at all important: 1; Somewhat important: 7; Moderately important: 12; Very important: 15; Extremely important: 2. Source: GAO‘s survey of ETA‘s research priorities and dissemination methods. [End of figure] Experts also reported that it is important to fund research on what works for selected population groups. Of the population groups identified, the experts on our virtual panel most often ranked the long-term unemployed, economically disadvantaged workers, and adults with low basic skills as the top populations on which to focus future research. Specifically, several experts commented that research could help to identify the challenges some of these groups face, as well as identify effective strategies that may help these population groups obtain employment. (See appendix III for a complete list of responses to these items.) In addition to population groups, experts also identified several employment and training programs that they believe warrant research attention. In particular, experts most often ranked three components of the WIA program--WIA Adult, WIA Dislocated Workers, and WIA Youth [Footnote 7]--as key to evaluate in ETA's future research. Among those three, WIA Adult was ranked the highest. (See appendix III for a complete list of experts' responses on employment and training programs to evaluate.) ETA's Research Studies Generally Answered the Questions Posed, but Their Usefulness Was Limited: Research organizations and academic institutions with responsibility for implementing ETA-funded research generally used methodologies appropriate for the questions posed, but the studies were not always useful for informing policy and practice. From January 2008 through March 2010, ETA published 17 large research and evaluation reports--14 evaluations and 3 research reports--that each cost $1 million or more. Four of these reports were designed to demonstrate what works and for whom. Each of these four reports compared the employment-related outcomes of individuals or regions who participated in training or employment programs with the employment outcomes of similar individuals who did not participate in the programs. The remaining 13 reports were descriptive and were not designed to assess program outcomes. In several studies we examined that cost $1 million or more, we found that, for a number of reasons, ETA's research studies were limited in their usefulness and in their ability to inform policy and practices. For example, in a study of the Prisoner Re-entry Initiative, shortcomings in the data collection phase limited the strength of the findings and, as a result, limited the study's opportunity to influence policy directions. Among other things, while the study provided information on employment-centered opportunities for ex- offenders, the study relied on self-reported baseline data, did not account for differences across sites where services were received, lacked the capacity to record differences in the intensity of those services, and researchers failed to ensure that data collectors were properly trained. In another study, researchers did not control for bias in selecting participants, compromising their ability to draw conclusions about the cause and effect of program outcomes. Authors of this study on the Workforce Innovations in Regional Economic Development (WIRED) initiative acknowledged that the study would be unable to attribute outcomes to program services because it did not use random assignment in selecting participating regions. We have previously criticized ETA for failing to adequately provide for evaluating the effectiveness of its WIRED initiative.[Footnote 8] Moreover, some studies were limited due to observation periods that did not match the needs of the studies' objectives. For example, an evaluation of an entrepreneurship training project was unable to assess the effectiveness of the project in meeting its long-term goals of increasing business ownership and self-sufficiency because the time frames for the study were too short. In this study, data collection was limited to 18 months after participants were randomly assigned, a period far shorter than the 60-month period recommended by experts. [Footnote 9] (See appendix IV for additional information on the methodological characteristics of these studies.) Experts generally agreed that ETA's research had limited usefulness in informing policy and practice. Over one-third of the 39 experts reported that over the past 5 years, ETA's research informed employment and training policy and state and local practices to a little extent or not at all. (See figure 5.) Some experts commented that the design of these studies and the length of time to complete them and disseminate results reduced their usefulness. For example, many of the reports that we reviewed costing $1 million or more were multiyear projects that took, in most cases, about 3 to 5 years to complete. Some experts commented that the inclusion of shorter-length studies may be useful in times of rapidly changing economic conditions. At least one expert noted that some mixed-methods studies would be useful--studies that would allow for short-term interim findings that could facilitate changes in practice during the course of the research study. Figure 5: Expert Panel Views on the Extent to Which ETA's Research Has Informed Policy and Practice: [Refer to PDF for image: stacked horizontal bar graph] In the past 5 years, to what extent has ETA‘s research informed employment and training policy: Number of respondents answering: Don‘t know: 1; Not at all: 1; Little extent: 12; Some extent: 17; Moderate extent: 8; Great extent: 0. In the past 5 years, to what extent has ETA‘s research informed state and local employment and training practices: Number of respondents answering: Don‘t know: 4; Not at all: 1; Little extent: 15; Some extent: 13; Moderate extent: 5; Great extent: 1. Source: GAO‘s survey of ETA‘s research priorities and dissemination methods. [End of figure] Members of our expert panel stressed the importance of ETA incorporating varied methodological approaches into its future research proposals to best position the agency to address key employment and training issues. Twenty-seven of the 39 experts reported it was very important that ETA evaluate its pilots and demonstrations. Twenty-three reported that it was very important that more randomized experimental research designs be integrated into ETA's future research.[Footnote 10] (See figure 6.) While several experts noted that these randomized experiments will allow ETA to identify the effectiveness of particular interventions or strategies, at least one expert suggested that ETA should be strategic in choosing the interventions it tests more rigorously, basing those decisions on what appears most promising in preliminary studies. Figure 6: Methodological Approaches Frequently Cited by Experts as Very Important for ETA to Address in Future Research: [Refer to PDF for image: stacked horizontal bar graph] Evaluation of pilots and demonstrations: Number of respondents answering: Very unimportant: 0; Somewhat unimportant: 2; Neither important nor unimportant: 0; Somewhat important: 9; Very important: 27. Randomized experimental design: Number of respondents answering: Very unimportant: 0; Somewhat unimportant: 0; Neither important nor unimportant: 3; Somewhat important: 12; Very important: 23. Quasi-experimental design: Number of respondents answering: Very unimportant: 1; Somewhat unimportant: 1; Neither important nor unimportant: 3; Somewhat important: 17; Very important: 16. Analysis of administrative data for research purposes: Number of respondents answering: Very unimportant: 0; Somewhat unimportant: 1; Neither important nor unimportant: 5; Somewhat important: 19; Very important: 14. Descriptive studies: Number of respondents answering: Very unimportant: 0; Somewhat unimportant: 3; Neither important nor unimportant: 6; Somewhat important: 18; Very important: 12. Meta-analyses (i.e., review or analysis of prior research): Number of respondents answering: Very unimportant: 2; Somewhat unimportant: 2; Neither important nor unimportant: 11; Somewhat important: 18; Very important: 6. Secondary analysis of existing research data: Number of respondents answering: Very unimportant: 1; Somewhat unimportant: 4; Neither important nor unimportant: 13; Somewhat important: 17; Very important: 4. Source: GAO‘s survey of ETA‘s research priorities and dissemination methods. [End of figure] Furthermore, 16 of the 39 experts also reported that it is very important for ETA to consider including more quasi-experimental studies in the future. As previously discussed, such studies would include designs that compare outcomes between groups with similar characteristics, but do not use random assignment. By including more quasi-experimental designs, ETA may be able to better understand the link between services and outcomes in those settings where random assignment is not possible, ethical, or practical.[Footnote 11] ETA Has Taken Steps to Improve Its Research Program, but Additional Actions Are Needed: Labor Has Taken Steps to Reform Its Research Program: Labor has taken several steps designed to improve the way it conducts research, both at the department level and within ETA. Department-level efforts. Labor has changed the organizational structure of research within the department. In 2010, acknowledging the need for better and more rigorous evaluations to inform its policy, Labor established the Chief Evaluation Office to oversee the department's research and evaluation efforts. The office, which resides within the Office of the Assistant Secretary for Policy, has no authority to direct research within Labor's agencies, according to officials. It does, however, manage evaluations supported by funds from a departmentwide account, oversee departmentwide evaluations, and provide consultation to Labor agencies, including ETA. Specifically, the office is responsible for creating and maintaining a comprehensive inventory of past, ongoing, and planned evaluation activities within Labor and for ensuring that Labor's evaluation program and findings are transparent, credible, and accessible to the public. In fiscal year 2010, the Chief Evaluation Office had an estimated budget of $8.5 million, and two of its four staff were on board by the beginning of fiscal year 2011. ETA efforts. ETA has recently made changes to some of its research practices--chief among them is the involvement of stakeholders and outside experts in the research process. We previously criticized ETA for failing to consistently involve a broad range of stakeholders, outside experts, or the general public in deciding what areas of research it should undertake. We recommended that ETA take steps to routinely involve outside experts in the research agenda-setting process.[Footnote 12] For the upcoming 2010 to 2015 research plan, ETA has awarded a grant to the Heldrich Center at Rutgers University to convene an expert panel to help inform the research plan. The center is expected to issue a report in May 2011 that outlines the panel's recommendations for research areas to include in the plan. In addition, ETA will work with other Labor agencies, as well as the Departments of Education and Health and Human Services, before finalizing its research agenda. Officials told us that they will also solicit public comments before the research plan is finalized. In addition to engaging stakeholders, ETA has also established a formal research process. As we previously reported, ETA developed and documented its research process in 2007.[Footnote 13] The agency's actions were in response to a request by the Office of Management and Budget (OMB) to establish more formal policies and procedures to guide its research--a request that came out of OMB's concerns about the manner in which ETA's research was being carried out. Prior to 2007, ETA lacked a documented research process, and its research was often conducted in an ad hoc manner. ETA's current research process identifies the steps, activities, and time frames it uses to carry out its research. Figure 7 illustrates critical components of ETA's 8-step research process. Figure 7: ETA's 8-Step Research Process: [Refer to PDF for image: illustrated table] Research stage: Agenda setting; Steps in ETA‘s process: No formalized process; Critical components of each step: (none). Research stage: Selecting research; Steps in ETA‘s process: ETA Step 1: Select projects for evaluation and pilots, demonstrations, and research (PD&R) funding; Critical components of each step: * Review ETA/Labor stated priorities and recommendations in the 5-year plan and solicit project ideas from within; * Review congressional mandates and requests to co-fund projects from other agencies; * Review existing project commitments; * Submit agenda to Assistant Secretary of ETA and finalize plan; * Prepare approval memos for approved projects. Research stage: Designing research; Steps in ETA‘s process: ETA Step 2: Design Statements of Work (SOW), Requests for Proposal (RFP) and Solicitations for Grant Applications (SGA); Critical components of each step: * Review time frames, available funding, and methodology for each project idea and determine appropriate funding method; * Prepare draft reflecting research questions and methodology and obtain feedback from program offices; * Revise SOWs, RFPs, or SGAs, as appropriate, and obtain final approval from Assistant Secretary; * Submit to OMB for review any SGA, SOW, or RFP that costs $250,000 or more. Research stage: Designing research; Steps in ETA‘s process: ETA Step 3: Contractor/grantee selection process; Critical components of each step: * Work with the Office of Grants and Contracts Management to select contractors or grantee; * Ensure contracts and/or grants are awarded in accordance with Federal Acquisition Regulations and departmental procurement rules. Research stage: Conducting research; Steps in ETA‘s process: ETA Step 4: Initial project design (Post-award implementation); Critical components of each step: * Hold initial meetings to discuss study objectives and gather information; * Review contractor‘s draft design report and approve? Review grantee‘s implementation plans and approve; * Submit contractor‘s final design report, and if applicable, submit report for peer review. Research stage: Conducting research; Steps in ETA‘s process: ETA Step 5: Paperwork Reduction Act clearance process; Critical components of each step: * Prepare and submit an information collection request for internal review to reduce burdens associated with federal information collection and reporting requirements; * Inform the public about the information collection request and allow 60 days for comment; * Finalize information collection request and submit to OMB for review and approval. Research stage: Conducting research; Steps in ETA‘s process: ETA Step 6: Project monitoring; Critical components of each step: * Monitor project implementation and raise any issues to senior management; * Develop corrective action plans (as needed) and review/approve project progress and finance reports. Research stage: Disseminating research; Steps in ETA‘s process: ETA Steps 7 and 8: Review and dissemination of reports. Critical components of each step: * Review and finalize report deliverables, working with contractors and internal offices to incorporate revisions; * Finalize research reports and obtain approval from the Assistant Secretary; * Produce Training and Employment Notice (TEN) and abstract to announce the availability of approved research reports; * Disseminate reports through ETA‘s research Web site and in hard copy (as available); * Alert OMB of reports submitted for dissemination and not approved within 9 months. Source: GAO analysis of ETA documentation. [End of figure] ETA's process contains several of the key elements identified by leading organizations as important for guiding research activities. For example, the process includes specific steps the agency should take to identify the types of evaluations it will perform, as well as the administrative steps it should take to develop evaluation plans and select the research projects to fund.[Footnote 14] In addition, the process also specifies key events and time frames, and provides for monitoring the implementation of the research. For example, the process stipulates that ETA should alert OMB of research reports that have not been approved for dissemination within 9 months of being submitted and allows contractors to publicly release their research reports within those same time frames.[Footnote 15] Some Areas of ETA's Research Program Merit Further Attention: Despite ETA's efforts, more action is needed to improve its research program. While ETA has taken steps to document its research process, its process lacks specific details in some areas, creating ambiguities that could undermine efforts to adhere to a formal process. For example, as we previously reported, its process lacks clear criteria, such as a dollar threshold or a particular methodological design feature, for determining which projects require peer review. And while the process specifies the actions project officers should take if reports are not released in a timely manner, it does not specify the consequences for failing to do so. We previously recommended that ETA establish more specific processes, including time frames for disseminating research reports. ETA has taken some action, such as revising the performance standards for project officers to hold them accountable for meeting time frames, but these steps do not fully satisfy the recommendation because the changes are not yet reflected in the formal research process. Moreover, ETA's process is missing some critical elements that are needed to ensure that the current improvements become routine practices. * Consulting with the Chief Evaluation Officer. ETA's process lacks a formal provision requiring consultation with the newly established Chief Evaluation Officer at important points in the research process. For example, it contains no provision for consulting with the Chief Evaluation Officer when developing its annual list of research projects or when determining how ETA will invest its research and evaluation resources. Such consultation could help Labor better coordinate its research and evaluation efforts and better leverage its research funding. Moreover, the process contains no provision for involving the Chief Evaluation Officer in the early stages of developing its research projects. In the recent past, Labor officials told us that ETA has had difficulty developing requests for research and evaluation proposals that can pass OMB technical reviews. In particular, OMB has been critical of ETA's research designs because they failed to provide for adequate sample size and appropriate methodologies that are needed to obtain useful results. In addition, OMB has also expressed concerns with ETA's reliance on process evaluations rather than focusing on outcomes. These difficulties have resulted in delays in the research process. ETA has begun to consult with the Chief Evaluation Officer; however, these consultations are not a routine component in the formal process. * Setting the research agenda. ETA's current process, as documented, begins with phase two--selecting specific research studies--and misses the important first step of setting the overall research agenda. This first phase of the process should include the steps that ETA will take to establish its research priorities and to update them on a regular basis. It should also include provisions for ensuring critical issues are considered and internal and external stakeholders are included in developing the plan. Officials noted that they plan to incorporate the agenda-setting phase into its formal process, but have not yet done so. Setting the research agenda is key to ensuring that an appropriate mix of studies is included in future research. Failing to make this phase part of the formal process, including the specific steps to involve outside stakeholders that are currently under way, may leave ETA with little assurance that these efforts will continue in the future. Beyond ETA's process for conducting research, current research practices fall short of ensuring research transparency and accountability--essential elements of a sound research and evaluation program. The research program has few, if any, safeguards to protect it from undue influence. According to officials, at times in the past decade, many key research decisions have been made outside of the office that is responsible for research. For example, decisions about which research studies would and would not be publicly released were made at the highest levels within ETA, and the criteria used to make those decisions were unclear. Of the 34 reports that ETA released to the public in 2008, 20 had waited between 2 and 5 years to be approved for public release.[Footnote 16] Several reports that had experienced long delays had relatively positive and potentially useful findings for the workforce system, according to our analysis. Among the studies delayed by almost 5 years was an evaluation of labor exchange services in the one-stop system that found certain employment services to be highly cost-effective in some situations. Another study, delayed for about 3.5 years, was a compendium of past and ongoing experimental studies of the workforce system, including early findings and recommendations for future research. In our previous report, we noted that ETA's research and evaluation center lacked a specific mechanism to insulate it from undue influence. We reported that other federal agencies, such as the Department of Education's Institute of Education Sciences and the National Science Foundation, engage advisory bodies in the research process. While not without tradeoffs in terms of additional time and effort, such an approach may serve to protect the research program from undue influence and improve accountability. ETA is currently involving outside experts in setting the research agenda for 2010 to 2015, but is not involving experts more broadly on research policy and practices. ETA Has Recently Included More Random Assignment Studies in Its Research Program: ETA has recently begun to include more rigorous studies in its ongoing research. Of the 10 large, ongoing studies costing $2 million or more that began during the period of our review, three--the WIA Gold Standard Evaluation of the Adult and Dislocated Worker Programs, the Impact Evaluation of the Young Parents Demonstration, and the Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II)--use experimental design with random assignment, as recommended by our experts.[Footnote 17] These ongoing studies--which range in cost from $2 million to nearly $23 million--have the potential to determine the effectiveness of some of the program services. Table 4 outlines some key characteristics of these three studies. Table 4: Characteristics of Three Ongoing Studies: Study: Purpose; WIA Gold Standard Evaluation: This experimental evaluation will provide rigorous, nationally representative estimates of the WIA's net impacts. The study will also produce a detailed description of the program's implementation and estimates of its benefits and costs. Key questions include: * Does access to WIA-intensive training services lead adult and dislocated workers to achieve better educational and employment outcomes than they would achieve in the absence of those services? * Does the effectiveness of WIA vary by population subgroup? * Is the effectiveness of WIA--and its components--commensurate with its costs? Impact Evaluation of the Young Parents Demonstration: This experimental evaluation will determine the impacts of supplemental short-term mentoring services to complement core workforce development programs, including the following: * Measure differences in services that the participants in the treatment and control groups receive; * Assess how services received differ across sites and among individuals in the treatment and control groups; * Identify the resource requirements involved in providing services and identify implementation issues that might have affected outcomes; Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II): This experimental evaluation will compare the outcomes of randomly assigned Project GATE II participants to the outcomes of individuals who did not receive Project GATE II services; * Measure the impacts of Project GATE on participants' labor market and self-employment outcomes; * Test the viability of providing Project GATE services to dislocated workers in rural areas over the age of 45; * Assess the linkages between the local one-stop centers' programs and other key groups providing services. Study: Scope; WIA Gold Standard Evaluation: Thirty local workforce investment areas, selected at random, stratified to ensure that they are representative in terms of the size, geography, and customer training rates. Following this selection, a subset of participants at each of the 30 selected sites will be randomly assigned to one of three groups; Impact Evaluation of the Young Parents Demonstration: Five to 10 sites serving approximately 2,500 participants; Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II): Examination of four state grants. Study: Performance period; WIA Gold Standard Evaluation: 7 years. A subset of participants will be tracked and administered surveys at 15 and 36 months after random assignment; Impact Evaluation of the Young Parents Demonstration: 7 years. A sample of participants will be tracked and administered a follow-up survey approximately 12 months after random assignment. An additional survey will be conducted 30 months after random assignment; Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II): 3 years. A sample of project participants will be tracked and administered a follow-up survey approximately 12 months after random assignment. Study: Peer review; WIA Gold Standard Evaluation: The contractor will coordinate three meetings of the peer review board made up of 10 experts. ETA will select seven and the contractor three. The purpose of the board will be to review and comment on the project methodology and key project deliverables; Impact Evaluation of the Young Parents Demonstration: The contractor will convene up to four peer review panels composed of four to five researchers selected by ETA to review the evaluation methodology, analytic approaches and key deliverables; Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II): No requirements under contract. However, ETA officials told us that they have plans for an independent peer review. Study: Cost; WIA Gold Standard Evaluation: Total: $22,951,040 (Evaluation contract); Impact Evaluation of the Young Parents Demonstration: Total: $6,154,570; (Impact evaluation contract: $4.8 million; process evaluation: $717,000; and technical assistance contracts: $199,998 and $437,572); Evaluation of Project Growing America Through Entrepreneurship II (Project GATE II): Total: $2,014,996; (Impact evaluation: $1.5 million; technical assistance contracts: $485,000 and $29,996). Source: GAO analysis of ETA's documents. [End of table] Experimental designs with random assignment are an important means to understand whether various program components or services are effective, but they are also often difficult to design and implement in real-world settings.[Footnote 18] For example, in doing evaluations of employment and training programs, researchers often have difficulty in recruiting sample sizes large enough to detect meaningful outcomes. Because employment and training services may vary by location, and participants and their socio-economic environments are diverse, researchers must find ways to standardize procedures and treatment or service options. This often means recruiting relatively large samples. However, studies can be intrusive, often requiring program sites to change how they operate or to increase the resources available to participants. As a result, recruiting sites and sufficient numbers of participants may be difficult. Some of ETA's ongoing research studies face challenges in recruiting sample sizes large enough to meet the studies' objectives. For example, based on an OMB review, it was determined that the sample size for the Impact Evaluation of the Young Parents Demonstration had to be much larger in order to be able to assess the effectiveness of the program. At that time, ETA had already awarded two phases of grants. After consulting with the new Chief Evaluation Officer, ETA changed the number of participants required for the third phase from 100 to 400 to obtain a sample large enough to address OMB's concerns and provide reliable estimates. However, grantees found it difficult to recruit even the 100 participants in the smaller sample, and it remains unclear whether they will be able to recruit all of the needed participants for the expanded design. The WIA Gold Standard Evaluation of the Adult and Dislocated Worker Programs: The WIA Gold Standard Evaluation illustrates ETA's difficulties in planning and executing large-scale, rigorous random assignment studies. WIA required that the Secretary of Labor conduct at least one multi-site control-group evaluation of the services and programs under WIA by the end of fiscal year 2005.[Footnote 19] ETA, however, delayed executing such a study, finally soliciting proposals in November 2007 and awarding the contract in June 2008. The contractor submitted the initial design report in January 2009 and provided ETA with design revisions in May 2010. Officials tell us researchers will soon begin randomly assigning participants. ETA expects to receive the first report (on implementation) during the winter of 2012-2013 and the final report in 2015--10 years later than the WIA-mandated time frame. An OMB-selected panel of government experts--a technical working group composed of experts chosen by ETA, the evaluation contractor, and OMB staff--reviewed the original design for this study.[Footnote 20] Reviewers agreed the design contained many strengths, including: * the selection of an experimental design and a net impact approach; * the addition of a process or implementation study to evaluate differences among sites and other implementation and data collection issues; * the use of administrative and survey data; * the collection of information on services received by participants in the control group; and: * the collection of a wide range of outcome data for participants. However, reviewers raised several concerns regarding the design. For example, they were skeptical that the researchers would be able to obtain a sufficiently large and representative sample to draw meaningful conclusions about the effectiveness of the national workforce system. In order to maximize participation, officials told us that the Assistant Secretary of ETA made personal phone calls to all selected sites to emphasize the importance of the study, offered an open door policy to site officials to discuss issues, and followed up with an appreciation letter. Furthermore, ETA required the evaluation contractor to provide reimbursement payments to each site to offset implementation costs. Reviewers also had several other concerns regarding which groups would be included in the study and which groups would not. For example, some experts raised concerns about getting accurate information on the youth program because of the large, one-time infusion of funds the program received from the American Recovery and Reinvestment Act of 2009. Reviewers were further concerned about the appropriateness of the evaluation objectives, the adequacy of steps taken to account for the effect of variation in services across sites on evaluation outcomes, and the external validity or generalizability of the study. In order to address these concerns, ETA made substantial adjustments to the original design. Specifically, ETA officials told us that based on an agreement with OMB, they instructed the contractor to drop the youth component from the evaluation and to focus only on the Adult and Dislocated Worker programs. While we received information on the new design and time frames for the WIA Gold Standard Evaluation, a finalized design plan is not yet available. According to officials, a finalized design is being prepared and will be available in June 2011. ETA Has Improved the Availability of Its Research but There Are Opportunities to Improve Its Search Page and Dissemination Methods: ETA Has Improved the Timeliness of Its Disseminated Research: ETA has recently improved the timeliness with which it disseminates its research reports. In our last review in January 2010, we found that 20 of the 34 reports that ETA disseminated in 2008 had been waiting 2 to 5 years to be publicly released.[Footnote 21] The 34 research reports published by ETA in 2008 took, on average, 804 days from the time the report was submitted to ETA until the time it was posted to ETA's research database. By contrast, from 2009 through the first quarter of 2010, the average time between submission and public release was 76 days, which represents a more than 90 percent improvement in dissemination time compared with 2008. Additionally, there were no research reports in 2009 that were delayed for more than 6 months. Further, the average time to dissemination improved significantly even when we excluded such outliers as the 20 research reports that were delayed for 2 years or more. Without these outliers, average time to dissemination for reports in 2008 was 100 days, indicating that time to dissemination in 2009 through the first quarter of 2010 still improved by 24 percent. ETA Has Improved Its Research Database but Lacks Plans for Assessing the Usability of Its Search Page: In 2010, ETA updated its online, Web-based search page in order to improve the usability of its research database--the primary tool for making ETA research available to policymakers and the general public. Officials told us that ETA's old Web-based search page was so error- prone and difficult to use that they opted to substitute it with one that had not yet completed internal testing. Our review of the old Web- based search page confirmed that it had serious limitations and did not consistently return the same results. For example, when we searched the database by title for a known ETA research report titled Registered Apprenticeship, we successfully retrieved that report once. One month later, when we entered the exact same search terms, we were unable to retrieve the report. (For a more complete description of our analysis of ETA's search capability, see appendix II.) In our review of the updated Web-based search page, we found that the updates make the research database more usable. Labor officials told us they have taken other steps, as well, in efforts to improve its Web- based search page. For example, they have developed a project plan that articulates the steps Labor will take to update ETA's Web-based search page. In addition, they have assigned a database administrator whose responsibilities include performing daily quality control spot checks in order to monitor performance and address technical problems. Although these changes have the potential to improve the usability of ETA's database, Labor has not developed a formal plan for assessing the overall effectiveness of its Web-based search page, including user satisfaction. Labor has made a number of changes to the way the page operates, but it has not provided users with tips on how to use the search functions, even though it is an industry standard to do so.[Footnote 22] Even skilled users who were familiar with the old Web- based search page may need guidance on the exact meaning of new terms and functions now available on the new page. For example, the old Web- based search page gave users the option of searching by "key word," which is no longer an option in the new page. Instead, "key word" searches have been replaced with a variety of other options, including the ability to search the full text or abstract of a research report. However, there is no guidance on the Web site on how to use these new search options. Industry best practices suggest that a Web site evaluation plan that incorporates data from routine reviews of Web site performance and that assesses user satisfaction can help agencies ensure the usability of their Web sites. ETA currently has no plans to do such assessments. ETA Uses Various Methods to Disseminate Research, but Experts Suggest Additional Methods: At present, ETA's research database is the primary method that ETA uses to make its research reports publicly available, according to officials. In order to call attention to new reports available in that database, ETA sends a Training and Employment Notice, also commonly known as a TEN, to an e-mail list of the more than 40,000 subscribers who have signed up to receive them. ETA's research process specifies that for each new research report that is approved for dissemination, ETA must draft a TEN and an abstract before it is posted to ETA's Web site. Beyond posting reports to its database, ETA also distributes hard copies of some of its research reports. In addition to electronic distribution, ETA also organizes various presentations to disseminate its research findings. These presentations, however, are done on an ad hoc basis. As mentioned in our prior report, ETA hosted a research conference in 2009 to present some of its research findings, renewing a practice that had been discontinued in 2003. As ETA looks to the future, officials tell us they will plan and organize similar research conferences as resources permit. In addition to these research conferences, ETA's regional offices occasionally hold smaller, regional conferences as well. Beyond these formal conferences, ETA also hosts an internal briefing series at Labor headquarters where research contractors present their findings to various officials. For each of these briefings, ETA has a list of stakeholders that it invites, including various Labor officials, outside agency officials, congressional staff, and other outside stakeholders. Experts who participated in our virtual panel provided their views on the effectiveness of different methods for disseminating research reports, and several of those rated more highly are methods currently employed by ETA. (See figure 8.) Figure 8: Experts' Opinion on the Effectiveness of Various Research Dissemination Methods: [Refer to PDF for image: stacked horizontal bar graph] E-mail notifications announcing newly released research: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 3; Moderately effective: 6; Very effective: 19; Extremely effective: 11. Searchable database of papers on ETA‘s Web site: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 3; Moderately effective: 10; Very effective: 10; Extremely effective: 16. Distributing a one-page summary of the research findings from each report: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 4; Moderately effective: 9; Very effective: 17; Extremely effective: 9. Briefings at ETA for external audiences (including stakeholders and policymakers): Number of respondents answering: No response or left blank: 1; Not at all effective: 0; Somewhat effective: 7; Moderately effective: 9; Very effective: 14; Extremely effective: 8. Publicizing research findings through press articles, industry publications, etc.: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 7; Moderately effective: 13; Very effective: 14; Extremely effective: 5. Presenting at outside conferences: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 6; Moderately effective: 14; Very effective: 15; Extremely effective: 4. National ETA-sponsored conferences: Number of respondents answering: No response or left blank: 2; Not at all effective: 0; Somewhat effective: 5; Moderately effective: 16; Very effective: 13; Extremely effective: 3. Regional ETA-sponsored conferences: Number of respondents answering: No response or left blank: 1; Not at all effective: 0; Somewhat effective: 4; Moderately effective: 19; Very effective: 13; Extremely effective: 2. Publishing articles in journals: Number of respondents answering: No response or left blank: 0; Not at all effective: 0; Somewhat effective: 14; Moderately effective: 11; Very effective: 8; Extremely effective: 6. Contracting with relevant organizations or associations to disseminate research: Number of respondents answering: No response or left blank: 1; Not at all effective: 0; Somewhat effective: 7; Moderately effective: 17; Very effective: 10; Extremely effective: 4. Webinars: Number of respondents answering: No response or left blank: 0; Not at all effective: 1; Somewhat effective: 9; Moderately effective: 17; Very effective: 9; Extremely effective: 3. Distributing a compendium of ETA‘s research: Number of respondents answering: No response or left blank: 0; Not at all effective: 2; Somewhat effective: 7; Moderately effective: 20; Very effective: 5; Extremely effective: 5. Videos or podcasts: Number of respondents answering: No response or left blank: 2; Not at all effective: 4; Somewhat effective: 13; Moderately effective: 12; Very effective: 8; Extremely effective: 0. Social media (such as Twitter): Number of respondents answering: No response or left blank: 3; Not at all effective: 11; Somewhat effective: 12; Moderately effective: 10; Very effective: 3; Extremely effective: 0. Source: GAO‘s survey of ETA‘s research priorities and dissemination methods. [End of figure] Most of the experts (30 of the 39 respondents) in our panel reported that using e-mail notifications, a searchable database of ETA papers, and briefings at ETA for external audiences (including stakeholders and policymakers) would be very effective or extremely effective approaches for disseminating research. In addition, a majority of the experts (26 of the 39 respondents) in our panel reported that publishing one-page summaries of research findings, not currently done by ETA, would be very or extremely effective. Conclusions: ETA plays an important role in developing workforce policies and helping to identify the most effective and efficient ways to train and employ workers for jobs in the twenty-first century. With the current economic crisis and high unemployment rates, ETA's role has become even more critical. The agency has made some improvements in its research program, even since our last review a year ago. But officials can do more to ensure that the progress continues in the years to come. ETA has taken a major step forward in establishing a formal research process--one that documents most actions that must be taken in the life cycle of a research or evaluation project. But, it is missing some key elements that could help ensure the continuation of current practices. While ETA is currently using outside advisory bodies to help it establish its research agenda, the formal process does not include the agenda-setting phase. Officials tell us they have plans to incorporate this phase in the future, and we urge them to do so. Without a formalized agenda-setting phase, ETA may miss opportunities to ensure that its research agenda addresses the most critical employment and training issues and that outside stakeholders are routinely involved. Moreover, ETA's process has not formalized the now ad hoc advisory role of the Chief Evaluation Officer. Absent the routine involvement of the Chief Evaluation Officer at key steps in the process, ETA may find it difficult to ensure that research proposals are asking the right questions, are methodologically sound, and that they can quickly pass critical OMB reviews. ETA's research findings are now available to the public on its Web site in far less time than it took in 2008. Despite this clear improvement, ETA has not taken the necessary steps to ensure that research products remain readily available to the public. The decision regarding what and when to make research publicly available is left in the hands of too few, and the process lacks needed safeguards to ensure transparency and accountability. Absent safeguards, key research decisions may again be made in ways that harm the credibility of the program and prevent important research findings from being used to inform policy and practice. ETA's Web-based search page is the primary means ETA uses to make the research studies it funds readily available to the public. And, while ETA has improved the functionality of its Web site, no effort has been made to ensure that the problems that plagued the system in the past do not recur. Absent such efforts, ETA will have little assurance that its research findings are actually available to users. Recommendations for Executive Action: To improve ETA's research program, we recommend that the Secretary of Labor require ETA to take the following three actions: * Formally incorporate into its research process the routine involvement of the Chief Evaluation Officer at key milestones, including at the development of ETA's annual research agenda and spending priorities, as well as at the early stages of developing specific research projects. * Develop a mechanism to enhance the transparency and accountability of ETA's research program. For example, such a mechanism might include involving advisory bodies or other entities outside ETA, in efforts to develop ETA's research policies and processes. * Develop a formal plan for ensuring that ETA's research products are easily accessible to stakeholders and to the general public through its searchable database. Such a plan could involve requiring Labor to assess the overall effectiveness of its Web-based search page, including user satisfaction with search features. Agency Comments and Our Evaluation: We provided a draft of this report to the Department of Labor for review and comment. Labor provided written comments, which are reproduced in appendix VII. In addition, ETA provided technical comments, which we incorporated where appropriate. In its response, Labor generally agreed with our findings and all of our recommendations, noting its ongoing efforts in support of the recommendations. * Regarding our recommendation to formally incorporate into its research process the routine involvement of the Chief Evaluation Officer at key research milestones, Labor noted that it is currently taking steps to do so. Officials reported that they have worked closely with this office in various aspects of its research, including discussing research, demonstration projects, and evaluations in the early stages of development and plans to continue this collaboration in the future. However, ETA's comments did not discuss plans to update its documentation on the formal research process. We found in our review that involving the Chief Evaluation Officer was not an official component of ETA's documented research process, and it occurred on an ad hoc basis. As ETA moves forward, we urge the agency to modify its current research process and document the involvement of the Chief Evaluation Officer at critical research milestones. * Regarding our recommendation for ETA to develop a mechanism to enhance the transparency and accountability of its research program, officials cited several steps they are taking to improve the program, including involving outside experts in the development of their 5-year research plan and establishing advisory and peer review groups to review major evaluations. While officials note they plan to engage outside experts in broader research policies and processes, we encourage ETA to formalize this involvement. Moreover, we encourage ETA to continue to move forward in its efforts to further clarify components of its research process that are not well defined, including, for example, the criteria to be used when deciding when a peer review should be performed. * Regarding our recommendation to develop a formal plan to ensure that disseminated research is easily accessible to stakeholders and the general public, officials cited specific steps the agency has taken to improve its Web-based research database. While these actions are a step in the right direction, we believe that it is still important for Labor to develop a formal and comprehensive plan to ensure that disseminated research continues to be accessible to the public. Furthermore, Labor expressed concerns about how we characterized the agency's budget for pilots, demonstrations, and research. Recognizing these concerns, we made changes to the report to better capture the amount of funding ETA has available for research. As agreed with your offices, unless you publicly announce the contents of this report earlier, we plan no further distribution until 30 days from the report date. At that time, we will send copies to the appropriate congressional committees, the Secretary of Labor, and other interested parties. The report will also be available at no charge on GAO's Web site at [hyperlink, http://www.gao.gov]. If you or your staffs have any questions about this report, please contact me at (202) 512-7215 or scottg@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made key contributions to this report are listed in appendix VIII. Signed by: George A. Scott: Director, Education, Workforce and Income Security Issues: [End of section] Appendix I: Status of Prior GAO Recommendations to the Department of Labor, as of January 2011: GAO's recommendations: The Secretary of Labor should Department of Labor's response: [Empty]; Status: [Empty]. GAO's recommendations: The Secretary of Labor should take steps to clarify the Employment and Training Administration's (ETA) revised organizational structure and ensure that the Office of Policy Development and Research (OPDR) reports directly to ETA's Assistant Secretary; Department of Labor's response: The Department of Labor (Labor) does not agree with this recommendation as written. According to Labor officials, the Administrator of OPDR currently reports to the Deputy Assistant Secretary, not directly to ETA's Assistant Secretary. However, Labor officials acknowledge that important functions such as research and evaluation should not have too many intermediary reporting layers. To facilitate communication, officials further noted that the OPDR Administrator, the Deputy Assistant Secretary, and the Chief Evaluation Officer meet on a monthly basis with the Assistant Secretary to discuss evaluations; Status: Labor has taken no action. GAO's recommendations: The Secretary of Labor should provide sufficient authority to ETA's research and evaluation center to plan, conduct, and disseminate research; Department of Labor's response: Labor agrees with this recommendation, but authority to make key decisions still resides with the Office of the Assistant Secretary for ETA. OPDR currently provides recommendations to this office regarding plans for conducting and disseminating research. In an effort to improve evaluations departmentwide, the Secretary of Labor recently established the Chief Evaluation Office to monitor evaluation efforts across the department. OPDR has begun to work informally with the Chief Evaluation Officer and the Chief Economist to design and implement research and evaluation projects; Status: Labor's actions do not completely satisfy recommendation. GAO's recommendations: The Secretary of Labor should direct ETA's research and evaluation center to establish more specific processes, including time frames for dissemination of research, to promote transparency and accountability; Department of Labor's response: Labor agrees with this recommendation. ETA reports that it has taken some steps to establish more specific processes regarding dissemination of research, citing changes in performance standards for project officers. However, our recommendation would make broader changes to their research process and no such changes are reflected in the documents the agency provided; Status: Labor's actions do not completely satisfy recommendation. GAO's recommendations: The Secretary of Labor should create an information system to track research projects at all phases to ensure timely completion and dissemination; Department of Labor's response: Labor agrees with this recommendation. Officials report that they have begun working on a centralized, electronic tracking system for its research projects. However, the work is still under way and no time frames have been provided for its completion. Currently, OPDR uses an Excel document to keep inventory of all research, demonstration, and evaluation projects; Status: Labor's actions do not completely satisfy recommendation. GAO's recommendations: The Secretary of Labor should instruct ETA's research and evaluation center to develop processes to routinely involve outside experts in setting its research agenda and to the extent required, do so consistent with the Federal Advisory Committee Act; Department of Labor's response: Labor agrees with this recommendation. OPDR has taken steps to engage outside experts in setting its 5-year research plan for 2011 and collaborate with the research and evaluation centers of other federal agencies, such as the Departments of Education and Health and Human Services. OPDR also plans to convene an expert panel, solicit public comments, and incorporate feedback from its 2009 Reemployment Research Conference and its 2010 ETA Reemployment Summit. However, despite these current efforts, OPDR has not formally incorporated them in its standard research process; Status: Labor's actions do not completely satisfy recommendation. Source: GAO analysis of GAO-10-243. [End of table] [End of section] Appendix II: Scope and Methodology: We were asked to review the Employment and Training Administration's (ETA) research program to better understand its approach to conducting and disseminating research. Specifically, we answered the following research questions: (1) To what extent do ETA's research priorities reflect key national employment and training issues and how useful were the studies funded under them? (2) What steps has ETA taken to improve its research program? (3) How has ETA improved, if at all, the availability of its research since our last review in January 2010 and what other steps could ETA take to further ensure its research findings are readily available? To answer our research questions, we convened a virtual panel using a modified Delphi technique to obtain selected employment and training experts' opinions on ETA's research priorities and dissemination methods. We also visited two workforce agencies in Pennsylvania and Virginia that are implementing two of ETA's ongoing research studies to learn about implementation issues and how research is being conducted. In addition, we reviewed 58 ETA-funded research and evaluation reports disseminated between January 2008 and March 2010 and assessed the methodological soundness of completed studies that cost $1 million or more. We also reviewed ETA's ongoing studies that cost $2 million or more. To determine the availability of ETA's research, we measured the time between when the final version of a research report was submitted to ETA's Office of Policy Development and Research (ODPR) and when it was posted on ETA's Web site. We also conducted a series of systematic searches to test the reliability of ETA's research database. Furthermore, we interviewed Department of Labor (Labor) and ETA officials to better understand ETA's research capacity, processes, and the use of research findings to inform policy and practice. Lastly, we reviewed relevant agency documents and policies, as well as relevant federal laws. Web-Based Expert Panel: We convened a nongeneralizable Web-based virtual panel of 41 employment and training experts to obtain their opinions on ETA's research priorities and dissemination methods. We employed a modified version of the Delphi method to organize and gather these experts' opinions.[Footnote 23] To encourage participation by our experts, we promised that responses would not be individually identifiable and that results would generally be provided in summary form. To select the panel, we asked several employment and training experts, on the basis of their experience and expertise, to identify other experts who were knowledgeable of ETA and the research it conducts and disseminates. After receiving nominations from experts, we reviewed the list to ensure that it reflected a range of perspectives and backgrounds, including academics, researchers, and consultants. Our Delphi process entailed two survey phases. (See appendix V for a copy of our phase I and phase II questionnaires.) In phase I, which ran from June 22, 2010, to August 9, 2010, we asked the panel to respond to five open-ended questions about ETA's research priorities and dissemination methods. We developed these questions based on our study objectives and pretested them with four experts by phone to ensure the questionnaire was clear, unbiased, and did not place an undue burden on respondents. All relevant changes were made before we deployed the first Web-based questionnaire to experts. After the experts completed the open-ended questions in the first questionnaire, we performed a content analysis of the responses in order to identify the most important issues raised by our experts. Two members of our team categorized experts' responses to each of the questions. Any disagreements were discussed until consensus was reached. Thirty-six of the 41 panelists selected completed phase I of the survey (about an 88 percent response rate). Those that did not complete phase I were allowed to participate in phase II. (For a list of experts who participated in phase I and phase II, see appendix VI.) The experts' responses to phase I were used to create the questions for phase II. In phase II, we gathered more specific information on ETA's research and dissemination practices. Phase II, which ran from October 29, 2010, to December 14, 2010, consisted of 16 follow-up questions where panelists were asked to either rank or rate the responses from phase I. We pretested the questionnaire for the second phase with three experts to ensure the clarity of the instrument. We conducted two of our expert pretests in-person and one by phone. Thirty-nine of the 41 experts completed phase II (about a 95 percent response rate). Site Visits to Workforce Agencies Implementing ETA-Funded Research Studies: To further enhance our understanding of how ETA conducts its research, we visited two workforce agencies that are implementing ETA's ongoing research studies. First, we visited the Lancaster County Workforce Investment Board in Lancaster, Pa., which received funding from ETA to implement the Young Parents Demonstration project.[Footnote 24] This project provides educational and occupational skills training to promote employment and economic self-sufficiency for mothers, fathers, and expectant mothers ages 16 to 24. Second, we visited the Northern Virginia Workforce Investment Board in Falls Church, Va., which received funding from ETA to implement the second round of the Project Growing America Through Entrepreneurship, also referred to as Project GATE II.[Footnote 25] This grant helps dislocated workers aged 50 and over obtain information, classroom training, one-to-one technical assistance, counseling, and financial assistance to establish new businesses in order to help them start and sustain successful self- employment. We selected these workforce agencies because they were identified by ETA as having active research projects in the implementation stage. These sites also required minimum travel expenditure. During our site visits, we toured each workforce agencies' facilities and used a semistructured interview protocol to interview the project director and staff about their role and responsibilities, the extent to which they communicate with ETA, and whether or not they face challenges with regards to implementation. At the Lancaster County site, we participated in an informal on-site lunch forum where local community programs that the agency partners with talked with us about their collaboration with the program. At the Northern Virginia GATE II site, we observed a focus group operated by the program to facilitate information-sharing among participants. After our site visits, we conducted phone interviews with the contractors that received funding from ETA to evaluate the outcomes of two research projects. Specifically, we interviewed the Urban Institute, which evaluates the Young Parents Demonstration project, and IMPAQ International, which evaluates Project GATE II. Both projects include an experimental component with control and comparison groups to determine the effects of program interventions on participants. During our interviews we used a semistructured questionnaire and asked questions to better understand their roles and responsibilities for the project, the extent to which they communicate with ETA, and whether or not they experience methodological and implementation challenges. Analysis of Methodological Characteristics of ETA: We reviewed the 58 research and evaluation reports that ETA disseminated between January 2008 and March 2010 and assessed the methodological soundness of 11 completed studies that cost $1 million or more. In addition, we reviewed 10 ongoing studies costing $2 million or more to determine if research practices or the soundness of research designs had changed over time. We categorized the 58 studies disseminated between January 2008 and March 2010 by study type, cost, and research area. For the larger studies costing $1 million or more, we analyzed key characteristics including design features, scope, generalizability, and the appropriateness of analytical approaches and statistical procedures. These studies were analyzed independently by two analysts and the agreement between their ratings was 100 percent. (For results of this analysis, see appendix VI.) Analysis of the Timeliness and Effectiveness of ETA's Dissemination Activities: To evaluate the availability of ETA's research, we measured the time between when the final version of a research report was submitted to ODPR and when it was posted on ETA's Web site. Specifically, we measured the dissemination time frames for reports posted in 2008 and compared that with the dissemination time frames for reports issued between January 2009 through March 2010. In addition, we conducted a series of systematic searches to test the reliability of ETA's Web- based research database. To perform our searches, we selected a random sample of 30 reports from the 312 reports available on ETA's research database at the time of our review. Specifically, we tested a variety of search functions available at the time of our review to determine the extent to which research reports could be easily retrieved on ETA's research database. These functions included searches by title, keywords, author, and/or dates. We classified a report as retrievable if it appeared anywhere in our search results. We conducted our initial searches between June 30, 2010, and July 6, 2010. A second round of searches was conducted between August 6, 2010, and August 10, 2010. Further, we interviewed Labor and ETA officials to learn more about the search capabilities of ETA's research database and the processes used to address errors and implement changes. Finally, we interviewed officials to gather information about ETA's dissemination methods, including its current techniques and future plans for disseminating research reports. Interviews with Labor and ETA Officials: To better understand the agency's research capacity, we interviewed ETA officials and reviewed relevant agency and budget documentation. Similarly, to obtain information on ETA's research process and how research findings are used to inform employment and training policy and practice, we interviewed officials and reviewed agency documentation, including relevant policies and procedures that guide ETA's research. We also reviewed relevant federal laws. We conducted this performance audit from March 2009 through March 2011 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix III: The Panel's Ratings of Key Employment and Training Issues, Populations, and Programs That ETA Should Address in Its Future Research: In our Delphi phase II Web-based questionnaire, we asked the panel of experts to rate and rank the key employment and training issues, populations, and programs that ETA should address in its future research. These issues were identified by the panel during phase I. For our analysis, we calculated basic descriptive statistics on these issues, which are presented in tables 5 through 7. Table 5: Descriptive Statistics on Key Employment and Training Issues ETA Should Address in Its Future Research: The responses in this table are based on the following questions: Question 1: Taking into account ETA's limited resources, how important, if at all, is it for ETA to address the following in future research? (Q1): Question 2: Among the areas you checked as at least moderately important in question 1, which would you rank as the top 3 areas that ETA should address in future research. (Please rank only 3 areas, with 1 as your top priority area, 2 as your second highest area, etc.) Long-term outcomes of employment and training programs: Responses to question 2: Ranked 1: 9; Responses to question 2: Ranked 2: 3; Responses to question 2: Ranked 3: 3; Responses to question 2: Ranked 1, 2, or 3: 15; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 4; Responses to question 1: Moderately important: 4; Responses to question 1: Very Important: 14; Responses to question 1: Extremely important: 17. Short-term outcomes of employment and training programs: Responses to question 2: Ranked 1: 2; Responses to question 2: Ranked 2: 0; Responses to question 2: Ranked 3: 2; Responses to question 2: Ranked 1, 2, or 3: 4; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 5; Responses to question 1: Moderately important: 12; Responses to question 1: Very Important: 13; Responses to question 1: Extremely important: 9. Value of various credentials: Responses to question 2: Ranked 1: 0; Responses to question 2: Ranked 2: 1; Responses to question 2: Ranked 3: 2; Responses to question 2: Ranked 1, 2, or 3: 3; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 7; Responses to question 1: Moderately important: 12; Responses to question 1: Very Important: 15; Responses to question 1: Extremely important: 4. Impact of long-term and/or short-term training: Responses to question 2: Ranked 1: 5; Responses to question 2: Ranked 2: 9; Responses to question 2: Ranked 3: 4; Responses to question 2: Ranked 1, 2, or 3: 18; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 0; Responses to question 1: Moderately important: 5; Responses to question 1: Very Important: 18; Responses to question 1: Extremely important: 16. Job creation strategies, such as providing wage subsidies, tax credits, or public service employment: Responses to question 2: Ranked 1: 2; Responses to question 2: Ranked 2: 9; Responses to question 2: Ranked 3: 7; Responses to question 2: Ranked 1, 2, or 3: 18; Responses to question 1: Not at all important: 1; Responses to question 1: Somewhat important: 2; Responses to question 1: Moderately important: 5; Responses to question 1: Very Important: 18; Responses to question 1: Extremely important: 13. Employment and training approaches that work and for whom: Responses to question 2: Ranked 1: 9; Responses to question 2: Ranked 2: 11; Responses to question 2: Ranked 3: 8; Responses to question 2: Ranked 1, 2, or 3: 28; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 1; Responses to question 1: Moderately important: 1; Responses to question 1: Very Important: 18; Responses to question 1: Extremely important: 19. One-stop center management and operations: Responses to question 2: Ranked 1: 0; Responses to question 2: Ranked 2: 1; Responses to question 2: Ranked 3: 2; Responses to question 2: Ranked 1, 2, or 3: 3; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 10; Responses to question 1: Moderately important: 14; Responses to question 1: Very Important: 12; Responses to question 1: Extremely important: 3. Effective performance measurement systems: Responses to question 2: Ranked 1: 2; Responses to question 2: Ranked 2: 3; Responses to question 2: Ranked 3: 4; Responses to question 2: Ranked 1, 2, or 3: 9; Responses to question 1: Not at all important: 0; Responses to question 1: Somewhat important: 4; Responses to question 1: Moderately important: 10; Responses to question 1: Very Important: 16; Responses to question 1: Extremely important: 9. Linkages between the public workforce system and economic development entities: Responses to question 2: Ranked 1: 4; Responses to question 2: Ranked 2: 1; Responses to question 2: Ranked 3: 1; Responses to question 2: Ranked 1, 2, or 3: 6; Responses to question 1: Not at all important: 2; Responses to question 1: Somewhat important: 11; Responses to question 1: Moderately important: 11; Responses to question 1: Very Important: 8; Responses to question 1: Extremely important: 7. Needs of different labor markets or industries: Responses to question 2: Ranked 1: 1; Responses to question 2: Ranked 2: 1; Responses to question 2: Ranked 3: 2; Responses to question 2: Ranked 1, 2, or 3: 4; Responses to question 1: Not at all important: 2; Responses to question 1: Somewhat important: 6; Responses to question 1: Moderately important: 15; Responses to question 1: Very Important: 14; Responses to question 1: Extremely important: 2. Employment and training strategies for various economic conditions: Responses to question 2: Ranked 1: 0; Responses to question 2: Ranked 2: 0; Responses to question 2: Ranked 3: 1; Responses to question 2: Ranked 1, 2, or 3: 1; Responses to question 1: Not at all important: 1; Responses to question 1: Somewhat important: 4; Responses to question 1: Moderately important: 13; Responses to question 1: Very Important: 15; Responses to question 1: Extremely important: 6. Issues related to unemployment insurance: Responses to question 2: Ranked 1: 5; Responses to question 2: Ranked 2: 0; Responses to question 2: Ranked 3: 3; Responses to question 2: Ranked 1, 2, or 3: 8; Responses to question 1: Not at all important: 2; Responses to question 1: Somewhat important: 5; Responses to question 1: Moderately important: 10; Responses to question 1: Very Important: 12; Responses to question 1: Extremely important: 10. Source: GAO's survey of ETA's research priorities and dissemination methods. [End of table] Table 6: Descriptive Statistics on Key Populations ETA Should Address in Its Future Research: The responses in this table are based on the following questions: Question 3: How important, if at all, would it be for ETA to fund research that focuses on what works for the following populations given its resource constraints? Question 4: Among the populations you checked as at least moderately important in question 3, which would you rank as the top 5 populations on which ETA should fund research in the future? (Please rank only 5 populations, with 1 as your top population, 2 as your second highest population, etc.) Short-term unemployed: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 1; Responses to question 4: Ranked 4: 0; Responses to question 4: Ranked 5: 0; Responses to question 4: Ranked 1,2, 3, 4, or 5: 3; Responses to question 3: Not at all important: 3; Responses to question 3: Somewhat important: 16; Responses to question 3: Moderately important: 8; Responses to question 3: Very Important: 7; Responses to question 3: Extremely important: 2. Long-term unemployed: Responses to question 4: Ranked 1: 15; Responses to question 4: Ranked 2: 10; Responses to question 4: Ranked 3: 6; Responses to question 4: Ranked 4: 1; Responses to question 4: Ranked 5: 1; Responses to question 4: Ranked 1,2, 3, 4, or 5: 33; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 1; Responses to question 3: Moderately important: 0; Responses to question 3: Very Important: 19; Responses to question 3: Extremely important: 19. Dislocated workers: Responses to question 4: Ranked 1: 5; Responses to question 4: Ranked 2: 4; Responses to question 4: Ranked 3: 4; Responses to question 4: Ranked 4: 2; Responses to question 4: Ranked 5: 3; Responses to question 4: Ranked 1,2, 3, 4, or 5: 18; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 4; Responses to question 3: Moderately important: 7; Responses to question 3: Very Important: 19; Responses to question 3: Extremely important: 9. Older workers: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 2; Responses to question 4: Ranked 4: 1; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 7; Responses to question 3: Not at all important: 2; Responses to question 3: Somewhat important: 11; Responses to question 3: Moderately important: 9; Responses to question 3: Very Important: 13; Responses to question 3: Extremely important: 3. Veterans: Responses to question 4: Ranked 1: 1; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 2; Responses to question 4: Ranked 4: 0; Responses to question 4: Ranked 5: 0; Responses to question 4: Ranked 1,2, 3, 4, or 5: 5; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 9; Responses to question 3: Moderately important: 13; Responses to question 3: Very Important: 11; Responses to question 3: Extremely important: 4. Economically disadvantaged workers: Responses to question 4: Ranked 1: 11; Responses to question 4: Ranked 2: 8; Responses to question 4: Ranked 3: 8; Responses to question 4: Ranked 4: 0; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 29; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 2; Responses to question 3: Moderately important: 2; Responses to question 3: Very Important: 22; Responses to question 3: Extremely important: 13. Racial and ethnic minorities: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 0; Responses to question 4: Ranked 4: 4; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 8; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 8; Responses to question 3: Moderately important: 13; Responses to question 3: Very Important: 13; Responses to question 3: Extremely important: 3. Immigrants: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 0; Responses to question 4: Ranked 4: 0; Responses to question 4: Ranked 5: 1; Responses to question 4: Ranked 1,2, 3, 4, or 5: 3; Responses to question 3: Not at all important: 2; Responses to question 3: Somewhat important: 10; Responses to question 3: Moderately important: 12; Responses to question 3: Very Important: 11; Responses to question 3: Extremely important: 2. Adults with low basic skills: Responses to question 4: Ranked 1: 3; Responses to question 4: Ranked 2: 3; Responses to question 4: Ranked 3: 6; Responses to question 4: Ranked 4: 3; Responses to question 4: Ranked 5: 7; Responses to question 4: Ranked 1,2, 3, 4, or 5: 22; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 2; Responses to question 3: Moderately important: 6; Responses to question 3: Very Important: 19; Responses to question 3: Extremely important: 10. Workers with physical and mental disabilities: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 0; Responses to question 4: Ranked 3: 0; Responses to question 4: Ranked 4: 3; Responses to question 4: Ranked 5: 1; Responses to question 4: Ranked 1,2, 3, 4, or 5: 4; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 8; Responses to question 3: Moderately important: 15; Responses to question 3: Very Important: 13; Responses to question 3: Extremely important: 1. In-school youth: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 0; Responses to question 4: Ranked 3: 3; Responses to question 4: Ranked 4: 0; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 5; Responses to question 3: Not at all important: 2; Responses to question 3: Somewhat important: 14; Responses to question 3: Moderately important: 12; Responses to question 3: Very Important: 9; Responses to question 3: Extremely important: 1. Out-of-school youth: Responses to question 4: Ranked 1: 4; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 4; Responses to question 4: Ranked 4: 8; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 20; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 3; Responses to question 3: Moderately important: 10; Responses to question 3: Very Important: 17; Responses to question 3: Extremely important: 9. Ex-offenders: Responses to question 4: Ranked 1: 0; Responses to question 4: Ranked 2: 2; Responses to question 4: Ranked 3: 3; Responses to question 4: Ranked 4: 3; Responses to question 4: Ranked 5: 2; Responses to question 4: Ranked 1,2, 3, 4, or 5: 10; Responses to question 3: Not at all important: 0; Responses to question 3: Somewhat important: 5; Responses to question 3: Moderately important: 10; Responses to question 3: Very Important: 17; Responses to question 3: Extremely important: 6. Source: GAO's survey of ETA's research priorities and dissemination methods. [End of table] Table 7: Descriptive Statistics on Key Employment and Training Programs ETA Should Address in Its Future Research: The responses in this table are based on the following questions: Question 6: How important, if at all, is it for ETA to evaluate the following key employment or training programs (excluding UI)? Question 7: Given a limited amount of resources, which three of the key employment or training programs that you checked as at least moderately important in question 6 should ETA address in future research? (Please rank only 3 employment or training programs, with 1 as your top program, 2 as your second highest program, etc.) WIA Adult: Responses to question 7: Ranked 1: 12; Responses to question 7: Ranked 2: 9; Responses to question 7: Ranked 3: 7; Responses to question 7: Ranked 1, 2, or 3: 28; Responses to question 6: Not at all important: 0; Responses to question 6: Somewhat important: 2; Responses to question 6: Moderately important: 11; Responses to question 6: Very Important: 11; Responses to question 6: Extremely important: 13. WIA Dislocated Worker: Responses to question 7: Ranked 1: 8; Responses to question 7: Ranked 2: 11; Responses to question 7: Ranked 3: 3; Responses to question 7: Ranked 1, 2, or 3: 22; Responses to question 6: Not at all important: 0; Responses to question 6: Somewhat important: 3; Responses to question 6: Moderately important: 8; Responses to question 6: Very Important: 13; Responses to question 6: Extremely important: 13. WIA Youth: Responses to question 7: Ranked 1: 7; Responses to question 7: Ranked 2: 6; Responses to question 7: Ranked 3: 4; Responses to question 7: Ranked 1, 2, or 3: 17; Responses to question 6: Not at all important: 0; Responses to question 6: Somewhat important: 4; Responses to question 6: Moderately important: 11; Responses to question 6: Very Important: 15; Responses to question 6: Extremely important: 8. Wagner-Peyser employment Service: Responses to question 7: Ranked 1: 3; Responses to question 7: Ranked 2: 3; Responses to question 7: Ranked 3: 6; Responses to question 7: Ranked 1, 2, or 3: 12; Responses to question 6: Not at all important: 0; Responses to question 6: Somewhat important: 7; Responses to question 6: Moderately important: 13; Responses to question 6: Very Important: 9; Responses to question 6: Extremely important: 7. Job Corps: Responses to question 7: Ranked 1: 1; Responses to question 7: Ranked 2: 3; Responses to question 7: Ranked 3: 1; Responses to question 7: Ranked 1, 2, or 3: 5; Responses to question 6: Not at all important: 1; Responses to question 6: Somewhat important: 11; Responses to question 6: Moderately important: 12; Responses to question 6: Very Important: 10; Responses to question 6: Extremely important: 2. Trade Adjustment Assistance: Responses to question 7: Ranked 1: 2; Responses to question 7: Ranked 2: 2; Responses to question 7: Ranked 3: 3; Responses to question 7: Ranked 1, 2, or 3: 7; Responses to question 6: Not at all important: 2; Responses to question 6: Somewhat important: 5; Responses to question 6: Moderately important: 15; Responses to question 6: Very Important: 11; Responses to question 6: Extremely important: 5. Veterans' Employment and Training: Responses to question 7: Ranked 1: 1; Responses to question 7: Ranked 2: 2; Responses to question 7: Ranked 3: 4; Responses to question 7: Ranked 1, 2, or 3: 7; Responses to question 6: Not at all important: 1; Responses to question 6: Somewhat important: 9; Responses to question 6: Moderately important: 16; Responses to question 6: Very Important: 6; Responses to question 6: Extremely important: 6. Apprenticeship program: Responses to question 7: Ranked 1: 3; Responses to question 7: Ranked 2: 5; Responses to question 7: Ranked 3: 4; Responses to question 7: Ranked 1, 2, or 3: 3; Responses to question 6: Not at all important: 7; Responses to question 6: Somewhat important: 14; Responses to question 6: Moderately important: 1; Responses to question 6: Very Important: 7; Responses to question 6: Extremely important: 12. Source: GAO's survey of ETA's research priorities and dissemination methods. [End of table] [End of section] Appendix IV: Characteristics of Research Studies Disseminated between January 2009 and March 2010 That Cost $1 Million or More: Report name: 1. Implementation Analysis of the High Growth Job Training Initiative (HGJI) Programs; ETA report number: 2008-10; ETA research area: Integration of the workforce and regional economic development; Cost: $1,500,000; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Implementation review; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 2. Workforce Investment Act Non-Experimental Net Impact Evaluation: Final Report; ETA report number: 2009-10; ETA research area: Using state-level administrative data to measure progress and outcomes; Cost: $1,000,000; Study type: Evaluation; Study design: Quasi-experimental; Methods and techniques: * Other (comparison v. treatment group using propensity score matching); Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 3. Evaluation of the Prisoner Re-Entry Initiative - Final Report; ETA report number: 2009-3; ETA research area: Increasing the labor market participation of underutilized populations; Cost: $1,204,078; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Post-intervention only data collection; * Other (some characteristics collected before the intervention, but these were not used to make comparisons); Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 4. Implementing the National Fund for Workforce Solutions: The Baseline Evaluation Report; ETA report number: 2009-21; ETA research area: Integration of the workforce and regional economic development; Cost: $1,000,000; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Implementation review; * Secondary analysis; Consistency of analytical approach with research question: No basis to judge; Generalizability of findings: No response. Report name: 5. The Power of Partnership: American Regions Collaborating for Economic Competitiveness (Generation I WIRED Interim Eval); ETA report number: 2009-18; ETA research area: Integration of the workforce and regional economic development; Cost: $3,433,478[A]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Implementation review; * Survey; * Secondary analysis; * Other (social network analysis); Consistency of analytical approach with research question: No; Generalizability of findings: No. Report name: 6. Early Implementation of Generation I of the Workforce Innovation in Regional Development (WIRED) Initiative, 2007 Interim Evaluation Report; ETA report number: 2008-03; ETA research area: Integration of the workforce and regional economic development; Cost: [Empty]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Implementation review; * Survey; * Secondary analysis; * Other (social network analysis); Consistency of analytical approach with research question: No; Generalizability of findings: No. Report name: 7. Nurturing America's Growth in the Global Marketplace: An Interim Report on the Evaluation of Generations II and III of WIRED; ETA report number: 2009-19; ETA research area: Integration of the workforce and regional economic development; Cost: $3,345,036; Study type: Evaluation; Study design: Quasi-experimental; Methods and techniques: * Survey; * Secondary analysis; * Other (comparison v. treatment group analysis ); Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 8. Recent Changes in the Characteristics of Unemployed Workers; ETA report number: 2009-13; ETA research area: Unemployment insurance; Cost: $6,507,262[B]; Study type: Research; Study design: Descriptive; Methods and techniques: * Prestest/post-test intervention data collection; Consistency of analytical approach with research question: Yes; Generalizability of findings: Yes. Report name: 9. Trends in the Structure of the Labor Market and Unemployment; ETA report number: 2009-09; ETA research area: Unemployment insurance; Cost: [Empty]; Study type: Research; Study design: Descriptive; Methods and techniques: * Secondary analysis; Consistency of analytical approach with research question: Yes; Generalizability of findings: Yes. Report name: 10. Reemployment and Eligibility Assessment (REA) Study FY 2005 Initiative; ETA report number: 2008-02; ETA research area: Unemployment insurance; Cost: [Empty]; Study type: Research; Study design: * Experimental (randomized control trials; * Descriptive; Methods and techniques: * Survey; * Secondary analysis; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 11. Growing America Through Entrepreneurship: Findings from the Evaluation of Project GATE; ETA report number: 2008-08; ETA research area: Postsecondary education and job training; Cost: $11,400,000; Study type: Evaluation; Study design: Experimental (randomized control trials); Methods and techniques: * Survey; * Secondary analysis; * Prestest/post-test intervention data collection; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 12. Evaluation of Youth Build Offender Grants; ETA report number: 2009-11; ETA research area: Increasing the labor market participation of underutilized populations; Cost: $1,151,449; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; * Other (data collected on characteristics upon entry and outcome characteristics collected after completion); Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 13. Initial Implementation of the Trade Act; ETA report number: 2009-14; ETA research area: Methods of expanding U.S. workforce skills; Cost: $10,453,957[C]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; * Implementation review; * Secondary analysis; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 14. Assessment, Case Management, and Post-Training Assistance for TAA Participants; ETA report number: 2009-15; ETA research area: Methods of expanding U.S. workforce skills; Cost: [Empty]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 15. Linkages Between TAA, One-Stop Career Center Partners and Economic Development Agencies; ETA report number: 2009-16; ETA research area: Methods of expanding U.S. workforce skills; Cost: [Empty]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 16. Rapid Response and TAA; ETA report number: 2009-17; ETA research area: Methods of expanding U.S. workforce skills; Cost: [Empty]; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Report name: 17. Youth Offender Demonstration Project Process Evaluation Round Two; ETA report number: 2004-10; ETA research area: Increasing the labor market participation of underutilized populations; Cost: $1,734,393; Study type: Evaluation; Study design: Descriptive; Methods and techniques: * Case studies; * Secondary analysis; Consistency of analytical approach with research question: Yes; Generalizability of findings: No. Source: GAO analysis of ETA's disseminated studies between January 2008 and March 2010. [A] Two of four reports produced or expected from the WIRED Generation I Evaluation cost a total of $3,433,478 for the full evaluation. [B] Three of five reports produced from the UI Benefits Study cost a total of $6,507,262 for the full study. [C] Four of approximately 10 reports produced or expected from the National Evaluation of Trade Adjustment Assistance cost a total of $10,453,957 for the full evaluation. [End of table] [End of section] Appendix V: Delphi Phase I and Phase II Questionnaires: GAO's Review of ETA's Research Priorities and Dissemination Methods: Phase 1: U.S. Government Accountability Office: Introduction: The Congress has asked the U.S. Government Accountability Office (GAO), an independent agency that gathers information for the Congress, to review the Employment and Training Administration's (ETA) research activities. As part of this review, we have asked you to participate in this virtual expert panel to gather your input on national priorities in the area of employment and training, and on ETA's research agenda and dissemination practices. As you know, you have been identified as an expert in employment and training services or programs, service delivery approaches, and/or evaluation methodologies for these programs. Your participation allows you an opportunity to assist us in informing ETA's future work. The virtual expert panel will consist of multiple phases. For the first stage of the virtual panel, we ask that you complete a brief, five- question survey. We will use the group's responses to these questions to create a more detailed questionnaire in the second stage. This questionnaire will allow you to consider information provided by other members of the group. No information will be attributed to specific individuals. Thank you very much for agreeing to participate in the panel. Your participation in our study is essential for us to provide requested information to the Congress. Our work in this review will cover other areas in addition to those addressed in the virtual panel. To learn more about these additional areas,click here. Contact: If your response will be delayed or if you have any questions, please call or e-mail Ashanta Williams or Kathy White. Click here for their contact information. Confidentiality of Responses: We will not release individually identifiable information from this survey while our work is in progress unless compelled by law or required to do so by the Congress. Once the report is issued, we will destroy the file that links names to individual responses. While the results will generally be provided in summary form in our report, individual answers may be discussed, but they will not include any information that could be used to identify individual respondents. Names of participants will be listed in our final report; however, names will not be associated with responses. Furthermore, your comments and answers will not be provided to other participants during any stage in the process. Navigation, Exiting and Printing the Survey: To learn more about navigation, exiting and printing the survey, please click here. Key Employment and Training Issues: 1. In your opinion, what are the key employment and training issues that ETA's research agenda should address? Please consider population groups, programs, and/or methodological approaches in your response. ETA's research agenda: 2. We understand that ETA is currently reviewing its research agenda and that its research priorities may change moving forward. However, we would like you to consider the priority areas addressed in the 2007- 2012 research agenda along with selected examples of the topics covered in the agenda listed below. Note: This is not a complete or comprehensive list of ETA's disseminated research studies. We compiled this list to give you a better sense of the types of research conducted by ETA. ETA Research Priority Areas, with Example Topics Covered in ETA's Research Agenda: 1. Integration of Workforce and Regional Economic Development: * Job training initiatives to produce high-skill, high-wage jobs; * Strategic partnerships between private business sector and public entities; * Regional and economic development. 2. Methods of Expanding U.S. Workforce Skills: * Rapid response services for dislocated workers; * Costs and benefits of apprenticeship; * Trade adjustment assistance. 3. Increasing the Labor Market Participation of Underutilized Populations: * Enhanced services for the hard-to-employ; * Employment-centered programs for ex-offenders; * Current strategies to employ and retain older workers. 4. Using State-Level Administrative Data to Measure Progress and Outcomes: * Examination of local workforce investment areas o WIA and employment outcomes. 5. Post-Secondary Education and Job Training: * The role of community colleges in workforce development and training; * Career advancement accounts; * Growing America through entrepreneurship. 6. Unemployment Insurance: * Unemployment insurance benefits * Characteristics of unemployed workers * Trends in the structure of the labor market and unemployment Based on your knowledge, how, if at all, do you think ETA's 2007-2012 research agenda reflects the key employment and training issues you identified in question one? 3. In your opinion, how well does ETA's research inform policy and practice? Please consider ongoing research studies, of which you are familiar, during the past 5 years. 4. What additional thoughts, if any, do you have about ETA's research or research agenda? ETA's Dissemination Methods: 5. What dissemination methods do you think would allow ETA to effectively share its research findings? Completed: 6. Have you finished this questionnaire? (Check Only One Answer) 1. Yes. 2. No. Print. Submit. Cancel. [End of section] Menu: Display list of headings: GAO's Review of ETA's Research Priorities and Dissemination Methods: Phase 2: Prioritization for Future Work: Prioritization of ETA's Research: Policy and Practice: ETA's Dissemination Methods: Demographics: Summary: View and print a summary of your responses. Help: If you have any questions about this GAO survey, please contact Ashanta Williams at (202) 512-5110 (williamsar@qao.qoy) or Kathy White at (202) 512-8512 (whitek@gao.gov). Progress: Navigate: GAO's Review of ETA's Research Priorities and Dissemination Methods: Phase 2: U.S. Government Accountability Office: In phase 1, we asked you to provide information on: * key employment and training issues that ETA's research agenda should address; * the extent to which ETA's 2007-2012 research agenda addresses these key employment and training issues; and * dissemination methods that would allow ETA to effectively share its research findings. This is Phase II of our virtual panel. We developed our questions in this phase using the responses of 36 experts who provided input in Phase 1. Your responses will allow us to gather more specific information for the Congress on ETA's research and dissemination practices. We thank you again for your participation in our panel. We appreciate your time and found the information that you provided extremely helpful. Contact: If your response will be delayed or if you have any questions, please contact Ashanta Williams at (202) 512-5110 (williamsa@agao.gov) or Kathy White at (202) 512-8512 (whitek@gao.gov). Directions: * Click the "Exit" button at the bottom of the screen if you wish to save your responses and continue at a later time. Always use the "Exit" button to close the survey to avoid losing data you have entered. Confidentiality of Responses: We will not release individually identifiable information from this survey while our work is in progress unless compelled by law or required to do so by the Congress. Once the report is issued, we will destroy the file that links names to individual responses. While the results will generally be provided in summary form in our report, individual answers may be discussed, but they will not include any information that could be used to identify individual respondents. Names of participants may be listed in our final report; however, names will not be associated with responses. Furthermore, your comments and answers will not be provided to other participants during any stage in the process. Prioritization for Future Work: The questions in this section are based on respondents' answers to the question: In your opinion, what are the key employment and training issues that ETA's research agenda should address? 1. Taking into account ETA's limited resources, how important, if at all, is it for ETA to address the following in future research? a. Long-term outcomes of Employment and Training Programs: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: b. Short-term outcomes of Employment and Training Programs: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: c. Value of various credentials: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: d. Impact of long-term and/or short-term training: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: e. Job creation strategies, such as providing wage subsidies, tax credits, or public service employment: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: f. Employment and training approaches that work and for whom: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: g. One-stop center management and operations: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: h. Effective performance measurement systems: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: i. Linkages between the public workforce system and economic development entities: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: j. Needs of different labor markets or industries: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: k. Employment and training strategies for various economic conditions: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: 1. Issues related to Unemployment Insurance: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: 2. Among the areas you checked at least moderately important in question 1, which would you rank as the top 3 areas that ETA should address in future research. (Please rank only 3 areas, with 1 as your top priority area, 2 as your second highest area, etc.) a. Long-term outcomes of Employment and Training Programs: Rank: b. Short-term outcomes of Employment and Training Programs: Rank: c. Value of various credentials: Rank: d. Impact of long-term and/or short-term training: Rank: e. Job creation strategies, such as providing wage subsidies, tax credits, or public service employment: Rank: f. Employment and training approaches that work and for whom: Rank: g. One-stop center management and operations: Rank: h. Effective performance measurement systems: Rank: i. Linkages between the public workforce system and economic development entities: Rank: j. Needs of different labor markets or industries: Rank: k. Employment and training strategies for various economic conditions: Rank: I. Issues related to Unemployment Insurance: Rank: 3. How important, if at all, would it be for ETA to fund research that focuses on what works for the following populations given its resource constraints? a. Short-term unemployed: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: b. Long-term unemployed: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: c. Dislocated workers: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: d. Older workers: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: e. Veterans: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: f. Economically disadvantaged workers: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: g. Racial and ethnic minorities: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: h. Immigrants: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: i. Adults with low basic skills: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: j. Workers with physical and mental disabilities: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: k. In-school youth (ages 14 - 21): Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: I. Out-of-school youth: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: m. Ex-offenders: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: 4. Among the populations you checked as least moderately important in question 3, which would you rank as the top 5 populations on which ETA should fund research in the future, please rank only 5 populations, with 1 as your top population, 2 as your second highest population, etc.) a. Short-term unemployed: Rank: b. Long-term unemployed: Rank: c. Dislocated workers: Rank: d. Older workers: Rank: e. Veterans: Rank: f. Economically disadvantaged workers: Rank: g. Racial and ethnic minorities: Rank: h. Immigrants: Rank: i. Adults with low basic skills: Rank: j. Workers with physical and mental disabilities: Rank: k. In-school youth (ages 14 - 21): Rank: I. Out-of-school youth: Rank: m. Ex-offenders: Rank: 5. How important are each of the following Unemployment Insurance (UI) system issues for ETA to include in future research efforts? a. Examination of the incentives and disincentives in the UI system: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: b. Modernization of UI system (such as extending coverage to include additional income groups): Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: c. Benefits and costs of UI system in different business cycles: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: d. Accountability mechanisms in the UI system (such as data accuracy, enforcement of work search requirement): Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: e. Linkage between UI and employment and training programs or safety- net programs (such as SNAP, TANF): Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: 6. How important, if at all, is it for ETA to evaluate the following key employment or training programs (excluding Up? a. WIA Adult: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: b. WIA Dislocated Worker: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: c. WIA Youth: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: d. Wagner-Peyser Employment Service: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: e. Job Corps: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: f. Trade Adjustment Assistance: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: g. Veterans' Employment and Training: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: h. Apprenticeship program: Extremely important: Very important: Moderately important: Somewhat important: Not at all important: No response: 7. Given a limited amount of resources, which three of the key employment or training programs that you checked as at least moderately important in question 6 should ETA address in future research. Please rank only 3 employment or training programs, with 1 as your top program, 2 as your second highest program, etc.) a. WIA Adult: Rank: b. WIA Dislocated Worker: Rank: c. WIA Youth: Rank: d. Wagner-Peyser Employment Service: Rank: e. Job Corps: Rank: f. Trade Adjustment Assistance: Rank: g. Veterans' Employment and Training: Rank: h. Apprenticeship program: Rank: 8. How important, if at all, is it for ETA to integrate the following into its future research proposals? a. Randomized experimental design: Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: b. Quasi-experimental design: Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: c. Descriptive studies (including program implementation and operations): Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: d. Analysis of administrative data for research purposes: Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: e. Meta-analyses (including literature reviews or quantitative analyses of prior research): Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: f. Secondary analysis of existing research data: Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: g. Evaluation of pilots and demonstrations: Very important: Somewhat important: Neither important nor unimportant: Somewhat unimportant: Very unimportant: No response: Prioritization of ETA's Research: 9. Overall, to what extent do you feel ETA's 2007-2012 research agenda reflect key employment and training issues? (For a reminder of the topics covered during the 2007-2012 research agenda click here.) * Great extent; * Moderate extent; * Some extent; * Little extent; * Not at all; * No response. Policy and Practice: The questions in this section are based on respondents' answers to the following questions: "In your opinion, how well does ETA's research inform policy and practice? Please consider ongoing research studies, of which you are familiar, during the past 5 years." "What additional thoughts, if any, do you have about ETA's research or research agenda?" 10. In the past 5 years, to what extent has ETA's research informed employment and training policy? * Great extent; * Moderate extent; * Some extent; * Little extent; * Not at all; * Don't know; * No response. 11. In the past 5 years, to what extent has ETA's research informed state and local employment and training practices? * Great extent; * Moderate extent; * Some extent; * Little extent; * Not at all; * Don't know; * No response. ETA's Dissemination Methods: The questions in this section are based on respondents' answers to the following question: "What dissemination methods do you think would allow ETA to effectively share its research findings?" 12. How effective, if at all, do you think the following methods would be for ETA to disseminate its research findings? a. Searchable database of papers on ETA's website: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: b. Distributing a compendium of ETA's research: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: c. Distributing a one-page summary of the research findings from each report: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: d. Email notifications announcing newly released research: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: e. Publishing articles in journals: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: f. Publicizing research findings through press articles, industry publications, etc. Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: g. Contracting with relevant organizations and associations to disseminate research: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: h. Webinars: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: i. Videos or podcasts: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: j. Social media (such as Twitter): Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: k. National ETA-sponsored conferences: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: l. Regional ETA-sponsored conferences: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: m. Presenting at outside conferences: Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: n. Briefings at ETA for external audiences (including stakeholders and policymakers): Extremely effective: Very effective: Moderately effective: Somewhat effective: Not at all effective: No response: 13. Given a limited amount of resources, which five of the dissemination methods that you checked as at least moderately effective in question 12 should ETA focus on for disseminating research in the future. Please rank only 3 methods, with 1 as your top method, 2 as your second highest method, etc.). a. Searchable data base of papers on ETA's website: Rank: b. Distributing a compendium of ETA's research: Rank: c. Distributing a one-page summary of the research findings from each report: Rank: d. Email notifications announcing newly released research: Rank: e. Publishing articles in journals: Rank: f. Publicizing research findings through press articles, industry publications, etc. Rank: g. Contracting with relevant organizations and associations to disseminate research: Rank: h. Webinars: Rank: i. Videos or podcasts: Rank: j. Social media (such as Twitter): Rank: k. National ETA-sponsored conferences: Rank: l. Regional ETA-sponsored conferences: Rank: m. Presenting at outside conferences: Rank: n. Briefings at ETA for external audiences (including stakeholders and policymakers): Rank: 14. Do you have recent experience with any of the following ETA research dissemination methods? If yes, how satisfied or dissatisfied were you with each of the following ETA research dissemination methods? Do you have recent experience using any of the following ETA research dissemination methods? If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? a. Searchable database of research papers on ETA's website: Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. b. ETA's Training and Employment Notice (TENs) e-mail list: Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. c. Webinars: Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. d. National ETA-sponsored conferences: Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. e. Regional ETA-sponsored conferences: Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. f. Briefings at ETA for external audiences (including stakeholders and policymakers): Do you have recent experience using any of the following ETA research dissemination methods? Yes: (Go to next column); No. If yes, how satisfied or dissatisfied were you with the following ETA research dissemination methods? Very satisfied: Satisfied: Neither satisfied nor dissatisfied: Dissatisfied: Very dissatisfied. 15. Are there any additional thoughts you would like to share with us about ETA's research and dissemination practices? Demographics: 16. Which of the following best describe your recent professional experience? (CHECK ALL THAT APPLY) * Researcher * Professor/academic * Consultant; * State Official; * Local official; * Other; * No response. 17. Are you ready to submit your final completed survey to GAO? (This is equivalent to mailing a completed paper survey to us. It tells us that your answers are official and final.) * Yes, my survey is complete - To submit your final responses, please click on "Exit" below. * No, my survey is not yet complete -To save your responses for later, please click on "Exit" below. You may view and print your completed survey by clicking on the Summary link in the menu to the left. Print: Submit: [End of section] Appendix VI: Experts Who Agreed to Participate in GAO's Delphi Panel: Expert: Burt S. Barnow; Affiliation: Trachtenberg School of Public Policy and Public Administration, George Washington University. Expert: Jon Baron; Affiliation: Coalition for Evidence-Based Policy. Expert: Stephen Bell; Affiliation: Abt Associates Inc. Expert: Jacob Benus; Affiliation: IMPAQ International, LLC. Expert: Dan Bloom; Affiliation: MDRC. Expert: Gary Burtless; Affiliation: Brookings Institution. Expert: Paul Decker; Affiliation: Mathematica Policy Research, Inc. Expert: Randall Eberts; Affiliation: W.E. Upjohn Institute for Employment Research. Expert: Richard Freeman; Affiliation: Harvard University; and; National Bureau of Economic Research. Expert: Robert Giloth; Affiliation: Annie E. Casey Foundation. Expert: David Heaney; Affiliation: Maximus. Expert: Carolyn J. Heinrich; Affiliation: Robert M. LaFollette School of Public Affairs, University of Wisconsin-Madison. Expert: Kevin Hollenbeck; Affiliation: W.E. Upjohn Institute for Employment Research. Expert: Louis Jacobson; Affiliation: New Horizons Economic Research. Expert: Richard Kazis; Affiliation: Jobs for the Future. Expert: Jacob Alex Klerman; Affiliation: Abt Associates Inc. Expert: Kathy Krepcio; Affiliation: John J. Heldrich Center for Workforce Development, Rutgers, The State University of New Jersey. Expert: Robert L. Lerman; Affiliation: Department of Economics, College of Arts and Sciences, American University. Expert: Alberto Martini; Affiliation: Progetto Valutazione. Expert: Sheena McConnell; Affiliation: Mathematica Policy Research, Inc. Expert: Rick McGahey; Affiliation: Ford Foundation. Expert: Peter Mueser; Affiliation: Department of Economics, University of Missouri-Columbia. Expert: Lee Munnich; Affiliation: Humphrey School of Public Affairs, University of Minnesota. Expert: Burke Murphy; Affiliation: Minnesota Department of Employment and Economic Development. Expert: Sigurd R. Nilsen; Affiliation: Formerly of GAO. Expert: Demetra Smith Nightingale; Affiliation: Urban Institute. Expert: Christopher O'Leary; Affiliation: W.E. Upjohn Institute for Employment Research. Expert: Marion Pines; Affiliation: Institute for Policy Studies, Johns Hopkins University. Expert: James Riccio; Affiliation: MDRC. Expert: Neil Ridley; Affiliation: Center for Law and Social Policy. Expert: Howard Rosen; Affiliation: Peterson Institute for International Economics. Expert: Peter Z. Schochet; Affiliation: Mathematica Policy Research, Inc. Expert: Kenneth R. Troske; Affiliation: University of Kentucky, Department of Economics. Expert: Jason Turner; Affiliation: Heritage Foundation. Expert: John Twomey; Affiliation: New York Association of Training and Employment Professionals. Expert: Ray Uhalde; Affiliation: Job for the Future. Expert: Carl Van Horn; Affiliation: John J. Heldrich Center for Workforce Development, Rutgers, The State University of New Jersey. Expert: John W. Wallace; Affiliation: Formerly of MDRC. Expert: Jeffrey B. Wenger; Affiliation: Department of Public Administration and Policy, School of Public and International Affairs, University of Georgia. Expert: Michael Wiseman; Affiliation: George Washington Institute of Public Policy, George Washington University. Expert: Steve A. Woodbury; Affiliation: Department of Economics, Michigan State University; and; W.E. Upjohn Institute for Employment Research. Source: GAO. [End of table] [End of section] Appendix VII: Comments from the Department of Labor: U.S. Department of Labor: Assistant Secretary for Employment and Training: Washington, D.C. 20210 March 7, 2011: Mr. George Scott: Director, Education, Workforce and Income Security Issues: U.S. Government Accountability Office: 441 G. Street, N.W. Washington, D.C. 20548: Dear Mr. Scott: Thank you for the opportunity to review the Government Accountability Office (GAO) draft report entitled: "Employment and Training Administration: More Actions Needed to Improve Transparency and Accountability of Its Research Program." We continue to look for ways to strengthening the Employment and Training Administration's (ETA) research program, we appreciate the information provided in the report and the opportunity to comment. The Department of Labor (DOL) agrees with all three recommendations. GAO Recommendation 1: Formally incorporate into its research process the routine involvement of Chief Evaluation Office at key milestones, including at the development of ETA's annual research agenda and spending priorities, as well as at the early stages of developing specific research projects. DOL Response: The Department agrees with this recommendation and is in the process of formally involving the Chief Evaluation Officer (CEO) into its research process. In fact, ETA has worked closely with the CEO since its establishment in spring 2010. Before setting formal monthly meetings that began in fall 2010, ETA routinely sought involvement of CEO to discuss research, demonstration projects, and evaluations from the early development stages. Topics covered include implications of research findings, strategic planning, and project and evaluation development of the Green Jobs High Growth Health Care Impact Evaluation, the Young Parents Demonstration and Impact Evaluation, the Enhanced Transitional Jobs Demonstrations and Random Assignment Evaluation, and the Unemployment Insurance-related American Recovery and Reinvestment Act studies. Currently, ETA also routinely involves the CEO in: 1) Reviewing research and demonstration statements of work; 2) Discussing research and evaluation methodologies with ETA and the Office of Management and Budget and periodically, with other federal agencies; 3) Discussing, on a monthly basis, on-going ETA research and evaluation projects; 4) Reviewing research and evaluation design reports, interim reports and final reports prior to their dissemination; 5) Participating in meetings with grantees involved in major research demonstrations to help explain the importance of rigorous evaluations; and; 6) Providing input on how to best incorporate research in ETA's Solicitations for Grant Applications (SGAs) by strengthening the evaluation and report sections of the SGAs and crafting language to encourage applicants' use of evidence-based practices. Going forward, ETA plans to continue its collaboration with the CEO and involve the CEO in earlier stages of the research development process, including meetings to discuss potential research, demonstration and evaluation projects. GAO Recommendation 2: Develop a mechanism to enhance the transparency and accountability of ETA's research program. For example, involve advisory bodies or other entities outside ETA, in efforts to develop research policies and processes. DOL Response: The Department agrees with this recommendation and ETA already has taken several steps to engage outside experts more fully to enhance the transparency and accountability of its research program. Specifically, ETA is: (1) including outside experts in the development of its five-year research plan; (2) forming advisory and peer review groups for major evaluations, (3) developing, with the Coalition for Evidence-Based Policy, the capability to conduct peer reviews on an ad hoc basis, and (4) having the CEO, who is in the Department but independent of ETA, also review reports. In addition, ETA is collecting and synthesizing feedback from a variety of sources for developing its five-year research plan including: (1) input received from experts and workforce system practitioners at ETA's conferences, and (2) feedback solicited from stakeholders in the workforce community. ETA also will analyze public comments that it will solicit when the draft plan is announced in the Federal Register and posted on the ETA Web site. ETA has strengthened its collaboration with other Federal agency experts at the U.S. Departments of Education and Health and Human Services in planning and carrying out research and evaluation studies. Moving forward, ETA plans to continue engaging outside experts in developing research policies and processes. GAO Recommendation 3: Develop a formal plan for ensuring that ETA's research products are easily accessible to stakeholders and to the general public through its searchable database. Such a plan could involve requiring labor to assess overall effectiveness of its Web based search page, including user satisfaction with search features. DOL Response: The Department agrees with this recommendation, and ETA has taken steps to establish specific processes for disseminating research and evaluation findings in an effort to ensure that its research products are easily accessible to stakeholders and to the general public. Specifically, ETA has improved the web-based publication search engine, and featured research and evaluation findings at agency-sponsored or other Federally-sponsored conferences. ETA supports continuous improvement of its dissemination efforts to ensure public accessibility. For example, last summer ETA upgraded the web-based research database to improve its overall effectiveness and user satisfaction by implementing the Google Search on all doleta.gov sites and developing an interface between the Google Search Appliance and ETA Research Publication Database. Since the improvements cited in GAO's report, additional improvements have been made; when a keyword search is performed in the search box, located in the top navigation bar on doleta.gov, it returns a list of relevant publications housed in the ETA Research Publication Database. As part of ETA's plan to continue improving its dissemination practices, we have linked to the customer satisfaction survey form provided at the bottom of all Departmental Web pages. This survey provides feedback from visitors to help ensure that the Web site is useful, accurate and complete. ETA's Web Services team monitors the system log daily and takes immediate action necessary to rectify any issues. The Web Services team will periodically obtain feedback on operation of the Research Publication Database and will gather new requirements for further cost-effective enhancements that will be specified in an improvement plan. Additional Remarks: The Department believes that without further clarifications on the agency pilots, demonstration and research budget, the draft report would be misleading as to the funds available for undesignated pilots, demonstrations and research. We would encourage you to consider the Department's technical comments on the draft report, enclosed for your reference. Again, thank you for the opportunity to review the draft report. If you would like additional information, please do not hesitate to call me at (202) 693-2700. Sincerely, Signed by: Jane Oates: Assistant Secretary: Enclosure: [End of section] Appendix VIII: GAO Contact and Staff Acknowledgments: Contact: George A. Scott, (202) 512-7215 or scottg@gao.gov: Staff Acknowledgments: In addition to the contact listed above, Dianne Blank, Assistant Director, and Kathleen White, analyst-in-charge, managed all phases of the engagement. Ashanta Williams assisted in managing many aspects of the work and was responsible for final report preparation. Lucas Alvarez and Benjamin Collins made significant contributions to all aspects of this report. In addition, Amanda Miller assisted with study and questionnaire design; Joanna Chan performed the data analysis; Stephanie Shipman advised on evaluation approaches; James Bennett provided graphics assistance; David Chrisinger provided writing assistance; Alex Galuten and Sheila McCoy provided legal support; and Sheranda Campbell and Ryan Siegel verified our findings. [End of section] GAO Related Products: Program Evaluation: Experienced Agencies Follow a Similar Model For Prioritizing Research. [hyperlink, http://www.gao.gov/products/GAO-11-176]. Washington, D.C.: January 14, 2011. Employment and Training Administration: Increased Authority and Accountability Could Improve Research Program. [hyperlink, http://www.gao.gov/products/GAO-10-243]. Washington, D.C.: January 29, 2010. Workforce Investment Act: Labor Has Made Progress in Addressing Areas of Concern, but More Focus Needed on Understanding What Works and What Doesn't. [hyperlink, http://www.gao.gov/products/GAO-09-396T]. Washington, D.C.: February 26, 2009. Employment and Training Program Grants: Evaluating Impacts and Enhanced Monitoring Would Improve Accountability. [hyperlink, http://www.gao.gov/products/GAO-08-486]. Washington, D.C.: May 7, 2008. Federal Research: Policies Guiding the Dissemination of Scientific Research from Selected Agencies Should Be Clarified and Better Communicated. [hyperlink, http://www.gao.gov/products/GAO-07-653]. Washington, D.C.: May 17, 2007. Data Quality: Expanded Use of Key Dissemination Practices Would Further Safeguard the Integrity of Federal Statistical Data. [hyperlink, http://www.gao.gov/products/GAO-06-607]. Washington, D.C.: May 31, 2006. Workforce Investment Act: Substantial Funds Are Used for Training, but Little Is Known Nationally about Training Outcomes. [hyperlink, http://www.gao.gov/products/GAO-05-650]. Washington, D.C.: June 29, 2005. Program Evaluation: An Evaluation Culture and Collaborative Partnerships Help Build Agency Capacity. [hyperlink, http://www.gao.gov/products/GAO-03-454]. Washington, D.C.: May 2, 2003. Workforce Investment Act: Improvements Needed in Performance Measures to Provide a More Accurate Picture of WIA's Effectiveness. [hyperlink, http://www.gao.gov/products/GAO-02-275]. Washington, D.C.: February 1, 2002. [End of section] Footnotes: [1] A program year begins on July 1 of a year and ends on June 30 of the following year. Thus, program year 2010 began on July 1, 2010, and ends on June 30, 2011. Of the $103 million appropriated for program year 2010, Congress designated about $84 million for specific projects, including $30 million for Transitional Jobs activities and $5.5 million for the employment and training needs of young parents. The remainder of the funds, approximately $18.7 million, was available to ETA for undesignated pilots, demonstrations, and research. [2] GAO, Workforce Investment Act: Labor Has Made Progress in Addressing Areas of Concern, but More Focus Needed on Understanding What Works and What Doesn't, [hyperlink, http://www.gao.gov/products/GAO-09-396T] (Washington, D.C.: Feb. 26, 2009); Employment and Training Program Grants: Evaluating Impact and Enhancing Monitoring Would Improve Accountability, [hyperlink, http://www.gao.gov/products/GAO-08-486] (Washington, D.C.: May 7, 2008); and Employment and Training Program Grants: Labor Has Outlined Steps for Additional Documentation and Monitoring but Assessing Impact Still Remains an Issue, [hyperlink, http://www.gao.gov/products/GAO-08-1140T] (Washington, D.C.: Sept. 23, 2008). [3] GAO, Employment and Training Administration: Increased Authority and Accountability Could Improve Research Program, [hyperlink, http://www.gao.gov/products/GAO-10-243] (Washington, D.C.: Jan. 29, 2010). [4] 29 U.S.C. § 2916(a). [5] While ETA's program year 2010 began on July 1, 2010, ETA's funding for pilots, demonstrations, and research was available starting on April 1, 2010. [6] American Evaluation Association, An Evaluation Roadmap for a More Effective Government, (September 2010), available at [hyperlink, www.eval.org/eptf.asp]. [7] Congress enacted WIA in part to increase employment, retention, earnings, and occupational skill attainment of participants, thus improving the quality of the workforce. WIA Adult primarily focuses on disadvantaged adults and WIA Dislocated Workers focuses largely on workers who have been laid off or have been notified that they will be laid off. WIA Youth helps to prepare youth for employment and/or postsecondary education through linkages between academic and occupational learning. [8] See [hyperlink, http://www.gao.gov/products/GAO-08-486]. [9] ETA now has a contract for $1 million with the firm that performed the original GATE I evaluation to perform a 5-year follow-up to assess long-term outcomes for Project GATE II. [10] These studies should include designs that randomly assign individuals to groups that receive enhanced program services and to groups that do not. [11] GAO, Program Evaluation: A Variety of Rigorous Methods Can Help Identify Effective Interventions, [hyperlink, http://www.gao.gov/products/GAO-10-30] (Washington, D.C.: Nov. 23, 2009). [12] See [hyperlink, http://www.gao.gov/products/GAO-10-243]. [13] See [hyperlink, http://www.gao.gov/products/GAO-10-243]. [14] For more information on the criteria several other federal agencies use to prioritize their research, see GAO, Program Evaluation: Experienced Agencies Follow a Similar Model for Prioritizing Research, [hyperlink, http://www.gao.gov/products/GAO-11-176] (Washington, D.C.: Jan. 14, 2011). [15] While ETA has a provision that allows contractors to publicly release reports that have not been approved by the Assistant Secretary after 9 months, researchers may be reluctant to do so out of concern that this action may damage their credibility with the agency and limit their ability to win future research contracts. [16] See [hyperlink, http://www.gao.gov/products/GAO-10-243]. [17] A fourth study, the Impact Evaluation of the Senior Community Service Employment Program, was also designed as a random assignment study. We omitted it from our analysis because ETA was reconsidering the funding of this evaluation at the time of our report. [18] See [hyperlink, http://www.gao.gov/products/GAO-10-30]. [19] 29 U.S.C. § 2917(c). [20] In addition to OMB's review, the evaluation contractor also conducted a peer review of the study design. [21] [hyperlink, http://www.gao.gov/products/GAO-10-243]. [22] See Interagency Committee on Government Information, Recommended Policies and Guidelines for Federal Public Websites (2004). The Interagency Committee on Government Information was established by OMB in 2003 to improve the usability of federal public Web sites. In particular, they recommend that "organizations should provide help, hints, or tips, and include examples" for search users. [23] The Delphi method, developed by the RAND Corporation in the 1950s, is most commonly applied in a group-discussion forum. We modified the approach to have the group discussion take place in the form of a Web-based forum. [24] Labor awarded about $9.9 million to various entities to carry out its Young Parents Demonstration for fiscal years 2008 and 2009. Lancaster County Workforce Investment Board was 1 of 13 grantees that received funding. [25] In June 2008, ETA awarded GATE II grants to four states--one of which was to Virginia--for the extension of the GATE model for helping selected dislocated workers create their own business. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Phone: The price of each GAO publication reflects GAO‘s actual cost of production and distribution and depends on the number of pages in the publication and whether the publication is printed in color or black and white. Pricing and ordering information is posted on GAO‘s Web site, [hyperlink, http://www.gao.gov/ordering.htm]. Place orders by calling (202) 512-6000, toll free (866) 801-7077, or TDD (202) 512-2537. Orders may be paid for using American Express, Discover Card, MasterCard, Visa, check, or money order. Call for additional information. To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.