Yucca Mountain

Quality Assurance at DOE's Planned Nuclear Waste Repository Needs Increased Management Attention Gao ID: GAO-06-313 March 17, 2006

The Department of Energy (DOE) is working to obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain in Nevada. The project, which began in the 1980s, has been beset by delays. In a 2004 report, GAO raised concerns that persistent quality assurance problems could further delay the project. Then, in 2005, DOE announced the discovery of employee e-mails suggesting quality assurance problems, including possible falsification of records. Quality assurance, which establishes requirements for work to be performed under controlled conditions that ensure quality, is critical to making sure the project meets standards for protecting public health and the environment. GAO was asked to examine (1) the history of the project's quality assurance problems, (2) DOE's tracking of these problems and efforts to address them since GAO's 2004 report, and (3) challenges facing DOE as it continues to address quality assurance issues within the project.

DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the 1980s and 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. More recently, as it prepares to submit a license application for the repository to NRC, DOE has been relying on costly and time-consuming rework to resolve lingering quality assurance problems uncovered during audits and after-the-fact evaluations. DOE announced, in 2004, that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed. However, GAO found that the project's performance indicators and other key management tools were not effective for this purpose. For example, the management tools did not target existing areas of concern and did not track progress in addressing them. The tools also had weaknesses in detecting and highlighting significant problems for management attention. DOE continues to face quality assurance and other challenges. First, DOE is engaged in extensive efforts to restore confidence in scientific documents because of the quality assurance problems suggested in the discovered e-mails between project employees, and it has about 14 million more project e-mails to review. Second, DOE faces quality assurance challenges in resolving design control problems associated with its requirements management process--the process for ensuring that high-level plans and regulatory requirements are incorporated into specific engineering details. Problems with the process led to the December 2005 suspension of certain project work. Third, DOE continues to be challenged to manage a complex program and organization. Significant personnel and project changes initiated in October 2005 create the potential for confusion over roles and responsibilities--a situation DOE found to contribute to quality assurance problems during an earlier transition.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-06-313, Yucca Mountain: Quality Assurance at DOE's Planned Nuclear Waste Repository Needs Increased Management Attention This is the accessible text file for GAO report number GAO-06-313 entitled 'Yucca Mountain: Quality Assurance at DOE's Planned Nuclear Waste Repository Needs Increased Management Attention' which was released on March 23, 2006. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to the Chairman, Subcommittee on the Federal Workforce and Agency Organization, Committee on Government Reform, House of Representatives: March 2006: Yucca Mountain: Quality Assurance at DOE's Planned Nuclear Waste Repository Needs Increased Management Attention: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-06-313]: GAO Highlights: Highlights of GAO-06-313, a report to the Chairman, Subcommittee on the Federal Workforce and Agency Organization, Committee on Government Reform, House of Representatives: Why GAO Did This Study: The Department of Energy (DOE) is working to obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain in Nevada. The project, which began in the 1980s, has been beset by delays. In a 2004 report, GAO raised concerns that persistent quality assurance problems could further delay the project. Then, in 2005, DOE announced the discovery of employee e-mails suggesting quality assurance problems, including possible falsification of records. Quality assurance, which establishes requirements for work to be performed under controlled conditions that ensure quality, is critical to making sure the project meets standards for protecting public health and the environment. GAO was asked to examine (1) the history of the project‘s quality assurance problems, (2) DOE‘s tracking of these problems and efforts to address them since GAO‘s 2004 report, and (3) challenges facing DOE as it continues to address quality assurance issues within the project. What GAO Found: DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the 1980s and 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. More recently, as it prepares to submit a license application for the repository to NRC, DOE has been relying on costly and time-consuming rework to resolve lingering quality assurance problems uncovered during audits and after-the-fact evaluations. DOE announced, in 2004, that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed. However, GAO found that the project‘s performance indicators and other key management tools were not effective for this purpose. For example, the management tools did not target existing areas of concern and did not track progress in addressing them. The tools also had weaknesses in detecting and highlighting significant problems for management attention. DOE continues to face quality assurance and other challenges. First, DOE is engaged in extensive efforts to restore confidence in scientific documents because of the quality assurance problems suggested in the discovered e-mails between project employees, and it has about 14 million more project e-mails to review. Second, DOE faces quality assurance challenges in resolving design control problems associated with its requirements management process”the process for ensuring that high-level plans and regulatory requirements are incorporated into specific engineering details. Problems with the process led to the December 2005 suspension of certain project work. Third, DOE continues to be challenged to manage a complex program and organization. Significant personnel and project changes initiated in October 2005 create the potential for confusion over roles and responsibilities”a situation DOE found to contribute to quality assurance problems during an earlier transition. View of Yucca Mountain and the Exploratory Tunnel for the Repository: [See PDF for image] [End of figure] What GAO Recommends: GAO recommends five actions DOE can take to improve the project‘s management tools and identify and address quality assurance and other problems. In oral comments, DOE agreed with GAO‘s recommendations. www.gao.gov/cgi-bin/getrpt?GAO-06-313. To view the full product, including the scope and methodology, click on the link above. For more information, contact Jim Wells at (202) 512- 3841 or wellsj@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: DOE Has a Long History of Quality Assurance Problems at Yucca Mountain and Is Relying on Costly and Time-Consuming Measures to Correct Problems Before Submitting Its License Application for the Repository: DOE Cannot Be Certain Its Efforts to Improve Quality Assurance Have Been Effective Because of Weaknesses in Tracking Progress and Identifying Problems: DOE's 'New Path Forward' for Preparing to Submit Its License Application Faces Substantial Quality Assurance and Other Challenges: Conclusions: Recommendations for Executive Action: Agency Comments: Appendixes: Appendix I: Objectives, Scope, and Methodology: Appendix II: Yucca Mountain Project Employee Concerns Programs: Appendix III: GAO Contact and Staff Acknowledgments: Tables: Table 1: Visibility of Management Improvement Initiatives' Effectiveness Indicators in Annunciator Panel When Last Prepared (Using August 2005 Data): Table 2: Key Indicators for Processes with Quality Elements, Their Intended Focus, and Number of Times They Changed (March through August 2005): Table 3: Employee Concerns Opened for Investigation under DOE's Employee Concerns Program by Category of Concern, January through November 2005: Table 4: Employee Concerns Opened for Investigation under BSC's Employee Concerns Program by Category of Concern, January through November 2005: Figures: Figure 1: A Yucca Mountain Project Scientist Conducts Water Infiltration Tests inside Yucca Mountain: Figure 2: The 1.7-Mile Tunnel Built for Scientific Studies near the Potential Repository Area: Abbreviations: AMR: Analysis and Model Report: BSC: Bechtel/SAIC Company, LLC: CAP: Corrective Action Program: CR: condition report: DOE: Department of Energy: EPA: Environmental Protection Agency: Initiatives: Management Improvement Initiatives: INPO: Institute of Nuclear Power Operations: LSN: Licensing Support Network: NRC: Nuclear Regulatory Commission: OCRWM: Office of Civilian Radioactive Waste Management: RIT: Regulatory Integration Team: USGS: U.S. Geological Survey: Letter March 17, 2006: The Honorable Jon C. Porter: Chairman: Subcommittee on the Federal Workforce and Agency Organization: Committee on Government Reform: House of Representatives: Dear Mr. Chairman: The nuclear waste created as a by-product of the nuclear power process in reactors can remain highly radioactive for hundreds of thousands of years, and will require proper disposal to protect public health and the environment. Over 50,000 metric tons of this waste is currently being stored at 72 sites around the country, principally at commercial nuclear power plants. These wastes have been accumulating for several decades in surface storage designed to be temporary. The Nuclear Waste Policy Act of 1982 obligated the Department of Energy (DOE) to construct an underground geological repository for permanent storage and begin accepting these wastes by January 31, 1998. However, it was not until 2002, after more than 15 years of scientific study, that Yucca Mountain in Nevada was approved by Congress as a suitable location for the repository. DOE is continuing to experience delays, and it does not currently have a schedule for when construction of the repository will begin. The project to build and operate a repository at Yucca Mountain is highly complex. It is also highly controversial among some of the public, in large part, because of their concern that the repository may not be adequate, over the long term, to prevent the release of radioactive material to the environment. DOE has established quality assurance procedures to ensure that its work relative to the project and the technical information it produces are of high quality and defensible. However, persistent problems with implementing these procedures and resulting questions about the quality of the work have significantly contributed to project delays. Resolving these quality issues is essential to proceeding with construction. To construct a repository at the Yucca Mountain site, DOE must obtain a license from the Nuclear Regulatory Commission (NRC). As part of the licensing process, DOE must demonstrate to NRC that its plans for the repository will meet standards for protecting public health and the environment from harmful exposure to the radioactive waste. The Environmental Protection Agency (EPA) set these standards in 2001, but as a result of a 2004 court ruling, EPA is proposing to revise the standards to extend the protection period from 10,000 years to 1 million years.[Footnote 1] To demonstrate that it can meet these standards, DOE has been conducting scientific and technical studies at the Yucca Mountain site that will serve as supporting documentation for DOE's planned license application. For example, it has developed mathematical models to measure the probability that various combinations of natural and engineered (human-made) features of a repository could safely contain waste for the long term; the models take into account possible water infiltration through the mountain (see fig. 1), earthquakes, volcanic action, or other scenarios. Thus, one of DOE's most important tasks during the licensing process will be to demonstrate the adequacy of its data, software, and models. Accordingly, NRC requires nuclear facilities to develop a quality assurance program that ensures that the technical information submitted in support of a license application-- such as scientific data, models, and details on design and construction--is well documented and defensible. More specifically, data used to support conclusions about the safety and design of the repository must meet transparency and traceability standards. That is, the data must be clear in justifying and explaining any underlying assumptions, calculations, and conclusions, and must be capable of being traced back to original source materials. Figure 1: A Yucca Mountain Project Scientist Conducts Water Infiltration Tests inside Yucca Mountain: [See PDF for image] [End of figure] To meet NRC's requirements, DOE established a quality assurance program for the Yucca Mountain project. The program establishes requirements that scientific, design, engineering, and other work, such as procurement and record keeping, is to be performed under controlled conditions that ensure quality and enable the work to be verified by others. For example, the program establishes general requirements for calibrating equipment before conducting tests, stipulating when and how the equipment should be calibrated and how to document the results. The project's line organizations, which are responsible for carrying out various functions or aspects of the work, then create their own policies and procedures to implement the requirements. Project employees are required to follow such procedures to help ensure the reliability of project information. Quality assurance auditors periodically verify that the procedures have been followed. Project employees, including quality assurance auditors, are required to identify when procedures are not being followed or when they encounter problems with the procedures. These problems can be identified in "condition reports" under the project's Corrective Action Program, which establishes procedures for the prompt identification and correction of problems. Alternatively, project employees can submit problems for resolution through the Employee Concerns Program, which allows for submissions to be confidential or anonymous. Because quality assurance plays a key role in ensuring that the information DOE uses to support its license application is of high quality and fully defensible, problems in this area raise concerns about delays to DOE's submission and NRC's review of the license application. In April 2004, for example, we reported that recurring quality assurance problems at the Yucca Mountain project could delay the licensing and operation of the repository.[Footnote 2] As we noted, a 2004 NRC evaluation found quality assurance problems such as data that could not be readily traced back to their sources. NRC indicated that unless DOE rectified such problems before submitting the license application, NRC could be in the position of requesting large volumes of additional information, which could prevent it from making a decision on the license within the time required by law. Then, in early 2005, DOE reported it had discovered a series of e-mail messages among some U.S. Geological Survey (USGS) employees working on the Yucca Mountain project under a contract with DOE that appeared to imply that workers had falsified records for scientific work. Several of these messages, written in the late 1990s, appeared to show disdain for the project's quality assurance program and its requirements. As a result of these e-mails, DOE is engaging in an extensive review of records to restore confidence in scientific documents that will be used to support its license application. DOE's recent efforts to better manage quality assurance problems include its Management Improvement Initiatives (Initiatives), which began in 2002 and were reported completed in April 2004. The Initiatives' purpose was to ensure that work and products consistently met quality objectives and were fully defensible by establishing a foundation for continuous improvement in areas of identified management weaknesses. In our 2004 report, we concluded that, while DOE considered the Initiatives to have been completed, it could not assess their effectiveness in addressing the management weaknesses because its performance goals lacked objective measures and time frames for determining success.[Footnote 3] By the end of the Initiatives, DOE had established two tools to alert management about quality-related and other problems: (1) a one-page summary of performance indicators for key project activities and processes (the summary, which DOE refers to as a "panel," is prepared monthly for discussion and action by project managers) and (2) quarterly trend evaluation reports analyzing patterns and trends in problems identified through the Corrective Action Program. Then, in October 2005, DOE initiated planning for an aggressive series of changes to the facility design, organization, and management of the Yucca Mountain project. This effort, known as the "new path forward," is intended to address quality assurance and other challenges prior to submission of a license application. According to the project's Acting Director, DOE will be considering changes in performance indicators and other management tools to better support the new path forward. In this context, you requested that we provide additional information on the project's quality assurance problems and DOE's efforts to correct them. As agreed with your office, this report discusses (1) the history of the project's quality assurance problems since its start in the 1980s, (2) DOE's tracking of quality problems and progress implementing quality assurance requirements since our April 2004 report, and (3) challenges DOE faces as it continues to address quality assurance issues at the project. In addition, you asked for information about concerns raised in recent years through the project's Employee Concerns Program, which is provided in appendix II. To determine the history of quality assurance problems, we reviewed previous GAO, DOE, and NRC documents, visited the project, and interviewed officials from DOE, NRC, and Bechtel/SAIC Company, LLC (BSC), which is DOE's management contractor for the Yucca Mountain project. To assess DOE's tracking of quality-related problems and progress in addressing them, we examined management tools and associated documentation, such as monthly indicator panels and quarterly trend reports, and interviewed BSC and DOE officials regarding those tools. To identify current quality assurance and other challenges, we attended quarterly NRC management meetings, interviewed the Acting Director and other senior managers of the DOE project, and gathered information on management turnover. Due to the criminal investigation under way related to possible falsification of records implied in USGS e-mail exchanges, we did not examine the investigated issues beyond confirming that a concern about the e-mails had been submitted to the Employee Concerns Program. However, to determine if concerns about other instances of potential falsification of records had been raised by project employees, we reviewed employee concerns filed with the project's Employee Concerns Program from January 2004 to December 2005. More information on our scope and methodology is provided in appendix I. We conducted our work from July 2005 through January 2006 in accordance with generally accepted government auditing standards. Results in Brief: DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the late 1980s and early 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. For example, as GAO reported in 1988, NRC had found that DOE's quality assurance procedures were inadequate and its efforts to independently identify and resolve weaknesses in the procedures were ineffective. By the late 1990s, DOE had largely addressed NRC's concerns about its plans and procedures, but its own audits identified quality assurance problems with the data, software, and models used in the scientific work supporting its potential license application. For example, in 1998, a team of project personnel determined that 87 percent of the models used to simulate the site's natural and environmental conditions, and to demonstrate the future repository's performance over time, did not comply with requirements for demonstrating their accuracy in predicting geologic events. More recently, as it prepares to submit the license application for the planned repository to NRC, DOE has been relying on costly and time- consuming rework to resolve lingering quality assurance concerns. For example, to address problems with the transparency and traceability of scientific work in technical documents, DOE implemented, in the spring of 2004, a roughly $20 million, 8-month project called the Regulatory Integration Team. This effort involved about 150 full-time employees from DOE, USGS, and multiple national laboratories, such as Sandia, Los Alamos, and Lawrence Livermore, working to inspect technical documents to identify and resolve quality problems. DOE cannot be certain that its efforts to improve the implementation of its quality assurance requirements have been effective because it adopted management tools that did not target existing management concerns and did not track progress with significant and recurring problems. Although DOE announced, in 2004, that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed, its adopted management tools have not been effective for this purpose. Specifically, the one-page summary, or "panel," of selected performance indicators that project managers used in monthly management meetings was not an effective tool for assessing progress. The indicators selected for the panel poorly represented the major management concerns and changed frequently. For example, the panel did not include an indicator to represent the management concern about unclear roles and responsibilities--a problem that could undermine accountability within the project. Use of the indicator panel was discontinued in late 2005, and DOE is deciding on a tool to replace it. Moreover, a second management tool--trend evaluation reports--also did not track relevant concerns. The reports generally had technical weaknesses for identifying recurrent and significant problems and inconsistently tracked progress in resolving the problems. For example, lacking reliable data and an appropriate performance benchmark for determining the significance of human errors as a cause of quality problems, DOE's trend reports offered no clear basis for tracking progress on such problems. In addition, under the trend reports' rating categories, the rating assigned to convey the significance of a problem was overly influenced by a judgment that there were already ongoing management actions to address the problem, rather than solely assessing the problem's significance. For example, the trend report's rating of one particular problem at the lowest level of significance did not accurately describe the problem or sufficiently draw management's attention to it. Before DOE submits a license application, its aggressive "new path forward" effort faces substantial quality assurance and other challenges. First, the March 2005 announcement of the discovery of USGS e-mails suggesting the possible falsification of quality assurance records has resulted in extensive efforts to restore confidence in scientific documents, and DOE is conducting a wide-ranging review of approximately 14 million e-mails to determine whether they raise additional quality assurance issues. Such a review creates a challenge not just because of the sheer volume of e-mails to be reviewed, but also because DOE will have to decipher their meaning and determine their significance, sometimes without clarification from authors who have left the project. Furthermore, if any of the e-mails raise quality assurance concerns, further review, inspection, or rework may need to be performed to resolve any newfound problems. Second, DOE faces quality assurance challenges in resolving design control problems associated with an inadequate requirements management process--the process responsible for ensuring that broad plans and regulatory requirements affecting the project are tracked and incorporated into specific engineering details. In December 2005, DOE issued a stop-work order on some design and engineering work until DOE can determine that the requirements management process has been improved. Third, DOE continues to be challenged by managing a changing and complex program and organization. Significant project changes initiated in October 2005 under the new path forward create the potential for confusion over accountability as roles and responsibilities change--a situation DOE found to contribute to quality assurance problems during an earlier transition period. For example, a proposed reorganization, establishing a lead laboratory to assist the project, not only would have to be effectively managed, but also would introduce a new player whose accountability DOE would have to ensure. DOE has also experienced turnover in 9 of 17 key management positions since 2001--including positions related to quality assurance--that has created management continuity challenges. For example, the director position for the project has been occupied by three individuals since 1999 and is currently occupied by an acting director. Since DOE is still formulating its plans, it is too early to determine whether its new effort will effectively resolve these challenges. We are making recommendations to DOE aimed at improving the effectiveness of its management tools for monitoring performance in key areas, including quality assurance, by improving the tools' ability to identify problems and track progress in addressing them. We provided DOE and NRC with draft copies of this report for their review and comment. In comments, DOE agreed with our recommendations. Both DOE and NRC provided technical and editorial comments that we incorporated into the report, as appropriate. Background: Congress enacted the Nuclear Waste Policy Act of 1982 to establish a comprehensive policy and program for the safe, permanent disposal of commercial spent fuel and other highly radioactive wastes in one or more mined geologic repositories. The act charged DOE with (1) establishing criteria for recommending sites for repositories; (2) "characterizing" (investigating) three sites to determine each site's suitability for a repository (1987 amendments to the act directed DOE to investigate only the Yucca Mountain site); (3) recommending one suitable site to the President, who, if he considered the site qualified for a license application, would submit a recommendation to Congress; and (4) seeking a license from NRC to construct and operate a repository at the approved site. The act created the Office of Civilian Radioactive Waste Management within DOE to manage its nuclear waste program. Since the 1980s, DOE has spent years conducting site characterization studies at the Yucca Mountain site to determine whether it is suitable for a high-level radioactive waste and spent nuclear fuel repository. DOE, for example, has completed numerous scientific studies of the mountain and its surrounding region for water flow and the potential for rock movement, including volcanoes and earthquakes that might adversely affect the performance of the repository. To allow scientists and engineers greater access to the rock being studied, DOE excavated two tunnels for studying the deep underground environment: (1) a five- mile main tunnel that loops through the mountain, with several research areas or alcoves connected to it; and (2) a 1.7-mile tunnel that crosses the mountain (see fig. 2). This second tunnel allows scientists to study properties of the rock and the behavior of water near the potential repository area. In July 2002, Congress approved the President's recommendation of the Yucca Mountain site for the development of a repository. Figure 2: The 1.7-Mile Tunnel Built for Scientific Studies near the Potential Repository Area: [See PDF for image] [End of figure] The Yucca Mountain project is currently focused on preparing an application to obtain a license from NRC to construct a repository. The required application information includes both repository design work and scientific analyses. DOE is engaged in necessary tasks such as compiling information and writing sections of the license application, and is conducting technical exchanges with NRC staff and addressing key technical issues identified by NRC to ensure that sufficient supporting information is provided. It also plans to further develop the design of the repository, including revised designs for the repository's surface facilities and canisters to hold the waste. DOE is also identifying and preparing potentially relevant documentary material that it is required to make available on NRC's Web-based information system, known as the Licensing Support Network. This is a critical step because DOE is required to certify that the documentary material has been identified and made electronically available no later than 6 months in advance of submitting the license application.[Footnote 4] In February 2005, DOE announced that it does not expect the repository to open until 2012 at the earliest, which is more than 14 years later than the 1998 goal specified by the Nuclear Waste Policy Act of 1982. More recently, the conference report for DOE's fiscal year 2006 appropriations observed that further significant schedule slippages for submitting a license application are likely. Further delays could arise from factors such as the time needed for EPA to establish revised radiation standards for Yucca Mountain and for DOE to revise its technical documents in response. Such delays could be costly because nuclear utilities, which pay for most of the disposal program through a fee on nuclear power, have sued DOE, seeking damages for not starting the removal of spent nuclear fuel from storage at commercial reactors by the 1998 deadline. Estimates of the potential damages vary widely, from DOE's estimate of about $5 billion to a nuclear industry's estimate of about $50 billion, but the cost for the damages will likely rise if there are further delays to opening the repository. Given these schedule slippages, Congress has considered other options for managing existing and future nuclear wastes, such as centralized interim storage at one or more DOE sites. The conference report for DOE's fiscal year 2006 appropriations directed DOE to develop a spent nuclear fuel recycling plan to reuse the fuel. However, according to the policy organization of the nuclear energy industry, no technological option contemplated will eliminate the need to ultimately dispose of nuclear waste in a geologic repository. In October 2005, the project's Acting Director issued a memorandum calling for the development of wide-ranging plans for the "new path forward," DOE's effort to address quality assurance and other challenges prior to applying for a license. To restore confidence in scientific documents that will support the license application, some of the plans will address the need to review and replace USGS work products, a requirement for USGS to certify its scientific work products, and establishing a lead national laboratory to assist the project. Other plans are focused on a new simplified design for the waste canisters and repository facilities, a design that is expected to improve the safety and operation of the repository by eliminating the need to directly handle and process the spent fuel at the repository. Further, this aggressive effort called for management changes, including a transition plan; more rigorous project management, including a new baseline schedule; rescoping existing contracts and developing new contracts; tracking project hiring actions; a financial plan; and new reporting indicators. After DOE submits the license application, NRC plans to take 90 days to examine the application for completeness to determine whether DOE has addressed all NRC requirements. One of the reviews for completeness will include an examination of DOE's documentation of the quality assurance program to assess whether it addresses all NRC criteria. These criteria include, among other things, organization, design control, document control, corrective actions, quality assurance records, and quality audits. If it deems any part of the application is incomplete, NRC may either reject the application or require that DOE furnish the necessary documentation before proceeding with the detailed technical review of the application. If it deems the application is complete, NRC will docket the application, indicating its readiness for a detailed technical review.[Footnote 5] Once the application is accepted and placed on the docket, NRC will conduct its 18-month technical review of the application to determine if the application meets all NRC requirements, including the soundness of scientific analyses and preliminary facility design, and NRC quality assurance criteria. If NRC discovers problems with the technical information used to support the application, it may conduct specific reviews, including inspections, to determine the extent and effect of the problem. Because the data, models, and software used in modeling repository performance are integral parts of this technical review, quality assurance plays a key role since it is the mechanism used to verify the accuracy of the information DOE presents in the application. NRC may conduct reviews, including inspections, of the quality assurance program if technical problems are identified that are attributable to quality problems. NRC will hold public hearings chaired by its Atomic Safety and Licensing Board to examine specific topics. After completing the proceedings, the board will forward its initial decision to the NRC commissioners for their review. Finally, within 3 to 4 years from the date that NRC dockets the application, NRC will make a decision to grant the construction authorization, reject the application, or grant the construction authorization with conditions.[Footnote 6] NRC will grant a construction authorization only if it concludes from its reviews that the repository would meet its reasonable expectation that the safety and health of workers and the public would be protected. DOE Has a Long History of Quality Assurance Problems at Yucca Mountain and Is Relying on Costly and Time-Consuming Measures to Correct Problems Before Submitting Its License Application for the Repository: DOE has repeatedly experienced quality assurance problems with its work on the Yucca Mountain project. In the late 1980s, DOE had been challenged to fix and develop adequate plans and procedures related to quality assurance. By the late 1990s, audits by GAO, DOE, and others identified recurring quality assurance problems with several aspects of key scientific data, models, and software. Currently, in preparing to submit the license application to NRC, DOE is relying on costly and time-consuming rework to resolve lingering quality assurance problems with the transparency and traceability of data and in project design and engineering documents uncovered during audits and after-the-fact evaluations. DOE Has Had Problems Implementing and Maintaining an Effective Quality Assurance Program: DOE has a long-standing history of attempting to address NRC concerns about its quality assurance program. Although NRC will have responsibility for regulating the construction, operation, and decommissioning (closure) phases of the project, its regulatory and oversight role does not begin until DOE submits a license application. As a result, NRC's role in the project has been limited to providing guidance to DOE to ensure an understanding of NRC regulations and that the years of scientific and technical work will not later be found inadequate for licensing purposes. Specifically, since 1984, NRC has agreed to point out problems it identifies with the quality assurance program so that DOE can take timely corrective action. Initially, this NRC guidance was mainly focused on ensuring that DOE had the necessary quality assurance organization, plans, and procedures. As we reported in 1988, NRC had reviewed DOE's quality assurance plans and procedures comprising the principal framework of its quality assurance program, and concluded that they were inadequate and did not meet NRC requirements.[Footnote 7] NRC also concluded that DOE's efforts to independently identify and resolve weaknesses in the plans and procedures were ineffective. After observing DOE quality assurance audits, NRC determined that the audits were ineffective for measuring whether quality assurance procedures were being effectively implemented. Further, NRC identified additional concerns, during the 1980s, related to DOE management and organizational deficiencies relating to the quality assurance program. Specifically, among other things, NRC found the following: * DOE had a small staff and relied heavily on contractors to provide quality assurance oversight. Based on its experience in regulating nuclear power plants, NRC found that these types of organizations frequently developed major quality-related problems. * DOE had indirect project control, with administrative and functional control over the project split between different offices. NRC found that such project control arrangements tend to have serious quality assurance-related problems because conflicts can arise between quality and other organizational goals, such as cost and schedule. * During a 1984 NRC visit to Nevada, DOE project participants had expressed the opinion that quality assurance is "unnecessary, burdensome, and an imposition." Further, in 1986, DOE issued a stop- work order to the USGS based on a determination that USGS staff did not appreciate the importance of quality assurance and that USGS work would not meet NRC expectations. NRC believed that organizational attitudes can indicate whether a project is likely to experience problems relating to quality assurance and found such examples troublesome. Finally, based in part on the information obtained from its oversight activities, NRC concluded, in 1989, that DOE and its key contractors had yet to develop and implement an acceptable quality assurance program. However, by March 1992, NRC came to the conclusion that DOE had made significant progress in improving its quality assurance program. NRC noted that DOE had addressed many of its concerns, specifically that, among other things, (1) all of the contractor organizations had developed and were in the process of implementing quality assurance programs that met NRC requirements, (2) quality assurance management positions had been filled with full-time DOE personnel with appropriate knowledge and experience, and (3) DOE had demonstrated that it is capable of evaluating and correcting deficiencies in the overall quality assurance program. Nevertheless, in October 1994, NRC found problems with quality assurance, particularly with the site contractor's ability to effectively implement corrective actions and DOE's ability to oversee the site contractor's quality assurance program. Recurring Issues with Project Data, Models, and Software Illustrate DOE's Difficulties Addressing Quality Assurance Problems: As DOE's quality assurance program matured, it resolved NRC concerns about its organization, plans, and procedures, and in the late 1990s began successfully detecting new quality assurance problems in three areas critical to the repository's successful performance: the adequacy of the data sources, the validity of scientific models, and the reliability of computer software developed at the site. These problems surfaced in 1998 when DOE began to run the initial version of its performance assessment model. Specifically, DOE was unable to ensure that critical project data had been collected and tracked back to the original sources. In addition, DOE did not have a standardized process for developing scientific models used to simulate a variety of geologic events or an effective process for ensuring that computer software used to support the scientific models would work properly. As required by DOE's quality assurance procedures, the department conducted a root cause analysis and issued a corrective action plan in 1999. After corrective actions were taken, DOE considered the issues resolved. However, in 2001, similar deficiencies associated with models and software resurfaced. DOE attributed the recurrence to ineffective procedures and corrective actions, improper implementation of quality procedures by line managers, and personnel who feared reprisal for expressing quality concerns. Recognizing the need to correct these recurring problems, DOE conducted a comprehensive root cause analysis that included reviews of numerous past self-assessments and independent program assessments, and identified weaknesses in management systems, quality processes, and organization roles and responsibilities. Following the analysis, in July 2002, DOE issued its Management Improvement Initiatives (Initiatives) that addressed quality problems with software and models. In addition, DOE added other corrective actions to address management weaknesses that it found in areas such as roles and responsibilities, quality assurance processes, written procedures, corrective action plans, and work environment. However, DOE continued to face difficulties in resolving quality assurance problems concerning the data, software, and modeling to be used in support of the licensing application: * Data management. As part of NRC's quality assurance requirements, data used to support conclusions about the safety and design of the repository must be either collected under a quality assurance program or subjected to prescribed testing procedures to ensure the data are accurate for their intended use. In addition, the data supporting these conclusions must also be traceable back to its original source. In 1998, DOE identified quality assurance problems with the quality and traceability of data--specifically that some data had not been properly collected or tested to ensure their accuracy and that data used to support scientific analysis could not be properly traced back to their source. DOE again found similar problems in April and September 2003, when a DOE audit revealed that some data sets did not have the documentation necessary to trace them back to their sources; the processes for data control and management were unsatisfactory; and faulty definitions were developed, which allowed unqualified data to be used. * Software management. DOE quality assurance procedures require that software used to support analysis and conclusions about the performance and safety of the repository be tested or created in such a way to ensure that it is reliable. From 1998 to 2003, multiple DOE audits found recurring quality assurance problems that could affect confidence in the adequacy of software codes. For example, in 2003, DOE auditors found problems related to software similar to those found previously in areas such as technical reviews, software classification, planning, design, and testing. Further, a team of industry professionals hired by DOE to assess quality assurance problems with software reported in February 2004 that these problems kept recurring because DOE did not assess the effectiveness of its corrective actions and did not adequately identify the root causes of the problems. * Model validation. Models are used to simulate natural and environmental conditions at Yucca Mountain, and to demonstrate the performance of the future repository over time. However, before models can be used to support the license application, DOE must demonstrate through a process called validation that the models are able to accurately predict geologic events. In 1998, a team of project personnel evaluated the models and determined that 87 percent did not comply with the validation requirements. In 2001, and again in 2003, DOE audits found that project personnel were not properly following procedures--specifically in the areas of model documentation, model validation, and checking and review. Further, the 2003 audit concluded that previous corrective actions designed to improve validation and reduce errors in model reports were not fully implemented. After many years of working to address these quality assurance problems with data, software, and models, DOE had mostly resolved these problems and closed the last of the associated condition reports by February 2005. DOE Is Now Relying on Costly and Time-Consuming Rework to Resolve Additional Problems: As DOE prepares to submit the Yucca Mountain project license application to NRC, it has relied on costly and time consuming rework to ensure that the documents supporting the application are accurate and complete. Specifically, DOE has relied on inspections and rework by DOE personnel to resolve quality assurance problems with the traceability and transparency of technical work products. These efforts to deal with quality problems at the end, rather than effectively ensuring that work organizations are producing quality products from the beginning, add to the project's cost and could potentially delay DOE's submission of the license application to NRC. In addition, DOE's efforts indicate that some corrective actions have been ineffective in resolving problems with the quality assurance process. Further, DOE is now detecting quality assurance problems in design and engineering work that are similar to the quality assurance problems it experienced with its scientific work in the late 1990s. Although DOE did not initiate its major effort to address these problems until 2004, the department and NRC for years had known of quality assurance problems with the traceability and transparency of technical work products called Analysis and Model Reports (AMR). AMRs are a key component of the license application, and contain the scientific analysis and modeling data demonstrating the safety and performance of the planned repository. Among other quality requirements, AMRs must be traceable back to their original source material and data, and must also be transparent in justifying and explaining their underlying assumptions, calculations, and conclusions. In 2003, based in part on these problems as well as DOE's long-standing problems with data, software, and modeling, NRC conducted an independent evaluation of three AMRs. The scope of the review was to determine if the AMRs met NRC requirements for being traceable, transparent, and technically appropriate for their use in the license application. NRC found significant problems.[Footnote 8] First, in some cases DOE was not transparent in explaining the basis on which it was reaching conclusions. For example, in some circumstances, DOE selected a single value from a range of data without sufficient justification. Other times, DOE did not explain how a range of experimental conditions were representative of repository conditions. Second, where DOE did sufficiently explain the basis for a conclusion, it did not always provide the necessary technical information, such as experimental data, analysis, or expert judgment, to trace the support for that explanation back to source materials. For example, DOE did not explain how information on one type of material provided an appropriate comparison for another material. Moreover, while DOE had identified similar problems in the past, the actions taken to correct them did not identify and resolve other deficiencies. NRC concluded that these findings suggested that other AMRs possibly had similar problems, and that if not resolved, such problems could delay NRC's review of the license application as it would need to conduct special inspections to resolve any issues it found with the quality of technical information. To address problems of traceability and transparency, DOE in the spring of 2004 initiated an effort called the Regulatory Integration Team (RIT) to perform a comprehensive inspection and rework of the AMRs to ensure they met NRC requirements and expectations.[Footnote 9] According to DOE officials, the RIT involved roughly 150 full-time personnel from DOE, USGS, and multiple national laboratories such as Sandia, Los Alamos, and Lawrence Livermore. First, the RIT screened all of the approximately 110 AMRs and prioritized its efforts on 89 that needed additional rework. Ten AMRs were determined to be acceptable, and 11 were canceled because they were no longer needed to support the license application. According to DOE officials, approximately 8 months later, the RIT project was completed at a cost of about $20 million, with a total of over 3,700 problems and issues addressed or corrected. In February 2005, in a letter to DOE, the site contractor stated that the RIT effort was successful and that the AMRs had been revised to improve traceability and transparency. Subsequently, however, additional problems with traceability and transparency have been identified, requiring further inspections and rework. For example, after the March 2005 discovery of e-mails from USGS employees written between May 1998 and March 2000 implying that employees had falsified documentation of their work to avoid quality assurance standards, DOE initiated a review of additional AMRs that were not included in the scope of the 2004 RIT review. The additional AMRs contained scientific work performed by the USGS employees and had been assumed by the RIT to meet NRC requirements for traceability and transparency. However, according to DOE officials, DOE's review determined that these AMRs did not meet NRC's standards, and additional rework was required. Further, similar problems were identified as the focus of the project shifted to the design and engineering work required for the license application. In February 2005, the site contractor determined that in addition to problems with AMRs, similar traceability and transparency problems existed in the design and engineering documents that comprise the Safety Analysis Report--the report necessary for demonstrating to NRC how the facilities and other components of the repository site will meet the project's health, safety, and environmental goals and objectives. In a root cause analysis of this problem, the site contractor noted that additional resources were needed to inspect and rework the documents to correct the problems. DOE Cannot Be Certain Its Efforts to Improve Quality Assurance Have Been Effective Because of Weaknesses in Tracking Progress and Identifying Problems: DOE cannot be certain that it has met continuous improvement goals for implementing its quality assurance requirements, a commitment DOE made at the closure of its Management Improvement Initiatives (Initiatives) in April 2004. At that time, DOE told us it expected that the progress achieved with the initiatives would continue and that its performance indicators would enable it to assess further progress and direct management attention as needed. However, DOE's performance indicators, as well as a second management tool--trend evaluation reports--have not been effective for this purpose. More specifically, the indicators panel did not highlight the areas of concern covered by the initiatives and had weaknesses in assessing progress because the indicators kept changing. The trend evaluation reports also did not focus on tracking the concerns covered by the Initiatives, had technical weaknesses for identifying significant and recurring problems, had inconsistently tracked progress in addressing problems, and could not fully analyze projectwide problems.[Footnote 10] In addition, the trend reports' tracking of problems for which corrective actions were already being taken was at times overly influenced by judgments about whether additional management action was warranted rather than the problems' significance. The Panel's Focus and Frequent Changes Hindered the Tracking of Progress with Management Concerns and Quality Problems: By the time that the actions called for by the Initiatives had been completed in April 2004, project management had already developed the indicators panel, which DOE refers to as the annunciator panel, to use at monthly management meetings to monitor project performance. The panel was a single page composed of colored blocks representing selected performance indicators and their rating or level of performance. A manager viewing the panel would be able to quickly see the color rating of each block or indicator. For example, red indicated degraded or adverse performance warranting significant management attention; yellow indicated performance warranting increased management attention or acceptable performance that could change for the worse; and green indicated good performance. The panel represented a hierarchy of indicators in which the highest level, or primary, indicators were shown; secondary indicators that determined the primary indicators' ratings were shown for some primary indicators; but lower third-or fourth-level indicators were not shown. Our review analyzed a subset of these indicators that DOE designated as the indicators that best predict performance in areas affecting quality. While we were conducting our review, DOE suspended preparation of the panel after August 2005 while it reconsiders its use of indicators to monitor project performance. DOE had also suspended preparation of the panel from late 2004 to early 2005 in order to make substantial revisions. These revisions were made, in part, to emphasize fewer, more important indicators for management attention. The Initiatives raised concerns about five key areas of management weakness as adversely affecting the implementation of quality assurance requirements: 1. Roles and responsibilities were becoming confused as the project transitioned from scientific studies to activities supporting licensing. The confusion over roles and responsibilities was undermining managers' accountability for results. The Initiatives' objective was to realign DOE's project organization to give a single point of responsibility for project functions, such as quality assurance and the Corrective Action Program, and hold the project contractor more accountable for performing the necessary work in accordance with quality, schedule, and cost requirements. 2. Product quality was sometimes being achieved through inspections by the project's Office of Quality Assurance rather than being routinely implemented by the project's work organizations. As a result, the Initiatives sought to increase work organizations' responsibility for being the principle means for achieving quality. 3. Work procedures were typically too burdensome and inefficient, which impeded work. The Initiatives sought to provide new user-friendly and effective procedures, when necessary, to allow routine compliance with safety and quality requirements. 4. Multiple corrective action programs existed, processes were burdensome and did not yield useful management reports, and corrective actions were not completed in a timely manner. The Initiatives sought to implement a single program to ensure that problems were identified, prioritized, and documented and that timely and effective corrective actions were taken to preclude recurrence of problems. 5. The importance of a safety-conscious work environment that fosters open communication about concerns was not understood by all managers and staff, and they had not been held accountable when inappropriately overemphasizing the work schedule, inadequately attending to work quality, and acting inconsistently in practicing the desired openness about concerns. Through issuing a work environment policy, providing training on the policy, and improving the Employee Concerns Program, the Initiatives sought to create an environment in which employees felt free to raise concerns without fear of reprisal and with confidence that issues would be addressed promptly and appropriately. As shown in table 1, the Initiatives' effectiveness indicators for tracking progress in addressing these management weaknesses did not have equivalent performance indicators visible in the annunciator panel when it was prepared for the last time, using August 2005 data. Table 1: Visibility of Management Improvement Initiatives' Effectiveness Indicators in Annunciator Panel When Last Prepared (Using August 2005 Data): Key area of management weakness identified in the Initiatives: Roles, responsibilities, accountability, authority; Effectiveness indicators from the Initiatives: An improving trend in quality and work schedule performance; DOE response on coverage of the management weakness in the panel's performance indicators: No integrated analysis of trends in quality and schedule performance; GAO comments and observations: No indicator was visible in, or underlies, the panel; Some indicators measured aspects of quality or schedule, but provided no integrated analysis of these trends. Effectiveness indicators from the Initiatives: A consistently decreasing trend in quality problems related to roles and responsibilities; DOE response on coverage of the management weakness in the panel's performance indicators: No aspect measured; GAO comments and observations: No indicator was visible in, or underlies, the panel. Key area of management weakness identified in the Initiatives: Quality assurance programs and processes; Effectiveness indicators from the Initiatives: The numbers of high- priority (significant) quality problems that are self-identified are at least 80 percent of all significant quality problems; DOE response on coverage of the management weakness in the panel's performance indicators: One indicator looked at work organizations' identification of problems, including less significant ones; GAO comments and observations: No indicator, focused only on significant problems, was visible in the panel; One fourth-level indicator tracked work organizations' identification of significant problems. Effectiveness indicators from the Initiatives: A decreasing trend in average time to resolve significant quality; problems and in number of delinquent corrective actions for significant quality problems; DOE response on coverage of the management weakness in the panel's performance indicators: A new timeliness measure has been developed.[A]; GAO comments and observations: No indicator was visible in panel; Aspect of fourth-level indicator tracked average time of resolution. Key area of management weakness identified in the Initiatives: Work procedures; Effectiveness indicators from the Initiatives: A decreasing number of quality problems related to ineffective; Procedures; DOE response on coverage of the management weakness in the panel's performance indicators: No aspect measured; GAO comments and observations: No indicator was visible in, or underlies, the panel. Effectiveness indicators from the Initiatives: A decreasing trend in time needed to revise procedures; DOE response on coverage of the management weakness in the panel's performance indicators: No aspect measured; GAO comments and observations: No indicator was visible in, or underlies, the panel. Effectiveness indicators from the Initiatives: A decreasing trend in average time of interim procedure changes; DOE response on coverage of the management weakness in the panel's performance indicators: No aspect measured; GAO comments and observations: No indicator was visible in, or underlies, the panel. Key area of management weakness identified in the Initiatives: Corrective Action Program; Effectiveness indicators from the Initiatives: A decreasing trend in number of repetitive quality problems; DOE response on coverage of the management weakness in the panel's performance indicators: No aspect measured; GAO comments and observations: No indicator was visible in, or underlies, the panel. Effectiveness indicators from the Initiatives: A decreasing trend in average time to resolve significant quality; problems; DOE response on coverage of the management weakness in the panel's performance indicators: A new timeliness measure has been developed.[A]; GAO comments and observations: No indicator was visible in panel; Aspect of fourth-level indicator tracked average time of resolution. Effectiveness indicators from the Initiatives: Less than 10 percent of quality problems are resolved late; DOE response on coverage of the management weakness in the panel's performance indicators: A new timeliness measure has been developed.[A]; GAO comments and observations: No indicator was visible in the panel; A third-level indicator tracked percentage of problems with timely resolution. Key area of management weakness identified in the Initiatives: Work environment; Effectiveness indicators from the Initiatives: A decreasing number of substantiated employee concerns for harassment, retaliation, intimidation, and discrimination; DOE response on coverage of the management weakness in the panel's performance indicators: Aspects of this issue are measured by work environment indicators; GAO comments and observations: No indicator was visible in panel; A third-level indicator measured this performance. Effectiveness indicators from the Initiatives: Evaluation of routine employee concerns in less than 30 days, or; 90 days for complex employee concerns involving harassment or intimidation; DOE response on coverage of the management weakness in the panel's performance indicators: Goals have remained at 30 and 90 days; GAO comments and observations: No indicator was visible in panel; Third- level indicators measured the timely completion of routine and other concerns. Effectiveness indicators from the Initiatives: External evaluation of work environment shows positive changes; DOE response on coverage of the management weakness in the panel's performance indicators: External evaluation is accomplished through independent employee surveys, reflected in third-level indicators; GAO comments and observations: No indicator was visible in panel; Four third-level indicators were based on the employee surveys. Source: GAO analysis of DOE data. [A] New timeliness indicator was not implemented by the time of the final panel using August 2005 data. [End of table] Two of the Initiatives' key areas of concern--(1) roles, responsibilities, authority, and accountability; and (2) work procedures--and their associated effectiveness indicators were not represented in the panel's visible or underlying indicators. The Initiatives' effectiveness indicator for tracking trends in recurring problems also was not represented. In other cases, the Initiatives' effectiveness indicators were represented in underlying lower-level indicators that had very little impact on the rating of the visible indicator. An example is the Initiatives' indicator for timely completion of employee concerns. The panel's related visible indicator was work environment, whose rating was based on 4 secondary and 23 tertiary indicators. Of the third-level indicators, two were for timeliness of completion of employee concerns, and combined they contributed 3 percent toward the rating of the work environment indicator. As a result of the weighting of these many underlying indicators, ratings for individual lower-level indicators could be different from the visible indicator. For example, in August 2005, the work environment indicator showed good performance. However, the ratings of four underlying indicators from the project's employee survey on the work environment--collectively accounting for 25 percent of the work environment indicator's score--indicated the need for increased management attention. Moreover, some of the Initiatives' indicators, such as the work organizations' self-identification of significant problems, had their impact on visible indicators diluted by the inclusion of other indicators that were not focused solely on the detection of significant problems. Another shortcoming of the annunciator panel was that frequent changes to the indicators hindered the ability to identify problems for management attention and track progress in resolving them. The indicators could change in many ways, such as changes in their definition, calculation, or data sources used in calculations, or from the deletion or addition of a subindicator. When such changes were made to the indicators, progress became less clear because changes in reported performance levels may have been the result of the indicator changes rather than actual performance changes. Some of the indicators for key project processes with quality elements changed from one to five times during the 8-month period from April 2004 through November 2004. Even after the major revision of the panel in early 2005, most of the performance indicators tracking quality issues continued to change over the next 6 months--that is, from March 2005 through August 2005. As shown in table 2, only one of the five relevant indicators did not change during this period. One indicator was changed four times during the 6-month period, resulting in it being different in more months than it remained the same. Table 2: Key Indicators for Processes with Quality Elements, Their Intended Focus, and Number of Times They Changed (March through August 2005): Indicators: Performance improvement; Intended focus: Effectiveness of self-assessment of quality and other issues, lessons learned, and Corrective Action Program; Number of months changed: 4. Indicators: Work management; Intended focus: Quality of work products and documents; Number of months changed: 1. Indicators: Safety-conscious work environment; Intended focus: Worker confidence in management support for raising quality and other concerns without fear of retaliation; management effectiveness in detecting and preventing retaliation for raising concerns; effectiveness of normal and alternative problem resolution; Number of months changed: 0. Indicators: Human performance; Intended focus: Preventing, detecting, and correcting human errors; Number of months changed: 3. Indicators: Quality performance; Intended focus: Composite of quality indicators, in areas of engineering products, self-assessment, Corrective Action Program, and work products and documents; Number of months changed: 1. Source: GAO analysis of DOE data. [End of table] Moreover, the panel was not always available to identify problems and track progress. The panel was not created for December 2004, January 2005, and February 2005 because it was undergoing a major revision. At that time, DOE told NRC that the performance indicators for the panel were revised to reflect the change in the work as the project moved into the engineering, procurement, and construction phase. DOE also reduced the total number of visible indicators from 60 to 30 to focus on fewer, more critical aspects of project management. Panels with the new indicators were then produced for 6 months, starting with March 2005 and ending after August 2005. This second interruption of the panels resulted from another major revision to the indicators; this time, indicators are being made congruent with project work as designated by DOE's "new path forward," again to focus on fewer, more important activities. In December 2005, a senior DOE official told us that the project would begin to measure key activities, but without use of the panel. Trend Evaluation Reports Have Not Specifically Tracked the Initiatives' Management Concerns and Have Had Weaknesses Tracking Significant and Recurrent Problems for Management Attention: According to DOE, some of the Initiatives' areas of concern and their associated effectiveness indicators--for example, trends in quality problems related to roles and responsibilities--were being captured, at least partially, in the project's quarterly trend evaluation reports rather than in the performance indicators. However, the trend reports are a management tool designed more to identify emerging and unanticipated problems than to monitor progress with already identified problems, such as those addressed by the Initiatives. In developing these reports, trend analysts seek to identify patterns and trends in condition reports (CR), which document problematic conditions through the project's Corrective Action Program. The trend reports analyze CRs for more significant problems (Levels A and B) and minor problems (Level C), but not at Level D (opportunities for improvement). The trend analysis typically separates the reported problems into categories such as organizational unit, type of problem, and cause. These categories are intended to provide insights into the problems. For example, analysis might reveal that most occurrences of a particular type of problem are associated with a certain organization. In practice, DOE missed opportunities to use trend reports to call attention to progress in the Initiatives' areas of concern. For example, the Initiatives sought to clarify roles and responsibilities within and between DOE and BSC to ensure clear accountability for project results during the project's transition from scientific studies to the design and engineering activities necessary to license a repository. Similar organizational transition problems were identified in the November 2004 trend report. While that report attributed increases in the number of causal factors associated with change management, supervisory methods, and work organization to recent BSC reorganizations and changes in the project from science-based to design and engineering activities, it did not specifically mention issues of roles and responsibilities or that roles and responsibilities was an Initiatives' area of concern. However, an analysis of the cause of the problems noted in various significant condition reports, which is performed for certain condition reports and outside of the process of developing trend reports, found evidence of weaknesses in the organizational interfaces among BSC organizations, as well as between BSC and DOE. According to this cause analysis, these organizational interface weaknesses were associated with some manner of change and represented weaknesses in the definition of roles and responsibilities. Trend reports are generally based on condition reports, and problems with roles and responsibilities seem to be identified in cause analyses rather than in the condition reports themselves. Similarly, DOE missed an opportunity to use trend reports to discuss the Initiatives' goal that the project's line or work organizations become more accountable for self-identifying significant problems. The August 2005 trend report briefly cited an evaluation of a CR highlighting the low rate of self-identification of significant problems during the previous quarter and reported the evaluation's conclusion that it was not a problem warranting management attention. However, the trend report did not mention that about 35 percent of significant problems were self-identified during the previous quarter, while the Initiatives' goal was that 80 percent of significant problems would be self-identified. Thus, the trend report missed an opportunity to either raise a performance problem or pose the question of whether the Initiatives' goal needed to be reassessed. Beyond whether they effectively tracked the Initiatives' areas of concern, trend reports face important obstacles, in general, to adequately identify recurrent and significant problems: * Recurring or similar conditions can be difficult to clearly identify for management's attention and resolution. A trend report noted that there will be few cases where recurrent conditions are obvious because each condition slightly differs. * Trend analysis tends to focus on the number of CRs issued, but the number of CRs does not necessarily reflect the significance of a problem. For example, the number of CRs involving requirements management decreased by over half from the first quarter to the second quarter of fiscal year 2005. However, this decrease was not a clear sign of progress. Not only did the number rise again in the third quarter, but the May 2005 trend report also noted that the number of all condition reports had dropped during the second quarter. According to the report, the volume of CRs in the first quarter had been high because of reviews of various areas, including requirements management. Another example is the records management problem. The November 2005 trend report stated that a records management problem identified in various CRs, despite accounting for about 50 percent of all business administration problems, reflected an underlying error rate of less than 1 percent and thus was not a significant problem. * The lack of an increasing trend in the number of reported problems does not necessarily mean the lack of a significant problem for management attention. Knowing the appropriate level of performance, regardless of the trend, is difficult without having clearly appropriate benchmarks from organizations engaged in activities similar to the Yucca Mountain project. Such benchmarks would clarify, for example, whether a project's percentages of human performance errors compare favorably, regardless of whether the numbers are increasing. Similarly, the trend in the number and types of CRs during any period is not necessarily a sign of improvement or worsening conditions. Trends can be attributed to various factors, including increases in the number of audits or self-assessments, which can lead to more CRs being issued. * At the time of analysis, some trend data may not be sufficiently reliable or complete to ensure sound findings for management's attention. For example, although some actions were taken in December 2004 to ensure that cause and other codes were properly assigned, a BSC audit in June 2005 again raised questions about the consistency of the coding. With respect to completeness, the fourth quarter report for 2005 noted that 28 percent of the Level B CRs did not have a cause code at the time of the trend analysis, and one finding was presented even though two-thirds of the data was missing. Due, in part, to these obstacles and changes to how the analysis is done, trend reports have not consistently determined the significance of problems or performed well in tracking progress in resolving problems. For example, trend reports have questionably identified significant human performance problems and ineffectively tracked progress in resolving the problem because of no clearly appropriate or precise benchmark for performance, inconsistent focus on the problem, and unreliable data on cause codes. The February 2004 trend report identified a human performance problem based on Yucca Mountain project data showing the project's proportion of skill-based errors to all human performance errors was two times higher than benchmark data from the Institute of Nuclear Power Operations (INPO).[Footnote 11] The report used this comparison to suggest that the project needed to adopt successful commercial nuclear practices for addressing skill-based errors. However, the report cautioned that other comparisons with these INPO data may not be appropriate because of differences in the nature, complexity, and scope of work performed, but did not explain why the report's comparison of INPO data for skill-based errors to the Yucca Mountain project should be an exception to this caution. The May 2004 trend report repeated this comparison to INPO, finding skill-based errors three times higher than the benchmark data. However, this INPO benchmark has not been used in subsequent reports. The November 2004 trend report redefined the problem as the predominance of human performance errors in general, rather than the skill-based component of these errors--but later reports reinterpreted this predominance as not a problem. The problem with skill-based errors was unclear in the November 2004 report because these errors were showing a decreasing trend, a finding that was attributed as likely the result of unreliable assignment of cause codes. Instead, the report cited an adverse trend based on the fact that the human performance cause category accounted for over half of the total number of causes for condition reports prepared during the quarter. Under the project's trend analysis guidelines, this large predominance of human performance causes--in contrast to management, communication or procedure, and other cause categories--was designated an adverse trend. Nevertheless, by February 2005, trend reports began interpreting this predominance as generally appropriate, given the type of work done by the project. That is, the project's work involves mainly human efforts and little equipment, while work at nuclear power plants involves more opportunities for errors caused by equipment. In our view, this interpretation that a predominance of human performance errors would be expected implies an imprecise benchmark for appropriate performance. Although trend reports continued to draw conclusions about human performance problems, the February 2005 report indicated that any conclusions were hard to justify because of data reliability problems with cause coding. For example, the majority of problems attributed to human performance causes are minor, or Level C, problems that receive less rigorous cause analysis, such as not completing a form. This less rigorous analysis tends to reveal only individual human errors--that is, human performance problems--whereas more rigorous analysis tends to reveal less immediately obvious problems with management and procedures. Trend reports have also inconsistently tracked progress in resolving the problem associated with the "flow-down" of requirements into the project's procedures--that is, with ensuring that program, regulatory, and statutory requirements are identified, allocated, and assigned to the project organizations that are responsible for applicable activities. Such requirements management problems can result in inadequate control over design inputs and, possibly, inputs to scientific models. Progress with this problem was less clear because of inconsistent methods of categorizing requirements management problems over time. Initially, based on reviews of annual trends in condition reports, the September 2004 and November 2004 trend reports observed a systemic and continuing problem in the flow-down of requirements from BSC's Project Requirements Document and identified this as an adverse trend. In subsequent reports, the requirements flow-down problem was variously treated as an aspect of requirements management or records management, or as a latent management weakness or weak change management. When treated as an aspect of these broader problems, the significance of the original flow-down problem and any progress in resolving it became diluted and less clear. The primary focus eventually became requirements management, which the February 2005 trend report designated as a potential trend, whereas the flow-down problem had earlier been designated an adverse trend. Consequently, as a result of this change, the flow-down of requirements got less direct attention and analysis--for example, receiving only a footnote in the August 2005 trend report stating that the April 2004 condition report issued to address the adverse trend was still overseeing implementation of corrective actions. In addition, because trend reports examine only condition reports issued to BSC, they do not always assess the projectwide significance of problems such as requirements management.[Footnote 12] When analyzing one category of issues associated with requirements management, the November 2005 report stated that BSC and DOE shared the process problems, which cannot be adequately addressed by just one of the organizations. However, for a second category of these issues, the report did not analyze most of the condition reports because 6 of the 10 relevant reports were assigned to DOE. For a third category of issues, no analysis or recommendation was provided because all of the reports were assigned to DOE and therefore did not fall within the scope of the trend report. DOE Has Not Adequately Tracked Problems Being Addressed by Ongoing Management Actions: The tracking of problems for which corrective actions are already being taken appeared at times to be overly influenced by judgments, rather than the problems' significance, about whether additional management action is warranted. As a result, problems might be rated as less significant, or not tracked further. The situation of assigning a lower rating to a problem's significance was apparently caused by the fact that ratings were simultaneously an assessment of a problem's significance and of the need for management action. In its current formulation, DOE's rating categories cannot accurately represent both the assessment of a problem's significance and a judgment that additional actions are not needed because the designated rating category will distort one or the other. For instance, the November 2005 trend report analyzed the four categories of requirements management issues and designated one category that included problems with requirements flow-down as a "monitoring trend"- -defined as a small perturbation in numbers that does not warrant action but needs to be monitored closely. Describing this trend as a small perturbation, or a disturbance in numbers, did not accurately describe the report's simultaneous recognition that significant process problems spanned both BSC and DOE and the fact that the numbers and types of problems were consistently identified over the previous three quarters. A more understandable explanation for the low rating is that designating the problem at any higher level of significance would have triggered guidelines involving the issuance of a condition report, which, according to the judgment expressed in the report, was not needed. Specifically, the report indicated that existing condition reports have already identified and were evaluating and resolving the problem, thereby eliminating the need to issue a new condition report. By rating the problem at the lowest level of significance and not calling for additional actions, the trend report did not sufficiently draw management's attention to the problem. The trend report's assessment did not convey that other serious problems might have been raised by the additional condition reports. At about the same time that the trend report judged that no new condition reports were necessary, an Employee Concerns Program's investigation of requirements management resulted in 14 new condition reports--3 at the highest level of significance and 8 at the second-highest level of significance.[Footnote 13] For example, the Employee Concerns Program's investigation resulted in condition reports calling for an analysis of the collective significance of the numerous existing condition reports and an assessment of whether the quality assurance requirement for complete and prompt remedial action had been met.[Footnote 14] As a result of the investigation and a concurrent DOE root cause analysis, during the December 2005 Quarterly Management Meeting with NRC, DOE stated that strong actions were required to address the problems with its requirements management system and any resulting uncertainty about the adequacy of its design products.[Footnote 15] Trend reports identified significant problems in the February 2005 report but did not continue to track the problems after a separate analysis identified ongoing improvement actions. According to the trend report, Level B condition reports collectively indicated organizational weaknesses associated with change management involving cross- departmental interfaces. The trend report recommended that management focus on these problems, and cited a condition report that would further investigate them. The cause analysis for that condition report and a related condition report found that the problems were well-known, in part through a BSC review, and related to a variety of ongoing BSC improvement actions. Since this was a broad category of problems with many initiatives under way, the cause analysis recommended no new actions other than for management to remain aware of the problems. However, the trend reports that followed provided no further analyses to focus management's awareness on these problems or to assess progress in resolving them. DOE's 'New Path Forward' for Preparing to Submit Its License Application Faces Substantial Quality Assurance and Other Challenges: In October 2005, DOE announced an aggressive series of proposed changes to the design, organization, and management of the Yucca Mountain project, but this effort--known as the "new path forward"--will face substantial challenges. Some key challenges facing DOE are (1) determining the extent of problems and restoring confidence in the documents supporting the license application after the discovery of e- mails raising the potential of falsified records, (2) settling design issues and associated problems with requirements management, and (3) replacing key personnel and managing the transition of new managers and other organizational challenges. The current Acting Director of the Office of Civilian Radioactive Waste Management (OCRWM) stated that DOE will not announce a schedule for submitting a license application until DOE addresses these important quality assurance and other challenges. Since DOE is still formulating its plans, it is too early to determine whether the new path will resolve these challenges. Determining the Extent of Problems with Relevant Documents Will Delay DOE's Submission of the License Application: In March 2005, after announcing the discovery of USGS e-mails suggesting the possible violation of quality assurance requirements, including the falsification of records, DOE has taken steps to address lingering concerns about the adequacy of the scientific work related to the flow of water into the repository and whether similar quality assurance problems are evident in other e-mails relevant to the licensing application. Specifically, DOE is (1) conducting an extensive review of approximately 14 million e-mails to determine whether these e- mails raise additional quality assurance concerns and whether they might be relevant to the licensing process, and (2) reworking the technical documents created by USGS personnel to ensure that the science underlying the conclusions on water infiltration are correct and supportable in the license application. The Acting Director of OCRWM has stated that DOE will not submit a license application until these efforts are complete. Consequently, given the early planning stage of these efforts, it is unknown how long this will delay the submission of a license application. As part of the licensing process, DOE is required to publicly disclose all documents relevant to the licensing application, including e-mails, by posting them on DOE's public Web site, which is accessible through the NRC-sponsored, Internet-based Licensing Support Network (LSN). To satisfy schedule requirements, DOE must certify that relevant documents have been posted to the network and made available for public review 6 months before the submission of the license application. In preparation for submitting the license application by December 2004, in June of that year, DOE submitted almost 700,000 e-mails to the LSN that had been reviewed by their original authors and determined to be relevant to the licensing process. They were part of a group of approximately 6 million archived e-mails authored by individuals still associated with the project. However, in August 2004, NRC's Atomic Safety and Licensing Board ruled that DOE had not met its regulatory obligation to make all relevant documentary material available. Specifically, DOE had not reviewed a group of approximately 4 million archived e-mails authored by individuals no longer affiliated with the project to determine whether the e-mails were relevant to the licensing process. As part of its effort to address the board's ruling, BSC began a review of e-mails authored by employees who were not currently working at the project. During this review, the contractor discovered and brought forward e- mails between USGS scientists working on water infiltration models that raised questions of the potential falsification of technical information in order to sidestep quality assurance requirements. Following the discovery of the e-mails, DOE conducted a search to determine if there were similar e-mails in the approximately 1 million e-mails previously determined relevant for licensing. However, the DOE Inspector General reported in November 2005[Footnote 16] that there was no evidence that the project requirements for identifying and addressing conditions adverse to quality, such as those contained in the USGS e-mails, were considered during the initial review of e-mails. Further, among the approximately 10 million e-mails that had already been reviewed for the licensing process, they found additional e-mails that identified possible conditions adverse to quality that had not been identified by project personnel as requiring further review. The DOE Inspector General recommended, among other things, that DOE (1) expand the review of archived e-mails to include both those deemed relevant and those deemed not relevant to the licensing process, and ensure that conditions adverse to quality are appropriately identified, investigated, reported, and resolved; and (2) ensure that current and future e-mails are reviewed for possible conditions adverse to quality and that such conditions are appropriately addressed under the Corrective Action Program (CAP) system. DOE accepted the Inspector General's recommendations. Specifically, DOE agreed to develop a corrective action plan to expand the review of archived e-mails to ensure that conditions adverse to quality are appropriately identified and processed under the CAP system. In addition to this review, the DOE Inspector General opened a criminal investigation into the USGS e-mails in March 2005. As of December 2005, the investigation was still in progress. According to NRC on-site representatives, completing these e-mail reviews will be challenging because DOE now has to screen millions of e- mails to ensure that records were not falsified. Further, many of these e-mails were written by employees who no longer work at the project or may be deceased, making it difficult to learn their true meaning and context. Moreover, if additional e-mails are found that raise quality assurance concerns, DOE may have to initiate further review, inspections, or rework to address the newfound problems. NRC officials stated that it takes the issue of potentially falsified documents by USGS employees very seriously, wants a full understanding of the situation regarding the USGS e-mails, and will conduct follow-up in this area. Because NRC wants DOE to submit a high-quality license application, it has encouraged DOE to take the time and actions necessary to fully and adequately resolve these and other quality assurance issues. Immediately following the discovery of the USGS e-mails, DOE undertook a scientific investigation into the technical documents created by USGS personnel. In October 2005, DOE began developing an action plan for reviewing, validating, augmenting, and replacing USGS work products that had come under scrutiny. Although the plan is not yet complete, the Acting Director told us that the license application would not be submitted until the USGS work is replaced and there is confidence that all requirements have been met. In an effort to ensure that the scientific work underlying water infiltration modeling is accurate, DOE is working to corroborate the original work by engaging multiple agencies and organizations to rework the models. For example, DOE has (1) had its lead project contractor work with the Idaho National Laboratories to extensively review the software and data used in the original science work, (2) engaged Sandia National Laboratories to rework the model and calculations using different software than was used originally, and (3) also asked USGS to rework the models. Consequently, when this additional rework is completed, DOE will have four sets of analysis (including the original scientific work) with which they can evaluate, compare, and corroborate results. DOE will then pick one set of scientific analysis for inclusion in the license application, and work to explain and defend its choice. Ongoing Design and Requirements Management Issues Could Delay DOE's Submission of the License Application: In October 2005, DOE announced significant changes to the design of the Yucca Mountain repository to simplify the project and improve its safety and operation. However, these changes will also require additional design and engineering work that will add uncertainty about the timing of the submission of a license application. DOE had been considering a design where radioactive waste would be shipped to the Yucca Mountain site, removed from its shipping container, placed and sealed in a special disposal container, and finally moved into the underground repository. As a result, DOE contemplated handling the waste up to four separate times. In late 2003, DOE engineers began identifying potential safety problems with this approach. First, possible fissures or holes in the cladding surrounding the spent nuclear fuel accidentally caused during the handling of the waste could cause air to mix with the fuel and oxidize. Consequently, this radioactive oxidized material could then leak and be dispersed into the air. Second, DOE engineers determined that the original facility design would not be able to adequately control the levels of radioactivity in the buildings where the waste would be repackaged before being moved in the repository. To address these problems, DOE researched a series of options, including only accepting radioactive waste that had already decayed to the point where oxidization would not be problematic, and testing the waste shipments for oxidization and treating them at another site before they arrived at the repository. In addition, DOE also considered changing the design by filling the processing buildings with inert gas to prevent oxidization and revising the electrical and ventilation systems. According to a DOE official, these options were impractical or added complexity to the design. However, in October 2005, DOE proposed a new design that relies on uniform canisters that would be filled and sealed before being shipped, eliminating the need for direct handling of the waste prior to being placed in the repository. As a result, DOE will not have to construct several extremely large buildings costing millions of dollars for handling radioactive waste. DOE believes this change will improve the safety, operation, and long-term performance of the repository. However, this change will also pose a challenge to the project because of the widespread implications and the unknown time and effort required to implement it. For example, to implement the new design, DOE will need to, among other things, * get approval from the Energy Systems Acquisition Advisory Board[Footnote 17] for a new project plan, which, among other things, includes details on the conceptual design, cost estimates, risk management efforts, and acquisition strategies; * plan, design, and produce standardized canisters for the transportation of waste; * coordinate this new approach with commercial nuclear power plants, NRC, and government organizations that plan on shipping waste to the project; and: * revise procurement and contracting plans to support the new design. Finally, DOE will need to perform the detailed design and engineering work required to implement the new design, and create new technical documents to support the license application. However, before it can present its new plans and perform this design and engineering work, DOE officials have stated that it will need to resolve long-standing quality assurance problems involving requirements management. Requirements management is the process that ensures the broad plans and regulatory requirements affecting the project are tracked and incorporated into specific engineering details. According to DOE's root cause analyses, low-level documents were appropriately updated and revised to reflect high-level design changes through fiscal year 1995. However, from 1995 through 2002, many of these design documents were not adequately maintained and updated to reflect current designs and requirements. Further, a document that is a major component of the project's requirements management process was revised in July 2002, but has never been finalized or approved. Instead, the project envisioned a transition to a new requirements management system after the planned submission of the license application in December 2004. However, for various reasons, the license application was not submitted at that time, and the transition to a new requirements management system was never implemented. As a result, the document refers to the out-of-date NRC regulations contained in 10 CFR part 60, and not the regulations in 10 CFR part 63 that were finalized in October 2002. The scope and cause of requirements management problems have been identified in multiple DOE and NRC reviews. * Multiple condition reports issued in 2004 and 2005 have identified problems with requirements management. Due to these condition reports and NRC concerns that repetitive deficiencies and the failure to implement timely corrective actions could have direct implications on the quality of the planned license application, NRC performed a review of Corrective Action Program documents related to the requirements management program in the late summer of 2005. NRC determined that these reports identified approximately 35 deficiencies related to requirements management. Because the requirements management documents are not current and the new requirements management system has not been implemented, NRC concluded that there does not appear to be a requirements management mechanism in place. Further, based on the number of reports and other issues identified by DOE audits, NRC concluded that the project's Corrective Action Program was not effective in, among other things, eliminating the repeated identification of deficiencies relating to requirements management or initiating the actions to identify and appropriately address the root cause of these problems. * In September 2005, DOE began reviewing the root causes associated with CR-6278, a condition report identifying problems with requirements management. As part of the review, DOE personnel analyzed 135 condition reports and other events and allegations. Among other things, this review found that DOE expectations for requirements management were diluted and eventually neglected, that DOE reduced funding for requirements management due to reductions in its annual budget, and that these and other events caused the requirements management process to become "completely dysfunctional" from July 2002 to the time of the review in the fall of 2005. The analysis identified the root causes of these conditions as DOE's failure to fund, maintain, and rigidly apply a requirements management system. * In November 2005, a team of DOE personnel concluded an investigation into an employee's concerns regarding requirements management. The team substantiated all of the concerns they investigated and found instances of failures and breakdowns in the requirements management process. For example, among other things, the team found that no procedure was developed to describe how requirements management was to occur; some existing requirements management procedures were not implemented; and project management was aware of these conditions but corrective actions were deferred because the planned requirements management system was expected to address the problem. As a contributing factor, the team also observed that the project's lead contractor had not implemented a "traditional systems engineering approach" as it did not have, among other things, typical engineering management plans or a separate systems engineering organization responsible for requirements management. As a result of the investigation, the team initiated 14 condition reports, 13 of which identified quality-related problems. To address these problems, on December 19, 2005, DOE issued a stop-work order on design and engineering for the surface facility and certain other technical work. DOE stated that the root cause analysis for CR 6278 and the investigation into employee concerns revealed that the project has not maintained or properly implemented its requirements management system, resulting in inadequacies in the design control process. This stop-work order will be in effect until, among other things, the project's lead contractor improves the requirements management system; validates that processes exist and are being followed; and requirements are appropriately traced to implementing mechanisms and products. Further, DOE will establish a team to take other actions necessary to prevent inadequacies in requirements management and other management systems from recurring. An example of the potential risks of a breakdown with requirements management was noted during a BSC audit on the design process in March 2005. NRC on-site representatives observing this audit reported that the audit team noted problems with inconsistencies between the design documents of the planned fuel-handing facility that would be receiving, preparing, and packaging the waste before it is placed in the repository. The original set of requirements specified that no water from a fire protection system was to be used in the fuel-handling areas of the facility because under certain scenarios, water used for fire suppression could facilitate an accidental nuclear reaction, a condition known as criticality. Later, as the project began to review the design of the fuel-handling facility, the design was changed to allow the use of water sprinklers in the fuel-handling areas of the facility to suppress possible fires. NRC noted that personnel working on the design knew of the inconsistencies between older and newer design documents, but no formal tracking mechanism had been provided to ensure that those issues were rectified. According to an NRC on-site representative in December 2005, this was an example of a concern with requirements management, and that repetitive and uncorrected issues associated with the requirements management process could have direct implications on the quality of the license application. While the project may be able to resolve these inconsistencies through an informal process, the lack of a formal design control and requirements management process increases the risk that not all such problems will be addressed. These requirements management problems are potentially significant because if the high-level engineering needs of the project are not accurately or completely reflected in the detailed design, then the quality of the license application may be compromised and cause delays in the license application review process. For example, according to a 1989 speech prepared by NRC's Office of General Counsel stressing the importance of quality assurance, a West Coast nuclear power plant experienced similar quality assurance problems with requirements management. After a license was issued by NRC, power plant personnel discovered that the wrong diagrams were used to develop design requirements. As a result of this and other quality assurance weaknesses identified by NRC, the license was suspended and the power plant was required to initiate an independent program to verify the correctness of the design. Further, NRC reopened hearings on the issue of the adequacy of the power plant's quality assurance program related to the plant's design. DOE Faces Challenges in Managing the Transition, Complexity, and Continuity of Its 'New Path Forward': In October 2005, DOE announced a "new path forward" that would create a new project schedule and financial plan to address the completion of scientific and engineering work in support of a license application. However, DOE faces challenges to successfully implementing the new path, in terms of managing the transition, program and organizational complexities, and the continuity of management. According to DOE managers involved with planning the new path forward, the organizational transition could take several months to complete. It is too early to determine whether DOE's new effort will resolve quality assurance issues and move the project forward to the submission of a license application. Accountability for quality and results, which was identified as a significant transition issue in the Initiatives, will likely pose a challenge for managing the transition to the new path forward. The Initiatives sought to clarify roles and responsibilities within and between DOE and contractor organizations to ensure clear accountability for results and quality during the transition from OCRWM's organization, processes, procedures, and skills supporting scientific studies to those supporting the activities necessary to license a repository. As the project realigns organizations, processes, procedures, and skills to support the new path forward, it will also be faced with the challenge of ensuring that accountability is not undermined during the transition. For instance, according one DOE manager, transitioning project work to a lead laboratory under a direct contract with DOE could pose a significant challenge for quality assurance because the laboratories are currently working under BSC quality assurance procedures and will now have to develop their own procedures. Implicitly recognizing the importance of accountability issues, elements of the new path forward seek to address issues that can negatively affect quality assurance and project management in general. For instance, the new path includes plans for developing and transmitting requirements to USGS for the certification of scientific work. In addition, a senior project official told us that the lead laboratory would provide a single point of accountability that will enhance the quality of the science work. The Acting Director indicated that OCRWM's management structure may have to be reorganized to have a single manager clearly accountable for each of the new path's major tasks in science, engineering, and licensing. Moreover, the project is developing new performance indicators to allow the project to assess important activities under the new path forward. Outside of the new path, as the result of a September 2005 DOE Inspector General report[Footnote 18] on accountability problems with managing contract incentives, OCRWM agreed to develop a comprehensive corrective action plan to provide clearer and more objective performance standards in the BSC contract. Program complexity and other project characteristics are also likely to pose challenges to managing quality assurance. Based on its experience with licensing and regulating nuclear power plants, NRC observed in the mid-1980s that the Yucca Mountain project's characteristics, such as a large and complicated program, increased the likelihood of major quality-related problems. Although the new path is intended to simplify design, licensing, and construction, the project remains a complicated program that seeks to both restore confidence in its scientific studies and pursue new design and engineering activities. As a result, the project has to manage quality assurance issues simultaneously in both areas. Moreover, the project involves a complicated organizational structure. The project will continue contracting work with BSC, USGS, and the Sandia National Laboratory, which involves working with organizations in various locations. In our 1988 report, we noted that the geographic distance between the various organizations may hamper OCRWM's quality assurance communication and oversight objectives. The project also faces management challenges related to ensuring management continuity at the project, since DOE has experienced turnover in 9 of 17 key management positions since 2001. To ensure the right managers move the project forward to licensing, the project has a recruitment effort for replacing key departing managers. In the past year, the project has lost key managers through the departures of the director of Project Management and Engineering, the director of the License Application and Strategy, the director of Quality Assurance, and the contractor's general manager. According to NRC on-site representatives in August and October 2005, management turnover is a concern for NRC because it would like to see continuity of qualified managers rather than a series of acting managers. Recruiting replacement managers can impact project continuity, and newly acting managers may not take full rein of project tasks. However, the Acting Director told us that the recruitment process is an opportunity to improve project managers and staff, but recruiting the right people is challenging for various reasons--for example, government salaries are less than those in industry, and employment clauses restrict subsequent employment in related industries. Finally, since new directors sometimes give new direction to the project, a critical issue for sustaining the current new path forward is continuity with OCRWM's director. This position was occupied by three individuals between late 1999 and early 2005. The last OCRWM director assumed the position in April 2002, started the Management Improvement Initiatives in 2002, and left the position in February 2005. The current Acting Director began functioning in his position in the summer of 2005, and initiated the new path forward in October 2005. DOE is currently awaiting congressional confirmation of a nominee to take the director position. However, the Acting Director told us he expects that the new path forward will be sustained because it has been endorsed by the Secretary of Energy. Conclusions: DOE's Yucca Mountain project has been wrestling with quality assurance problems for a long time. Now, after more than 20 years of project work, DOE is again faced with substantial quality assurance and other challenges to submit a fully defensible license application to NRC. Unless these challenges are effectively addressed, further delays on the project are likely. Furthermore, even as DOE faces new quality assurance challenges, it cannot be certain it has resolved past problems, largely because the department has not been well served by management tools--specifically, its performance indicators and trend evaluation reports--that have not effectively identified and tracked progress on significant and recurring problems. First, the management tools have provided limited coverage of the areas of concern identified in the Management Improvement Initiatives and thus have not enabled DOE managers to effectively monitor progress in these important areas. Second, the tools have often not reflected the full extent or significance of problems because their scope has been limited and not based on projectwide analysis. Third, the trend evaluation reports have, at times, not accurately characterized problems because reliable and complete data and appropriate performance benchmarks were not available at the time of analysis. Fourth, frequent changes in performance indicators and the way analysis is done have made it difficult to accurately identify trends over time. Fifth, the tools' rating categories have sometimes been misleading as to the significance of problems because the ratings tend to be skewed by the fact that corrective actions were already being taken, without considering their effectiveness or considering the significance of the problem on its own terms. These shortcomings with the tools limit project managers' ability to direct and oversee such a large and complex undertaking as constructing an underground repository for nuclear wastes. Further complicating DOE's ability to manage the project are the vacancies in key managerial positions for the quality assurance program and elsewhere on the project. The tools become even more important for new managers who need to quickly understand project management issues. Recommendations for Executive Action: To improve the effectiveness of DOE's efforts to monitor performance in key areas at the Yucca Mountain project, including quality assurance, we recommend that the Secretary of Energy direct the Director, Office of Civilian Radioactive Waste Management, to take the following five actions to strengthen the project's management tools: * Reassess the coverage that the management tools provide for the areas of concern identified in the Management Improvement Initiatives and ensure that performance in these important areas is effectively monitored, especially in light of the more recent condition reports and associated cause analyses, trend reports, and other reviews indicating continuing problems. * Base future management tools, such as the trend evaluation reports, on projectwide analysis of problems, unless there are compelling reasons for a lesser scope. * Establish quality guidelines for trend evaluation reports to ensure sound analysis when reporting problems for management's attention. Such guidelines should address, among other things, having reliable and complete data and appropriate benchmarks. * To the extent practicable, make analyses and indicators of performance consistent over time so that trends or progress can be accurately identified and, where changes to analyses or indicators are made for compelling reasons, provide a clear history of the changes and their impact on measuring progress. * Focus the management tools' rating categories on the significance of the monitored condition, not on a judgment of the need for management action. Agency Comments: We provided DOE and NRC with draft copies of this report for their review and comment. In oral comments, DOE agreed with our recommendations and provided technical and editorial comments that we have incorporated in the report, as appropriate. We also incorporated, as appropriate, NRC's oral editorial comments, which primarily served to clarify its role. As agreed with your office, unless you publicly announce the contents of this report earlier, we plan no further distribution until 30 days from the report date. At that time, we will send copies to interested congressional committees and Members of Congress, the Secretary of Energy, and the Chairman of the Nuclear Regulatory Commission. We will also make copies available to others upon request. In addition, the report will be available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov]. If you or your staff have any questions about this report, please contact me at (202) 512-3841 or [Hyperlink, wellsj@gao.gov]. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made major contributions to this report are listed in appendix III. Sincerely yours, Signed by: Jim Wells: Director, Natural Resources and Environment: [End of section] Appendixes: Appendix I: Objectives, Scope, and Methodology: The objectives of this review were to determine (1) the history of the Yucca Mountain project's quality assurance problems since the project's start in the 1980s, (2) the Department of Energy's (DOE) tracking of quality problems and progress implementing quality assurance requirements since our April 2004 report, and (3) challenges that DOE faces as it continues to address quality assurance issues within the project. In addition, we were asked to provide information about implementation of the project's Employee Concerns Program and the types of concerns raised in recent years through the program. To determine the history of the project's quality assurance problems, we reviewed our prior reports and those of DOE's Office of the Inspector General concerning the Yucca Mountain project. We also reviewed internal DOE evaluations and audit reports written about the quality assurance program and Nuclear Regulatory Commission (NRC) reports and NRC-prepared summaries of NRC and DOE quarterly management meetings, technical exchange meetings, and quality assurance meetings dating to early 2004. In addition, we reviewed letters and communications between DOE and NRC regarding quality assurance from the NRC Web archives from the late 1980s. Furthermore, we reviewed plans for the Regulatory Integration Team (RIT) and subsequent correspondence between Bechtel/SAIC Company, LLC (BSC), DOE's management contractor for the Yucca Mountain project, and DOE. Moreover, we discussed quality assurance issues with officials of DOE's Office of Civilian Radioactive Waste Management (OCRWM), including the Acting Director and Deputy Director, at DOE headquarters in Washington, D.C., and at its field office in Las Vegas. In addition, we interviewed representatives of Navarro Quality Services, a DOE subcontractor, as well as BSC, and NRC officials in the agency's field office in Las Vegas, Nevada, and at its headquarters in Rockville, Maryland. To determine DOE's tracking of quality problems and progress implementing quality assurance requirements since our April 2004 report, we interviewed OCRWM, BSC, and NRC officials about the status of these efforts since the issuance of our prior report. We also reviewed DOE's Management Improvement Initiatives (2002), DOE's Management Improvement Initiatives Transition Approach (2003), and our 2004 report to understand the history of the improvement efforts. To understand DOE's management tools to monitor problems and progress, we reviewed the available performance indicators panels from April 2004 through August 2005, when it was last produced; the documentation on the individual indicators applied to August 2005 data; and the quarterly trend reports from the fourth quarter of fiscal year 2003 through the fourth quarter of fiscal year 2005. We also reviewed information from condition reports and examined documentation on DOE's Quality Assurance Requirements and Description (issued in August 2004), BSC's Trend Evaluation and Reporting, and DOE's Procedure: Condition Reporting and Resolution (issued in November 2005). To determine challenges that DOE faces as it continues to address quality assurance issues within the project, we reviewed information from condition reports, NRC on-site representative reports, DOE Inspector General reports, and an OCRWM Office of Concerns Program's investigative report on past quality assurance problems and DOE's efforts to address them. We obtained information on turnover in key management positions at DOE and BSC since 2000. In addition, we discussed with DOE and NRC officials DOE's difficulties in addressing recurring quality assurance problems and the quality assurance implications of the Yucca Mountain project moving from the site characterization phase to design and licensing. Also, to better understand issues and challenges, we attended quarterly meetings held between DOE and NRC in Rockville in September 2005 and Las Vegas in December 2005. To identify recent employee concerns related to quality assurance, such as falsification of records and a safety-conscious work environment, as well as to identify the actions taken to address those concerns, we reviewed all concerns received by the OCRWM and BSC Employee Concerns Programs from January through November 2005. For the OCRWM program, we reviewed all employee concerns files to identify concerns related to quality assurance. For the BSC program, we first read summary descriptions of each concerns file, and reviewed the concerns files for only those we identified as related to quality assurance. We then conducted a content analysis of all concerns files that we reviewed. Next, our three team members reached consensus about the correct classification of a concern as a quality assurance problem, such as potential falsification of records. Finally, through a second review of concerns files, we verified our recorded information for those concerns that seemed to be important illustrations of problems. In addition, we also spot-checked a sample of OCRWM and BSC concerns received in 2005 to verify the accuracy of their placement in various concerns categories. We found that the concerns were generally categorized accurately. We performed our work from July 2005 through January 2006 in accordance with generally accepted government auditing standards. [End of section] Appendix II: Yucca Mountain Project Employee Concerns Programs: NRC expects licensees to establish a safety conscious work environment- -that is, one in which (1) employees are encouraged to raise concerns either to their own management or to NRC without fear of retaliation and (2) employees' concerns are resolved in a timely and appropriate manner according to their importance. NRC encourages but does not require licensees to establish employee concerns programs to help achieve such a work environment, and both DOE and BSC have established such programs.[Footnote 19] DOE's Employee Concerns Program is currently operated under the requirements of DOE Order 442.1A, but the department, in anticipation of becoming a licensee, is in the process of establishing the program to meet NRC expectations. DOE and contractor employees at the Yucca Mountain project may raise concerns about quality, safety, or other work environment issues--such as harassment, intimidation, retaliation, and discrimination--through various means. Employees are encouraged to resolve concerns at the lowest possible level in the organization, in the following order: * Use normal supervisory channels, such as by raising an issue to a manager for resolution. * Initiate a condition report through the Corrective Action Program--a process in which any employee can formally identify a problem on the project, such as with policies, procedures, or the work environment, and have the issue investigated and, if necessary, fixed through corrective actions. * Submit a concern via e-mail, telephone, or in person to one of the project's two Employee Concerns Programs--a BSC program for BSC employees and other subcontractors and another run by DOE for either DOE or BSC employees. * Contact NRC directly. The DOE and BSC concerns programs are intended to supplement rather than replace the resolution of problems through managers or the Corrective Action Program. DOE and BSC Employee Concerns Programs have each established a communication network to allow employees to register concerns. These networks include brochures and regular newsletters on the programs and numerous links to the program on the project's intranet, where employees can obtain concerns forms. Both the DOE and BSC concerns programs of the Yucca Mountain project have four main steps: 1. Employees notify the concerns program staff about issues that they feel should be corrected, such as safety or health issues; harassment, intimidation, retaliation, or discrimination; concerns raised through the Corrective Action Program; and quality assurance problems. 2. The concerns program staff document and handle the concern in accordance with the requirements of DOE Order 442.1A. 3. The concerns program notifies the employees of the results of the investigation and notifies management of any deficiencies. 4. Project management develops corrective actions for deficiencies, and the program validates that the concerns have been effectively addressed by the actions. Under DOE Order 442.1A, concerns may be addressed through an investigation by the concerns program staff, an independent investigation, a referral, a transfer, or a dismissal of the concern. Employees can request or waive confidentiality. If a concern is submitted anonymously, interpreting the main issues and problems is left up to the concerns program staff, and action on the concern may be limited if the submitted information does not clearly or sufficiently define the concern. The concerns program may conduct its own investigation of the concern. Alternatively, it may refer the concern to another project organization for investigation or resolution. After the results of the investigation or resolution are reported to the concerns program within a specified period, the concerns program accepts the results or requires additional actions. In other cases, concerns may be transferred to another organization with the appropriate subject matter responsibility or expertise, such as the Office of Human Relations, Office of General Counsel, or Office of the Inspector General. After investigating a concern, the concerns programs determines whether the concern is substantiated, partially substantiated, unsubstantiated, or indeterminate. If a concern is substantiated or partially substantiated, the investigation results are presented to the responsible senior managers. A concern is considered indeterminate when evidence is insufficient to substantiate a concern or allow for a conclusion to be drawn. Some concerns can be resolved through a noninvestigative resolution, a method to address concerns promptly when minimal effort is required for resolution. Some resolutions involve the development of management corrective action plans that are tracked until they are closed. In addition, for deficiencies that identify systemic problems, the concerns programs may file a condition report through the Corrective Action Program. Moreover, DOE and contractor employees are required to report certain conditions or alleged conditions to DOE's Office of the Inspector General under DOE Order 221.1, which covers waste, fraud, and abuse. The concerns program handles some employee concerns in this way. From January through November 2005, DOE's concerns program opened 139 employee concerns for investigation, and the BSC concerns program opened 112 concerns for investigation.[Footnote 20] DOE's concerns program places concerns into 14 categories, while the BSC program uses 20 categories.[Footnote 21] For both DOE and BSC, the category receiving by far the most concerns for calendar year 2005 was management: "management/mismanagement" for DOE and "management practices" for BSC. According to DOE, management concerns generally involved conditions related to management behavior, policy practice, budget allocation, or use of resources. According to the manager of BSC's program, about half of the concerns in the management practices category involve hiring and human relations issues and the other half involve organizational policies and other issues. The "quality" category accounts for a relatively small portion of total concerns--18 percent of concerns for the DOE program and 4 percent for the BSC program. Tables 3 and 4 show the concerns received by the DOE and BSC programs for January through November 2005. Table 3: Employee Concerns Opened for Investigation under DOE's Employee Concerns Program by Category of Concern, January through November 2005: Concern category: Management/mismanagement; Percentage of total concerns: 42. Concern category: Workplace violence; Percentage of total concerns: 0. Concern category: Harassment, intimidation, retaliation, and discrimination; Percentage of total concerns: 6. Concern category: Reprisal; Percentage of total concerns: 0. Concern category: Chilling effect; Percentage of total concerns: 5. Concern category: Security; Percentage of total concerns: 0. Concern category: Health; Percentage of total concerns: 0. Concern category: Safety; Percentage of total concerns: 4. Concern category: Environment; Percentage of total concerns: 0. Concern category: Fraud, waste and abuse; Percentage of total concerns: 4. Concern category: Human resources; Percentage of total concerns: 12. Concern category: Equal Employment Opportunity; Percentage of total concerns: 2. Concern category: Quality; Percentage of total concerns: 18. Concern category: Other; Percentage of total concerns: 8. Concern category: Total; Percentage of total concerns: 100. Source: DOE. Note: Percentages may not add to 100 because of rounding. [End of table] Table 4: Employee Concerns Opened for Investigation under BSC's Employee Concerns Program by Category of Concern, January through November 2005: Concern category: Management practices; Percentage of total concerns: 48. Concern category: Industrial; Percentage of total concerns: 1. Concern category: Health; Percentage of total concerns: 4. Concern category: Fraud; Percentage of total concerns: 3. Concern category: Fitness for duty; Percentage of total concerns: 1. Concern category: Ethics; Percentage of total concerns: 5. Concern category: Cyber; Percentage of total concerns: 0. Concern category: Access authorization; Percentage of total concerns: 0. Concern category: Environmental; Percentage of total concerns: 1. Concern category: Employee relations; Percentage of total concerns: 5. Concern category: Intimidation; Percentage of total concerns: 1. Concern category: Harassment; Percentage of total concerns: 0. Concern category: Discrimination; Percentage of total concerns: 4. Concern category: Chilling effect; Percentage of total concerns: 4. Concern category: Abuse; Percentage of total concerns: 4. Concern category: Training; Percentage of total concerns: 1. Concern category: Safety-conscious work environment; Percentage of total concerns: 3. Concern category: Retaliation; Percentage of total concerns: 4. Concern category: Quality; Percentage of total concerns: 4. Concern category: Other; Percentage of total concerns: 6. Concern category: Total; Percentage of total concerns: 100. Source: BSC. Note: Percentages may not add to 100 because of rounding. [End of table] The Employee Concerns Programs, which are designed to provide an alternative to raising issues through the Corrective Action Program and issuing condition reports, have been playing an active and sometimes key role in identifying and addressing quality assurance problems, as can be seen in the following examples: * As part of an effort to identify e-mails relevant to the licensing process and that therefore should be included in the Licensing Support Network, BSC employees in late 2004 discovered e-mails suggesting potential falsification of technical records. The e-mails were submitted to the Employee Concerns Program in March 2005 and were eventually reported to the DOE Inspector General for investigation. The quality assurance issues raised by the e-mails have resulted in a substantial effort by DOE to restore confidence in the quality of technical documents that will support its license application to construct the repository. * In mid-2005, the DOE concerns program referred to the project's senior management an employee's allegation that the project's schedule was taking priority over quality in the review of technical documents. In this instance, the Office of Concerns Program Manager negotiated with senior management to address the time and resource needs for ensuring quality assurance, rather than simply communicating to the organization that quality should take priority over the schedule. * As the result of an employee's concerns referred to DOE by NRC in mid- 2005, the Employee Concerns Program initiated an extensive investigation of issues related to requirements management. That investigation substantiated the employee's concerns and led to the issuance of 14 condition reports for problem resolution. Signifying the importance of this issue, DOE discussed problems with requirements management with NRC at their quarterly meeting in December 2005. The Employee Concerns Programs' role in identifying and addressing quality assurance and other issues is dependent upon employees' willingness to submit concerns, but the employees' willingness has sometimes been in doubt. A late 2004 DOE survey of project employees indicated, for example, that less than two-thirds of employees were confident that submitted concerns would be thoroughly investigated and appropriately resolved. DOE recognizes the need to improve employee trust and willingness to use the concerns program, and both the DOE and BSC program are engaged in outreach efforts. However, employees' willingness to submit concerns may be affected by factors outside the programs' control. According to a DOE manager, the project's recent and pending workforce reductions may account for a decreasing number of concerns submitted to the DOE program in late 2005. Based on OCRWM Employee Concerns Program data, the program averaged about 13 concerns a month from January through November 2005. However, the number of monthly concerns dropped to 5 in October and 3 in November 2005. During our review of concerns opened for investigation from January 2004 through November 2005, we did not identify any concerns alleging problems similar to the falsification of technical records suggested by the USGS e-mails. Although we found records of an early 2004 concern about an instance of inappropriate management of a technical document, this instance was resolved and did not appear to be an intentional or systematic effort to falsify records. The manager of the BSC program told us of a concern raised about another set of e-mails, but this concern was not about record falsification. The manager of the DOE program told us that she had not seen any reportable allegations of falsification of technical records since she took her position in July 2004. [End of section] Appendix III: GAO Contact and Staff Acknowledgments: GAO Contact: Jim Wells, (202) 512-3841 or [Hyperlink, wellj@gao.gov]. Staff Acknowledgments: In addition to the contact named above, Raymond Smith (Assistant Director), Casey Brown, John Delicath, James Espinoza, and Terry Hanford made key contributions to this report. (360611): FOOTNOTES [1] Once EPA finalizes its new standards, NRC will revise its licensing regulations to make them consistent with the standards. [2] GAO, Yucca Mountain: Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation, GAO-04-460 (Washington, D.C.: Apr. 30, 2004). [3] GAO-04-460, 5. [4] In addition, DOE must update this certification at the time of license application submittal, as required by NRC regulations. [5] Docketing is the formal acceptance of the license application by NRC after it determines that it contains adequate information for a formal review. [6] NRC is required to issue or deny a construction authorization not later than 3 years after receiving a license application, unless it extends this period by not more than 1 year and reporting the reasons for doing so to the Secretary of Energy and Congress. [7] GAO, Nuclear Waste: Repository Work Should Not Proceed Until Quality Assurance Is Adequate, GAO/RCED-88-159 (Washington, D.C.: Sept. 29, 1988). [8] U.S. Nuclear Regulatory Commission, U.S. Nuclear Regulatory Commission Staff Evaluation of U.S. Department of Energy Analysis Model Reports, Process Controls, and Corrective Actions (Washington, D.C., Apr. 7, 2004). [9] In addition, the RIT edited the AMRs to assure consistency and ease of technical and regulatory reviews. [10] Similar to our findings, a January 2006 NRC observation audit report noted that a DOE audit had found the trend program was unsatisfactory. Specifically, the DOE audit found that the trend program was not handled as a priority for management attention, and has historically identified broad causal issues with no adverse trends. [11] Skill-based errors are defined in trend reports as unintentional errors resulting from people not paying attention to the task at hand. [12] DOE prepares a separate trend report for CRs issued to DOE, but that report's findings are not integrated with those of BSC to provide a projectwide analysis. [13] The difference in number of CRs issued also reflects the fact that the scope of the investigations was broader than the trending report's scope, which focused only on CRs assigned to BSC. The investigation resulted in CRs assigned to DOE as well as BSC. [14] Appendix II offers a detailed description of the employee concerns program and the variety of employee submitted concerns, which are often not related to quality assurance, and our examination of concerns submitted since 2004 that did not find concerns similar to those raised about the potential falsification of records by USGS employees. [15] A root cause analysis in such reports involves methods for determining the root cause of a problem, which is the underlying cause that must change in order to prevent the problem from reoccurring. A root cause analysis is required for the most significant CRs--those determined to be at Level A in the Corrective Action Program. [16] U.S. Department of Energy, Office of Inspector General, Office of Inspections and Special Inquiries, Quality Assurance Weaknesses in the Review of Yucca Mountain Electronic Mail for Relevancy to the Licensing Process, DOE/IG-0708 (Washington, D.C., November 2005). [17] The Energy Systems Acquisition Advisory Board is a special organization within DOE that advises the Secretary in approving and revising plans for major acquisition projects. [18] U.S. Department of Energy, Office of Inspector General, Office of Audit Services, DOE/IG-0702, Use of Performance Based Incentives by the Office of Civilian Waste Management (Washington, D.C., September 2005). [19] The discussion in this report involves the DOE OCRWM office in Las Vegas, Nevada and not the separately operated office in OCRWM headquarters in Washington, D.C. The DOE employee concerns program oversees BSC's program. [20] The number of concerns opened for investigation by BSC includes referrals from DOE. Both organizations track disposition of these concerns and include them in their totals. According to the manager of the BSC Employee Concerns Program, about one-fourth of BSC's concerns are DOE referrals. [21] We reviewed all of the concerns submitted to DOE and BSC from January 2004 through November 2005 for their relevance to quality assurance issues. Although we did not verify the accuracy of DOE's and BSC's placement of all concerns into the above categories, we spot- checked a sample of DOE and BSC concerns and found that they were generally categorized accurately. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.