Yucca Mountain

DOE's Planned Nuclear Waste Repository Faces Quality Assurance and Management Challenges Gao ID: GAO-06-550T April 25, 2006

The Department of Energy (DOE) is working to obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain in Nevada. The project, which began in the 1980s, has been beset by delays. In 2004, GAO raised concerns that persistent quality assurance problems could further delay the project. Then, in 2005, DOE announced discovery of employee e-mails suggesting quality assurance problems. Quality assurance, which establishes requirements for work to be performed under controlled conditions that ensure quality, is critical to making sure the project meets standards for protecting public health and the environment. This testimony, which summarizes GAO's March 2006 report (GAO-06-313), provides information on (1) the history of the project's quality assurance problems, (2) DOE's tracking of these problems and efforts to address them since GAO's 2004 report, and (3) challenges facing DOE as it continues to address quality assurance issues within the project.

DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the 1980s and 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. More recently, as it prepares to submit a license application for the repository to NRC, DOE has been relying on costly and time-consuming rework to resolve lingering quality assurance problems uncovered during audits and after-the-fact evaluations. DOE announced, in 2004, that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed. However, GAO found that the project's performance indicators and other key management tools were not effective for this purpose. For example, the management tools did not target existing areas of concern and did not track progress in addressing them. The tools also had weaknesses in detecting and highlighting significant problems for management attention. DOE continues to face quality assurance and other challenges. First, DOE is engaged in extensive efforts to restore confidence in scientific documents because of the quality assurance problems suggested in the discovered e-mails between project employees, and it has about 14 million more project e-mails to review. Second, DOE faces quality assurance challenges in resolving design control problems associated with its requirements management process--the process for ensuring that high-level plans and regulatory requirements are incorporated into specific engineering details. Problems with the process led to the December 2005 suspension of certain project work. Third, DOE continues to be challenged to manage a complex program and organization. Significant personnel and project changes initiated in October 2005 create the potential for earlier problem areas, such as confusion over roles and responsibilities, to reoccur.



GAO-06-550T, Yucca Mountain: DOE's Planned Nuclear Waste Repository Faces Quality Assurance and Management Challenges This is the accessible text file for GAO report number GAO-06-550T entitled 'Yucca Mountain: DOE's Planned Nuclear Waste Repository Faces Quality Assurance and Management Challenges' which was released on April 26, 2006. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. United States Government Accountability Office: GAO: Testimony: Before the Subcommittee on the Federal Workforce and Agency Organization, Committee on Government Reform, House of Representatives: Tuesday, April 25, 2006: Yucca Mountain: DOE's Planned Nuclear Waste Repository Faces Quality Assurance and Management Challenges: Statement of Jim Wells, Director Natural Resources and Environment: GAO-06-550T: GAO Highlights: Highlights of GAO-06-550T, a testimony before the Subcommittee on the Federal Workforce and Agency Organization, Committee on Government Reform, House of Representatives. Why GAO Did This Study: The Department of Energy (DOE) is working to obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain in Nevada. The project, which began in the 1980s, has been beset by delays. In 2004, GAO raised concerns that persistent quality assurance problems could further delay the project. Then, in 2005, DOE announced discovery of employee e-mails suggesting quality assurance problems. Quality assurance, which establishes requirements for work to be performed under controlled conditions that ensure quality, is critical to making sure the project meets standards for protecting public health and the environment. This testimony, which summarizes GAO‘s March 2006 report (GAO-06-313), provides information on (1) the history of the project‘s quality assurance problems, (2) DOE‘s tracking of these problems and efforts to address them since GAO‘s 2004 report, and (3) challenges facing DOE as it continues to address quality assurance issues within the project What GAO Found: DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the 1980s and 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. More recently, as it prepares to submit a license application for the repository to NRC, DOE has been relying on costly and time-consuming rework to resolve lingering quality assurance problems uncovered during audits and after-the-fact evaluations. DOE announced, in 2004, that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed. However, GAO found that the project‘s performance indicators and other key management tools were not effective for this purpose. For example, the management tools did not target existing areas of concern and did not track progress in addressing them. The tools also had weaknesses in detecting and highlighting significant problems for management attention. DOE continues to face quality assurance and other challenges. First, DOE is engaged in extensive efforts to restore confidence in scientific documents because of the quality assurance problems suggested in the discovered e-mails between project employees, and it has about 14 million more project e-mails to review. Second, DOE faces quality assurance challenges in resolving design control problems associated with its requirements management process”the process for ensuring that high-level plans and regulatory requirements are incorporated into specific engineering details. Problems with the process led to the December 2005 suspension of certain project work. Third, DOE continues to be challenged to manage a complex program and organization. Significant personnel and project changes initiated in October 2005 create the potential for earlier problem areas, such as confusion over roles and responsibilities, to reoccur. View of Yucca Mountain and the Exploratory Tunnel for the Repository: [See PDF for Image] Source: DOE: [End of Image] What GAO Recommends: In its March 2006 report, GAO recommended actions DOE can take to improve the project‘s management tools and their use in identifying and addressing quality assurance and other problems. In commenting on a draft of the report, DOE agreed with GAO‘s recommendations. To view the full product, including the scope and methodology, click on the link above. For more information, contact Jim Wells at (202) 512-3841 or wellsj@gao.gov. [End of Section] Mr. Chairman and Members of the Subcommittee: I am pleased to be here today to discuss our work concerning quality assurance and other management challenges facing the Department of Energy (DOE) as it prepares to construct a deep geological repository at Yucca Mountain in Nevada for the nation's nuclear wastes. My testimony is based on our March 2006 report entitled Yucca Mountain: Quality Assurance at DOE's Planned Nuclear Waste Repository Needs Increased Management Attention.[Footnote 1] DOE is preparing an application for a license from the Nuclear Regulatory Commission (NRC) to construct an underground geological repository at Yucca Mountain for the permanent storage of highly radioactive nuclear waste. Nuclear waste is a by-product of the production of nuclear power, which provides about 20 percent of U.S. electricity. About 50,000 metric tons of nuclear waste are stored at 72 sites around the country, principally at commercial nuclear power plants. These wastes have been accumulating for several decades in surface storage designed to be temporary. The Nuclear Waste Policy Act of 1982 required DOE to construct a repository for permanent storage and to begin accepting these wastes by January 31, 1998. In 2002, after more than 15 years of scientific study, the President recommended and the Congress approved Yucca Mountain as a suitable location for the repository. However, DOE continues to encounter delays, and it is not certain when it will apply for the license to construct the repository. The licensing process requires DOE to demonstrate to NRC that its plans for the repository will meet Environmental Protection Agency standards for protecting public health and the environment from harmful exposure to the radioactive waste. To show that it can meet these standards, DOE has been conducting scientific and technical studies at the Yucca Mountain site that will provide supporting documentation for its planned license application. DOE has also established a quality assurance program to meet NRC requirements and ensure that its work and the technical information it produces are accurate and defensible. To accomplish this goal, the program established procedures that require scientific, design, engineering, procurement, records keeping, and other work to be performed under controlled conditions that ensure quality and enable the work to be verified by others. However, persistent problems implementing these procedures and resulting questions about the quality of the work are significantly contributing to delays in DOE's submission of the license application. Resolving these quality assurance issues is essential to proceeding with construction. In April 2004, we reported that recurring quality assurance problems at the Yucca Mountain project could delay the licensing and operation of the repository. At that time, we also reported that DOE had completed efforts--known as Management Improvement Initiatives (Initiatives)--to better manage quality assurance problems, but could not assess their effectiveness because its performance goals lacked objective measures and time frames for determining success.[Footnote 2] Then, in early 2005, DOE reported that it had discovered a series of e-mail messages written in the late 1990s by U.S. Geological Survey (USGS) employees working on the Yucca Mountain project under a contract with DOE that appeared to imply that workers had falsified records for scientific work. Several of these messages appeared to show disdain for the project's quality assurance program and its requirements. In October 2005, DOE began planning an aggressive series of changes--known as the "new path forward"--to the facility design, organization, and management of the Yucca Mountain project. These efforts are intended to address quality assurance and other challenges, including those associated with the USGS e-mails, and advance the license application process. However, in December 2005 and again in February 2006, some project work was stopped due to continuing quality assurance problems. Our March 2006 report examined (1) the history of the project's quality assurance problems since its start in the 1980s, (2) DOE's tracking of quality problems and progress implementing quality assurance requirements since our April 2004 report, and (3) challenges that DOE faces as it continues to address quality assurance issues with the project. To determine the history of quality assurance problems, we reviewed previous GAO, DOE, and NRC documents, visited the Yucca Mountain site, and interviewed officials from DOE, NRC, and Bechtel/ SAIC Company, LLC (BSC), which is DOE's management contractor for the Yucca Mountain project. To assess DOE's tracking of quality-related problems and progress in addressing them, we examined management tools and associated documentation, and interviewed BSC and DOE officials regarding those tools. To identify current quality assurance and other challenges, we attended quarterly NRC management meetings, interviewed the Acting Director and other senior managers of the DOE project, and gathered information on management turnover. The work on our report was conducted from July 2005 through January 2006 in accordance with generally accepted government auditing standards. In summary, we found the following: * DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the late 1980s and early 1990s, DOE had problems assuring NRC that it had developed adequate plans and procedures related to quality assurance. As we reported in 1988, NRC had found that DOE's quality assurance procedures were inadequate and its efforts to independently identify and resolve weaknesses in the procedures were ineffective. By the late 1990s, DOE had largely addressed NRC's concerns about its plans and procedures, but its own audits identified quality assurance problems with the data, software, and models used in the scientific work supporting its potential license application. For example, in 1998, a team of project personnel determined that 87 percent of the models used to simulate the site's natural and environmental conditions, and to demonstrate the future repository's performance over time, did not comply with requirements for demonstrating their accuracy in predicting geologic events. More recently, DOE has relied on costly and time-consuming rework to resolve lingering quality assurance concerns. Specifically, in the spring of 2004, DOE implemented a roughly $20 million, 8-month project called the Regulatory Integration Team to ensure that scientific work was sufficiently documented and explained to support the license application. This effort involved about 150 full-time employees from DOE, USGS, and multiple national laboratories, such as Sandia and Los Alamos, working to inspect technical documents to identify and resolve quality problems. * DOE cannot be certain that its efforts to improve quality assurance have been effective because the management tools it adopted did not target existing management concerns and did not track progress in addressing significant and recurring problems. DOE announced in 2004 that it was making a commitment to continuous quality assurance improvement and that its efforts would be tracked by performance indicators that would enable it to assess progress and direct management attention as needed; however, its management tools have not been effective for this purpose. Specifically, its one-page summary, or "panel," of selected performance indicators that project managers used in monthly management meetings was not an effective tool for assessing progress because the indicators poorly represented the major management concerns and were changed frequently. For example, the panel did not include an indicator to represent the management concern about unclear roles and responsibilities--a problem that could undermine accountability within the project. Use of the indicator panel was discontinued in late 2005, and DOE is deciding on a tool to replace it. Moreover, a second management tool--trend evaluation reports--also did not track relevant concerns. The reports generally had technical weaknesses for identifying recurrent and significant problems and inconsistently tracked progress toward resolving the problems. For example, lacking reliable data and an appropriate performance benchmark for determining the significance of human errors as a cause of quality problems, DOE's trend reports offered no clear basis for tracking progress on such problems. In addition, under the trend reports' rating categories, the rating assigned to convey the significance of a problem was overly influenced by a judgment in the report that there were already ongoing management actions to address the problem, rather than solely assessing the problem's significance. For example, the trend report's rating of one particular problem at the lowest level of significance did not accurately describe the problem or sufficiently draw management's attention to it. * DOE's aggressive "new path forward" effort faces substantial quality assurance and other challenges, as it prepares to submit the license application to construct the repository at Yucca Mountain. First, the March 2005 announcement of the discovery of USGS e-mails suggesting the possible falsification of quality assurance records has resulted in extensive efforts to restore confidence in scientific documents, and DOE is conducting a wide-ranging review of approximately 14 million e- mails to determine whether they raise additional quality assurance issues. Such a review creates a challenge not just because of the sheer volume of e-mails to be reviewed, but also because DOE will have to decipher their meaning and determine their significance, sometimes without clarification from authors who have left the project. Furthermore, if any of the e-mails raise quality assurance concerns, further review, inspection, or additional work may need to be performed. Second, DOE faces quality assurance challenges associated with an inadequate requirements management process--the process responsible for ensuring that broad plans and regulatory requirements affecting the project are tracked and incorporated into specific engineering details. In December 2005, DOE issued a stop-work order on some design and engineering work until it can determine whether the requirements management process has been improved. Third, DOE continues to be challenged by managing a changing and complex program and organization. The significant project changes initiated under the new path forward create the potential for confusion over accountability as roles and responsibilities change--a situation DOE found to contribute to quality assurance problems during an earlier transition period. For example, one proposed reorganization--establishing a lead laboratory to assist the project--would not only have to be effectively managed, but also would introduce a new player whose accountability DOE would have to ensure. DOE has also experienced turnover in 9 of 17 key management positions since 2001--including positions related to quality assurance- -that has created management continuity challenges. For example, three individuals have directed the project since 1999, and the position is currently occupied by an acting director. Since DOE is still formulating its plans, it is too early to determine whether its new path forward effort will resolve these challenges. In our report, we recommend that DOE strengthen its management tools by (1) improving the tools' coverage of the Initiatives' areas of concern, (2) basing the tools on projectwide analysis of problems, (3) establishing quality guidelines, (4) making indicators and analyses more consistent over time, and (5) focusing rating categories on problem significance rather than a judgment on the need for management action. In commenting on the report, DOE agreed with our recommendations. Background: The Congress enacted the Nuclear Waste Policy Act of 1982 to establish a comprehensive policy and program for the safe, permanent disposal of commercial spent nuclear fuel and other highly radioactive wastes in one or more mined geologic repositories. The act charged DOE with (1) establishing criteria for recommending sites for repositories; (2) "characterizing" (investigating) three sites to determine each site's suitability for a repository (1987 amendments to the act directed DOE to investigate only the Yucca Mountain site); (3) recommending one suitable site to the President, who would submit a recommendation of such site to the Congress if he considered the site qualified for a license application; and (4) upon approval of a recommended site, seeking a license from NRC to construct and operate a repository at the site. The Yucca Mountain project is currently focused on preparing an application for a license from NRC to construct a repository. DOE is compiling information and writing sections of the license application, conducting technical exchanges with NRC staff, and addressing key technical issues identified by NRC to ensure that sufficient supporting information is provided. In February 2005, DOE announced that it does not expect the repository to open until 2012 at the earliest, which is more than 14 years later than the 1998 goal specified by the Nuclear Waste Policy Act of 1982. More recently, the conference report for DOE's fiscal year 2006 appropriations observed that additional significant delays to submitting a license application are likely. In October 2005, the project's Acting Director issued a memorandum calling for the development of wide-ranging plans for the "new path forward" to submitting the license application. The plans address the need to review and replace USGS work products, establish a lead national laboratory to assist the project, and develop a new simplified design for the waste canisters and repository facilities, among other things. In addition, DOE announced, in April 2006, that it was proposing legislation intended to accelerate licensing and operations. For example, the legislation provides that if NRC authorizes the repository, subsequent licensing actions would be conducted using expedited, simplified procedures. Given the delays, the Congress has considered other options for managing existing and future nuclear wastes, such as centralized interim storage at one or more DOE sites. In addition, the conference report for DOE's fiscal year 2006 appropriations directed DOE to develop a spent nuclear fuel recycling plan to reuse the spent fuel. However, according to the Nuclear Energy Institute, which represents the nuclear energy industry, none of technological options being considered will eliminate the need to ultimately dispose of nuclear waste in a geologic repository. DOE Has a Long History of Quality Assurance Problems at Yucca Mountain: DOE has had a long history of quality assurance problems at the Yucca Mountain project. In the project's early stages, DOE had problems assuring NRC that it had developed adequate quality assurance plans and procedures. By the late 1990s, DOE had largely addressed NRC's concerns about its plans and procedures, but its own audits identified quality assurance problems with the data, software, and models used in the scientific work supporting its potential license application. While recently resolving these quality problems, DOE is now relying on costly and time-consuming rework to ensure the traceability and transparency of several technical work products that are key components of the license application. As we reported in 1988, NRC reviewed DOE's quality assurance program for the Yucca Mountain project and concluded that it did not meet NRC requirements[Footnote 3] and that DOE's quality assurance audits were ineffective. In 1989, NRC concluded that DOE and its key contractors had yet to develop and implement an acceptable quality assurance program. However, by March 1992, NRC determined that DOE had made significant progress in improving its quality assurance program, noting among other things, that all of the contractor organizations had developed and were in the process of implementing quality assurance programs that met NRC requirements, and that DOE had demonstrated its ability to evaluate and correct deficiencies in the overall quality assurance program. By the late 1990s, however, the DOE quality assurance program began detecting new quality problems in three areas critical to demonstrating the repository's successful performance over time: data management, software management, and scientific models. * Data management. In 1998, DOE identified quality assurance problems with the quality and traceability of data, specifically that some data had not been properly collected or tested to ensure its accuracy and that data used to support scientific analysis could not be properly traced back to its source. DOE found similar problems in April and September 2003. * Software management. DOE quality assurance procedures require that software used to support analysis and conclusions about the performance and safety of the repository be tested or created in such a way to ensure that it is reliable. From 1998 to 2003, multiple DOE audits found recurring quality assurance problems that could affect confidence in the adequacy of software. * Model validation. In 1998, a team of project personnel evaluated the mathematical models used to simulate natural and environmental conditions and determined that 87 percent of them did not comply with validation requirements to ensure they accurately predict geologic events. In 2001, and again in 2003, DOE audits found that project personnel were not properly following procedures, specifically in the areas of model documentation, model validation, and checking and review. Further, the 2003 audit concluded that previous corrective actions designed to improve validation and reduce errors in model reports were not fully implemented. After many years of working to address these quality assurance problems with data, software, and models, DOE had mostly resolved these problems by February 2005. As DOE prepares to submit the Yucca Mountain project license application to NRC, it is relying on costly and time-consuming rework to ensure that the documents supporting its license application are accurate and complete. Although the department had known for years about quality assurance problems with the traceability and transparency of technical work products called Analysis and Model Reports (AMR)--a key component of the license application---DOE did not initiate a major effort to address these problems until 2004. AMRs contain the scientific analysis and modeling data that demonstrate the safety and performance of the planned repository and, among other quality requirements, must be traceable to their original source material and data and be transparent in justifying and explaining their underlying assumptions, calculations, and conclusions. In 2003, based in part on these problems, as well as DOE's long-standing problems with data, software, and modeling, NRC conducted an independent evaluation of three AMRs to determine if they met NRC requirements for being traceable, transparent, and technically appropriate for their use in the license application. In all three AMRs, NRC found significant problems with both traceability and transparency.[Footnote 4] NRC concluded that these findings suggested that other AMRs may have similar problems and that such problems could delay NRC's review of the license application, as it would need to conduct special inspections to resolve any problems it found with the quality of technical information. To address problems of traceability and transparency, DOE initiated an effort in the spring of 2004 called the Regulatory Integration Team (RIT) to perform a comprehensive inspection and rework of the AMRs and ensure they met NRC requirements and expectations.[Footnote 5] According to DOE officials, the RIT involved roughly 150 full-time personnel from DOE, USGS, and multiple national laboratories such as Sandia, Los Alamos, and Lawrence Livermore. The RIT decided that 89 of the approximately 110 AMRs needed rework. According to DOE officials, the RIT addressed or corrected over 3,700 problems, and was completed approximately 8 months later at a cost of about $20 million. In a February 2005 letter to DOE, the site contractor stated that the RIT effort had successfully improved the AMRs' traceability and transparency. Subsequently, however, DOE identified additional problems with traceability and transparency that required further inspections and rework. DOE initiated a review of additional AMRs that were not included in the scope of the 2004 RIT review after a March 2005 discovery of e-mails from USGS employees written between May 1998 and March 2000 implying that employees had falsified documentation of their work to avoid quality assurance standards. These additional AMRs contained scientific work performed by the USGS employees and had been assumed by the RIT to meet NRC requirements for traceability and transparency. However, according to DOE officials, DOE's review determined that these AMRs did not meet NRC's standards, and rework was required. DOE identified similar problems as the focus of the project shifted to the design and engineering work required for the license application. In February 2005, the site contractor determined that, in addition to problems with AMRs, similar traceability and transparency problems existed in the design and engineering documents that constitute the Safety Analysis Report--the report necessary to demonstrate to NRC that the repository site will meet the project's health, safety, and environmental goals and objectives. In an analysis of this problem, the site contractor noted that additional resources were needed to inspect and rework the documents to correct the problems. DOE Cannot Be Certain Its Efforts to Improve Quality Assurance Have Been Effective Because of Weaknesses in Tracking Progress and Identifying Problems: DOE's management tools for the Yucca Mountain project have not enabled it to effectively identify and track progress in addressing significant and recurring quality assurance problems. Specifically, its panel or one-page summary of selected performance indicators did not highlight the areas of management concern covered by its Management Improvement Initiatives (Initiatives) and had weaknesses in assessing progress because the indicators kept changing. Its trend reports also did not focus on tracking these management concerns, had technical weaknesses for identifying significant and recurrent problems, and has inconsistently tracked progress with problems. Furthermore, the trend reports have sometimes been misleading as to the significance of the problems being presented because their significance ratings tend to be lower if corrective actions were already being taken, without considering the effectiveness of the actions or the problem's importance to the project. In April 2004, DOE told us it expected that the progress achieved with its Initiatives for improving quality assurance would continue and that its performance indicators would enable it to assess further progress and direct management attention as needed. By that time, the actions called for by the Initiatives had been completed and project management had already developed a "panel" of indicators to use at monthly management meetings to monitor project performance. The panel was a single page composed of colored blocks representing selected performance indicators and their rating or level of performance. For example, a red block indicated degraded or adverse performance warranting significant management attention, a yellow block indicated performance warranting increased management attention or acceptable performance that could change for the worse, and a green block indicated good performance. The panel represented a hierarchy of indicators where the highest-level indicators were visible, but many lower-level indicators that determined the ratings of the visible indicators were not shown. Our review analyzed a subset of these indicators that DOE designated as the best predictors in areas affecting quality. We found that the panel was not effective for assessing continued progress because its indicators poorly represented the management concerns identified by the Initiatives. The Initiatives had raised concerns about five key areas of management weakness as adversely affecting the implementation of quality assurance requirements, and had designated effectiveness indicators for these areas. (These areas of concern are described in app. I.) However, two of the Initiatives' five key areas of concern--roles and responsibilities as well as work procedures--were not represented in the panel's visible or underlying indicators. In other cases, the Initiatives' effectiveness indicators were represented in underlying lower-level indicators that had very little impact on the rating of the visible indicator. For example, the Initiatives' indicator for timely completion of employee concerns was represented by two lower-level indicators that together contributed 3 percent of the rating for an indicator visible in the panel. Another shortcoming of the panel was that frequent changes to the indicators hindered the ability to identify problems for management attention and track progress in resolving them. The indicators could change in many ways, such as how they were defined or calculated. Such changes made it difficult to measure progress because changes in indicator ratings could reflect only the changes in the indicators rather than actual performance changes. Some of the indicators tracking quality issues changed from one to five times during the 8-month period from April 2004 through November 2004. Even after a major revision of the panel in early 2005, most of the performance indicators tracking quality issues continued to change over the next 6 months--that is, from March 2005 through August 2005. Only one of these five indicators did not change during this period. One indicator was changed four times during the 6-month period, resulting in it being different in more months than it remained the same. Moreover, the panel was not always available to track problems. It was not created for December 2004 through February 2005, and it has not been created since August 2005. In both cases, the panel was undergoing major revisions. In December 2005, a senior DOE official told us that the project would begin to measure key activities, but without use of the panel. According to DOE, a second management tool, the project's quarterly trend evaluation reports, captured some aspects of the Initiatives' areas of concern and their associated effectiveness indicators that were not represented in the performance indicators. However, the trend reports are designed more to identify emerging and unanticipated problems than to monitor progress with already identified problems, such as those addressed by the Initiatives. In developing these reports, trend analysts seek to identify patterns and trends in condition reports, which document problematic conditions through the project's Corrective Action Program. For example, analysis might reveal that most occurrences of a particular type of problem are associated with a certain organization. In practice, DOE missed opportunities to use trend reports to assess progress in the Initiatives' areas of concern. For example, DOE missed an opportunity to use trend reports to discuss the Initiatives' goal that the project's work organizations become more accountable for self- identifying significant problems. The August 2005 trend report briefly cited an evaluation of a condition report highlighting the low rate of self-identification of significant problems during the previous quarter and reported the evaluation's conclusion that it was not a problem warranting management attention. However, the trend report did not mention that about 35 percent of significant problems were self- identified during the previous quarter, while the Initiatives' goal was that 80 percent of significant problems would be self-identified. Beyond whether they effectively track the Initiatives' areas of concern, trend reports generally face serious obstacles to adequately identifying recurrent and significant problems. For example, trend analysis tends to focus on the number of condition reports issued, but the number of reports does not necessarily reflect the significance of a problem. For example, the number of condition reports involving requirements management decreased by over half from the first quarter to the second quarter of fiscal year 2005. However, this decrease was not a clear sign of progress. Not only did the number rise again in the third quarter, but the May 2005 trend report also noted that the number of all condition reports had dropped during the second quarter. According to the report, the volume of condition reports had been high in the first quarter because of reviews of various areas, including requirements management. Due, in part, to these obstacles, trend reports have not consistently determined the significance of problems or performed well in tracking progress in resolving them. For example, trend reports have questionably identified human performance as a significant problem for resolution and ineffectively tracked progress in resolving it because there was (1) no clearly appropriate or precise benchmark for performance, (2) a changing focus on the problem, and (3) unreliable data on cause codes. The February 2004 trend report identified a human performance problem based on Yucca Mountain project data showing the project's proportion of skill-based errors to all human performance errors was two times higher than benchmark data from the Institute of Nuclear Power Operations (INPO).[Footnote 6] Interestingly, the report cautioned that other comparisons with these INPO data may not be appropriate because of differences in the nature, complexity, and scope of work performed, but did not explain why this caution did not apply to the report's own comparison. While this comparison has not appeared in trend reports since May 2004, a November 2004 trend report changed the focus of the problem to the predominance of human performance errors in general, rather than the skill-based component of these errors. (Later reports reinterpreted this predominance as not a problem.) The report cited an adverse trend based on the fact that the human performance cause category accounted for over half of the total number of causes for condition reports prepared during the quarter. Nevertheless, by February 2005, trend reports began interpreting this predominance as generally appropriate, given the type of work done by the project. That is, the project's work involves mainly human efforts and little equipment, while work at nuclear power plants involves more opportunities for errors caused by equipment. In our view, this interpretation that a predominance of human performance errors would be expected implies an imprecise benchmark for appropriate performance. Further, although trend reports continued to draw conclusions about human performance problems, the February 2005 report indicated that any conclusions were hard to justify because of data reliability problems with cause coding. For example, the majority of problems attributed to human performance causes are minor problems, such as not completing a form, that receive less-rigorous cause analysis. This less-rigorous analysis tends to reveal only individual human errors--that is, human performance problems--whereas more-rigorous analysis tends to reveal less-obvious problems with management and procedures. Another shortcoming of the trend reports was that their rating categories made it difficult to adequately determine the significance of some problems. Specifically, trend reports sometimes assigned a problem a lower significance than justified because corrective actions were already being taken. The rating categories for a problem's significance also involve an assessment of the need for management action. In their current formulation, DOE's rating categories cannot accurately represent both these assessments, and the designated rating category can distort one or the other assessment. For instance, a November 2005 trend report rated certain requirements management issues as a "monitoring trend"--defined as a small perturbation in numbers that does not warrant action but needs to be monitored closely. However, this rating did not accurately capture the report's simultaneous recognition that significant process problems spanned both BSC and DOE and the fact that the numbers and types of problems were consistently identified over the previous three quarters. A more understandable explanation for why the problem received a low rating is that designating the problem at any higher level of significance would have triggered guidelines involving the issuance of a condition report, which, according to the judgment expressed in the report, was not needed. Specifically, the report indicated that existing condition reports have already identified and were evaluating and resolving the problem, thereby eliminating the need to issue a new condition report. However, by rating the problem at the lowest level of significance, the trend report did not sufficiently draw management's attention to the problem. At about the same time the trend report judged no new condition reports were necessary, a separate DOE investigation of requirements management resulted in 14 new condition reports--3 at the highest level of significance and 8 at the second-highest level of significance. These condition reports requested, for instance, an analysis of the collective significance of the numerous existing condition reports and an assessment of whether the quality assurance requirement for complete and prompt remedial action had been met. As a result of the investigation and a concurrent DOE root cause analysis,[Footnote 7] DOE stated during the December 2005 quarterly management meeting with NRC that strong actions were required to address the problems with its requirements management system and any resulting uncertainty about the adequacy of its design products. I would now like to update you on the project's February 2006 stop-work order, which occurred too late to be included in our report. We believe this incident is an example of how the project's management tools have not been effective in bringing quality assurance problems to top management's attention. After observing a DOE quality assurance audit at the Lawrence Livermore National Laboratory in August 2005, NRC expressed concern that humidity gauges used in scientific experiments at the project were not properly calibrated--an apparent violation of quality assurance requirements. According to an NRC official, NRC communicated these findings to BSC and DOE project officials on six occasions between August and December 2005, and issued a formal report and letter to DOE on January 9, 2006. However, despite these communications and the potentially serious quality assurance problems involved, the project's acting director did not become aware of the issue until January 2006, after reading about it in a news article. Due to concerns that quality assurance requirements had not been followed and the length of time it took top management to become aware of the issue, BSC issued a February 7, 2006, stop-work order affecting this scientific work. Project officials have begun a review of the issue. DOE's New Path Forward to Submitting a License Application Faces Substantial Quality Assurance and Other Challenges: In pursuing its new path forward, DOE faces significant quality assurance and other challenges, including (1) determining the extent of problems and restoring confidence in the documents supporting the license application after the discovery of e-mails raising the potential of falsified records, (2) settling the design issues and the associated problems with requirements management, and (3) replacing key personnel and managing the transition of new managers and other organizational challenges. The early 2005 discovery of USGS e-mails suggesting possible noncompliance with the project's quality assurance requirements has left lingering concerns about the adequacy of USGS's scientific work related to the infiltration or flow of water into the repository and whether other work on the project has similar quality assurance problems. As part of its new path forward, DOE has taken steps to address these concerns. It is reworking technical documents created by USGS personnel to ensure that the science underlying the conclusions on water infiltration is correct and supportable. In addition, DOE is conducting an extensive review of approximately 14 million e-mails to determine whether they raise additional quality assurance concerns. According to NRC on-site representatives, screening these millions of e- mails to ensure that records were not falsified will be challenging. Further, many of the e-mails were written by employees who no longer work at the project or may be deceased, making it difficult to learn their true meaning and context. Moreover, if additional e-mails raise quality assurance concerns, DOE may have to initiate further review, inspections, or rework. DOE officials have stated that it will need to resolve long-standing quality assurance problems involving requirements management before it can perform the design and engineering work needed to support the revised project plans called for by its new path forward. According to a 2005 DOE root cause analysis report, low-level documents were appropriately updated and revised to reflect high-level design changes through fiscal year 1995. However, from 1995 through 2002, many of these design documents were not adequately maintained and updated to reflect current designs and requirements. Further, a document that is a major component of the project's requirements management process was revised in July 2002, but has never been finalized or approved. Instead, the project envisioned a transition to a new requirements management system after the submission of the license application, which at that time was planned for December 2004. However, for various reasons, the license application was not submitted, and the transition to a new requirements management system was never implemented. The DOE report described this situation as "completely dysfunctional" and identified the root cause of these conditions as DOE's failure to fund, maintain, and rigidly apply a requirements management system. According to an NRC on-site representative, repetitive and uncorrected issues associated with the requirements management process could have direct implications for the quality of DOE's license application. In December 2005, DOE issued a stop-work order on design and engineering for the project's surface facility and certain other technical work. DOE stated that a root cause analysis and an investigation into employee concerns had revealed that the project had not maintained or properly implemented its requirements management system, resulting in inadequacies in the design control process. The stop-work order will be in effect until, among other things, the lead contractor improves the requirements management system, validates that processes exist and are being followed, and requirements are appropriately traced to implementing mechanisms and products. Further, DOE will establish a team to take other actions necessary to prevent inadequacies in requirements management and other management systems from recurring. Finally, DOE continues to be challenged to effectively manage a changing and complex program and organization. The significant project changes initiated under the new path forward create the potential for confusion over accountability as roles and responsibilities change--a situation DOE found to contribute to quality assurance problems during an earlier transition period. An important part of this challenge is ensuring that accountability for quality and results are effectively managed during the transition to the new path forward. For example, DOE's plan to establish a lead laboratory to assist the project would not only have to be effectively managed, but also would introduce a new player whose accountability DOE would have to ensure. According to one DOE manager, transitioning project work to a lead laboratory under a direct contract with DOE could pose a significant challenge for quality assurance because the various laboratories assisting with the project are currently working under BSC quality assurance procedures and will now have to develop their own procedures. In addition, the project faces management challenges related to ensuring management continuity at the project. DOE has experienced turnover in 9 of 17 key management positions since 2001. For example, in the past year, the project has lost key managers through the departures of the Director of Project Management and Engineering, the Director of the License Application and Strategy, the Director of Quality Assurance, and the contractor's General Manager. To ensure the right managers move the project forward to licensing, the project has a recruitment effort for replacing key departing managers. Further, the director position for the project has been occupied by three individuals since 1999 and is currently filled by an acting director. The current Acting Director took his position in summer 2005, and initiated the new path forward in October 2005. DOE is currently awaiting congressional confirmation of a nominee to take the director position. However, the current Acting Director told us he expects that the new path forward will be sustained after the new director assumes the position because it has been endorsed by the Secretary of Energy. Conclusions: DOE has a long history of trying to resolve quality assurance problems at its Yucca Mountain project. Now, after more than 20 years of work, DOE once again faces serious quality assurance and other challenges while seeking a new path forward to a fully defensible license application. Even as DOE faces new quality assurance challenges, it cannot be certain that it has resolved past problems. It is clear that DOE has not been well served by management tools that have not effectively identified and tracked progress on significant and recurring problems. As a result, DOE has not had a strong basis to assess progress in addressing management weaknesses or to direct management attention to significant and recurrent problems as needed. Unless these quality assurance problems are addressed, further delays on the project are likely. Mr. Chairman, this concludes my prepared statement, I would be happy to respond to any questions that you or other Members of the Subcommittee may have at this time. GAO Contact and Staff Acknowledgments: For further information about this testimony, please contact Jim Wells at (202) 512-3841 or wellsj@gao.gov. Casey Brown, John Delicath, Terry Hanford, and Raymond Smith also made key contributions to this statement. [End of section] Appendix I: The Management Improvement Initiatives' Key Areas of Concern: The Department of Energy's Management Improvement Initiatives (Initiatives) perceived five key areas of management weakness as adversely affecting the implementation of quality assurance requirements at the Yucca Mountain project: 1. Roles and responsibilities were becoming confused as the project transitioned from scientific studies to activities supporting licensing. The confusion over roles and responsibilities was undermining managers' accountability for results. The Initiatives' objective was to realign DOE's project organization to give a single point of responsibility for project functions, such as quality assurance and the Corrective Action Program, and hold the project contractor more accountable for performing the necessary work in accordance with quality, schedule, and cost requirements. 2. Product quality was sometimes being achieved through inspections by the project's Office of Quality Assurance rather than being routinely implemented by the project's work organizations. As a result, the Initiatives sought to increase work organizations' responsibility for being the principle means for achieving quality. 3. Work procedures were typically too burdensome and inefficient, which impeded work. The Initiatives sought to provide new user-friendly and effective procedures, when necessary, to allow routine compliance with safety and quality requirements. 4. Multiple corrective action programs existed, processes were burdensome and did not yield useful management reports, and corrective actions were not completed in a timely manner. The Initiatives sought to implement a single program to ensure that problems were identified, prioritized, and documented and that timely and effective corrective actions were taken to preclude recurrence of problems. 5.The importance of a safety-conscious work environment that fosters open communication about concerns was not understood by all managers and staff, and they had not been held accountable when inappropriately overemphasizing the work schedule, inadequately attending to work quality, and acting inconsistently in practicing the desired openness about concerns. Through issuing a work environment policy, providing training on the policy, and improving the Employee Concerns Program, the Initiatives sought to create an environment in which employees felt free to raise concerns without fear of reprisal and with confidence that issues would be addressed promptly and appropriately. GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site (www.gao.gov). Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to www.gao.gov and select "Subscribe to Updates." Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000 TDD: (202) 512-2537 Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Gloria Jarmon, Managing Director, JarmonG@gao.gov (202) 512-4400 U.S. Government Accountability Office, 441 G Street NW, Room 7125 Washington, D.C. 20548: Public Affairs: Paul Anderson, Managing Director, AndersonP1@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548: FOOTNOTES [1] GAO-06-313 (Washington, D.C.: Mar. 17, 2006). [2] GAO, Yucca Mountain: Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation, GAO-04-460 (Washington, D.C.: Apr. 30, 2004). [3] GAO, Nuclear Waste: Repository Work Should Not Proceed Until Quality Assurance Is Adequate, GAO/RCED-88-159 (Washington, D.C.: Sept. 29, 1988). [4] U.S. Nuclear Regulatory Commission, U.S. Nuclear Regulatory Commission Staff Evaluation of U.S. Department of Energy Analysis Model Reports, Process Controls, and Corrective Actions (Washington, D.C., Apr. 7, 2004). [5] In addition, the RIT edited the AMRs to assure consistency and ease of technical and regulatory reviews. [6] Skill-based errors are defined in trend reports as unintentional errors resulting from people not paying attention to the task at hand. [7] A root cause analysis seeks to determine the root cause of a problem, which is the underlying cause that must change in order to prevent the problem from reoccurring.

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.