Yucca Mountain

Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation Gao ID: GAO-04-460 April 30, 2004

The Department of Energy (DOE) must obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain, Nevada. In licensing, a quality assurance program helps ensure that the information used to demonstrate the safety of the repository is defensible and well documented. DOE developed a corrective action plan in 2002 to fix recurring problems with the accuracy of such information. This report assesses the status of corrective actions and the adequacy of DOE's plan to measure the effectiveness of actions taken.

DOE has reportedly implemented most of the actions in its 2002 corrective action plan, but recent audits and assessments have identified lingering quality problems with data, models, and software and continuing management weaknesses. Audits revealed that some data sets could not be traced back to their sources, model development and validation procedures were not followed, and some processes for software development and validation were inadequate or not followed. DOE believes these problems have not affected the technical basis of the project; however, they could adversely affect the licensing process. Recent assessments identified continuing management weaknesses in the areas of roles and responsibilities, quality assurance policies and procedures, and a work environment that did not foster employee confidence in raising concerns without fear of reprisal. NRC has acknowledged DOE's effectiveness in identifying quality problems, but recently concluded that quality problems could delay the licensing process. DOE cannot assess the effectiveness of its 2002 plan because the performance goals to assess management weaknesses lack objective measurements and time frames for determining success. The goals do not specify the amount of improvement expected, how quickly the improvement should be achieved, or how long the improvement should be sustained before the problems can be considered corrected. DOE recently developed a measurement tool that incorporates and revises some of the goals from the action plan, but most of the revised goals continue to lack the necessary time frames needed to determine whether the actions have corrected the recurring problems. A recently completed DOE review of the 2002 plan found that the corrective actions have been fully implemented. However, the review also noted the effectiveness of the actions could not be evaluated because many of the plan's goals lacked the level of objectivity and testing needed to measure effectiveness.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-04-460, Yucca Mountain: Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation This is the accessible text file for GAO report number GAO-04-460 entitled 'Yucca Mountain: Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation' which was released on April 30, 2004. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: April 2004: YUCCA MOUNTAIN: Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation: [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-460]: GAO Highlights: Highlights of GAO-04-460, a report to congressional requesters Why GAO Did This Study: The Department of Energy (DOE) must obtain a license from the Nuclear Regulatory Commission (NRC) to construct a nuclear waste repository at Yucca Mountain, Nevada. In licensing, a quality assurance program helps ensure that the information used to demonstrate the safety of the repository is defensible and well documented. DOE developed a corrective action plan in 2002 to fix recurring problems with the accuracy of such information. This report assesses the status of corrective actions and the adequacy of DOE‘s plan to measure the effectiveness of actions taken. What GAO Found: DOE has reportedly implemented most of the actions in its 2002 corrective action plan, but recent audits and assessments have identified lingering quality problems with data, models, and software and continuing management weaknesses. Audits revealed that some data sets could not be traced back to their sources, model development and validation procedures were not followed, and some processes for software development and validation were inadequate or not followed. DOE believes these problems have not affected the technical basis of the project; however, they could adversely affect the licensing process. Recent assessments identified continuing management weaknesses in the areas of roles and responsibilities, quality assurance policies and procedures, and a work environment that did not foster employee confidence in raising concerns without fear of reprisal. NRC has acknowledged DOE‘s effectiveness in identifying quality problems, but recently concluded that quality problems could delay the licensing process. DOE cannot assess the effectiveness of its 2002 plan because the performance goals to assess management weaknesses lack objective measurements and time frames for determining success. The goals do not specify the amount of improvement expected, how quickly the improvement should be achieved, or how long the improvement should be sustained before the problems can be considered corrected. DOE recently developed a measurement tool that incorporates and revises some of the goals from the action plan, but most of the revised goals continue to lack the necessary time frames needed to determine whether the actions have corrected the recurring problems. A recently completed DOE review of the 2002 plan found that the corrective actions have been fully implemented. However, the review also noted the effectiveness of the actions could not be evaluated because many of the plan‘s goals lacked the level of objectivity and testing needed to measure effectiveness. Quality Problems with Data, Models, and Software: [See PDF for image] [End of table] What GAO Recommends: GAO recommends that DOE revise action plan goals and close the plan once sufficient evidence exists showing that the actions have succeeded. In commenting on the report, DOE disagreed with the findings and recommendations, stating, among other things, that GAO mischaracterized the action plan and the results of independent reviews. GAO disagrees”the report correctly describes the plan and the findings of the reviews. NRC agreed with GAO‘s conclusions but suggested that DOE be given the flexibility to choose alternative approaches to achieve and measure performance. GAO agrees, provided that any approach include objective measures and time frames to assess effectiveness. www.gao.gov/cgi-bin/getrpt?GAO-04-460. To view the full product, including the scope and methodology, click on the link above. For more information, contact Robin M. Nazzaro at (202) 512-3841 or nazzaror@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: Quality Assurance Problems Persist at the Yucca Mountain Project: Corrective Action Plan Lacks Measurable Goals: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Scope and Methodology: Appendixes: Appendix I: Role of Quality Assurance in the Licensing Process: Appendix II: Employee Concerns Programs at the Yucca Mountain Project: Appendix III: 2002 Corrective Action Plan Process and Status: Appendix IV: Comments from the Department of Energy: GAO Comments: Appendix V: Comments from the Nuclear Regulatory Commission: GAO Comment: Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: Staff Acknowledgments: Tables: Table 1: Comparison of Goals in the July 2002 Corrective Action Plan to Goals in the December 2003 Performance Tool: Table 2: Employee Concerns Investigated by DOE and Bechtel in 2003: Figures: Figure 1: License Application Review Process and Timeline: Figure 2: 2002 Corrective Action Plan Process and Status: Abbreviations: DOE: Department of Energy: NRC: Nuclear Regulatory Commission: OCRWM: Office of Civilian Radioactive Waste Management: OMB: Office of Management and Budget: Letter April 30, 2004: The Honorable Harry Reid: United States Senate: The Honorable John Ensign: United States Senate: High-level nuclear waste, created as a by-product of the nuclear power process in reactors, can remain highly radioactive for hundreds of thousands of years, endangering the public if not properly disposed. Storing this waste safely is therefore of vital interest to the nation. Currently, more than 50,000 metric tons of this waste is being stored at 72 sites across the country. In 2002, Congress approved the President's recommendation of Yucca Mountain, Nevada, 90 miles from Las Vegas, as a suitable site for the Department of Energy (DOE) to construct and operate a geologic repository to safely and permanently dispose of this waste. To construct and operate the repository, DOE must obtain a license from the Nuclear Regulatory Commission (NRC). As part of the license application, DOE must, among other things, demonstrate an effective quality assurance program that ensures the safe construction and operation of a repository, protecting public health and safety. DOE plans to submit a license application by December 2004 and is following a demanding schedule to meet this date. NRC is reviewing an extensive amount of data as part of a prelicensing agreement with DOE. Before granting a license, NRC requires nuclear facilities to develop a quality assurance program that ensures that the technical information submitted in support of a license application--such as scientific data, models, and details on design and construction--is well documented and defensible. The quality assurance program involves a two-part process that (1) requires program staff to follow procedures to help ensure the reliability of information and (2) uses quality assurance auditors to verify that the procedures have been followed. Both program staff and quality assurance auditors are required to identify when procedures are not being followed or when they encounter problems with the procedures. DOE and contractor quality assurance auditors periodically assess compliance with procedures.[Footnote 1] In cases where a significant problem is found (DOE's criteria refer to significant problems as significant conditions adverse to quality), quality assurance personnel and program managers follow specific steps to analyze and correct the problem: 1. Quality assurance auditors or program personnel complete a corrective action report that describes the problem and the need for corrective action. 2. Program managers conduct a root-cause analysis of the problem. 3. Program managers identify corrective actions that address the root cause(s) to prevent the problem from recurring--these actions are included in a corrective action plan. 4. Program managers implement these corrective actions and quality assurance personnel verify that they have been implemented. 5. Quality assurance personnel close the corrective action report, and program managers later conduct an effectiveness review. a. If actions are determined ineffective, the process begins again with the issuance of a new corrective action report. b. In cases involving more significant problems, an effectiveness review may be conducted prior to closing the corrective action report. In 1998, DOE's quality assurance auditors identified significant problems with data sources, validation of scientific models, and software development and issued three corrective action reports. For data sources, DOE reported that it could not ensure that all data needed to support the scientific models could be tracked back to original sources or that the data had been properly collected. For validation of models, DOE reported that it had no standardized process to develop the scientific models needed to simulate geological events. For software, DOE reported that it had no process for ensuring that the software being developed to support the models would work. As required by DOE's quality assurance procedures, the department conducted a root- cause analysis and issued a corrective action plan in 1999 that identified the needed corrective actions. Following implementation of the actions, DOE considered the issues resolved and closed the corrective action reports. However, problems with models and software resurfaced during 2001 quality assurance audits. As a result, new corrective action reports were completed in May and June 2001, beginning another iteration of the corrective action process. Recognizing the need to correct these recurring problems, DOE (1) conducted a comprehensive root-cause analysis that included reviews of numerous past self-assessments and independent program assessments and (2) identified weaknesses in management systems, quality processes, and organizational roles and responsibilities. As a result, DOE issued a corrective action plan in July 2002 that addressed both the quality problems with data and models and the management weaknesses.[Footnote 2] In addition to the 37 actions in the 2002 plan that addressed models and software, DOE added 35 corrective actions to address management weaknesses that it found in five key areas: * roles and responsibilities, * quality assurance processes, * written procedures, * corrective action plans, and: * a work environment that allows employees to raise quality concerns without fear of reprisal. To correct these weaknesses, DOE completed a management reorganization and issued new policy statements to clarify roles and responsibilities, revised the primary quality assurance implementing document, reviewed and revised program procedures, revised the system to correct quality problems, and provided new training for employees to encourage them to raise concerns about quality. To assess the effectiveness of its actions in correcting the management weaknesses, DOE developed 13 goals to determine whether the corrective actions were successful, such as achieving decreasing trends in problems attributed to unclear roles and responsibilities, reducing the time required to revise procedures and complete corrective actions, and reducing the number of employee concerns related to the work environment. Because of the significance of these problems, DOE stated in the 2002 corrective action plan that an effectiveness review would be completed prior to closing the corrective action reports and reporting the results to NRC. In May 2003, at a congressional field hearing, we provided preliminary observations on the Yucca Mountain quality assurance program.[Footnote 3] Specifically, we noted DOE's poor track record in correcting recurring quality assurance problems and provided preliminary observations on recent actions taken to correct these problems. You requested that we continue our evaluation of the quality assurance program at Yucca Mountain, focusing on DOE's actions to correct the recurring quality problems. As agreed with your offices, this report (1) assesses the status of DOE's corrective actions to resolve recurring problems and (2) determines the adequacy of DOE's plan to measure the effectiveness of these actions. In conducting our work, we met with DOE and contractor officials, assessed the status of DOE's corrective actions, reviewed audits and deficiency reports, and visited the Yucca Mountain project office. We met with NRC officials and reviewed NRC-prepared documents, including observation audits, on-site representative reports, and correspondence between DOE and NRC. We attended several DOE-NRC quarterly quality assurance meetings and met with representatives of the State of Nevada Agency for Nuclear Projects and with representatives of the Nuclear Waste Technical Review Board. Our work was performed from April 2003 to April 2004 in accordance with generally accepted government auditing standards. Our scope and methodology for this review are presented at the end of this letter. Results in Brief: DOE reports that it has implemented almost all of the corrective actions detailed in its 2002 plan, but recent audits and assessments show that these actions have not solved the quality assurance problems or corrected management weaknesses, and that further actions are needed. Quality assurance audits found continuing problems with data, models, and software, including unqualified sources for the data used in modeling repository performance, noncompliance with processes used for the development and validation of models, and ineffective processes for developing software. DOE officials have stated that these findings represent problems with procedures and documentation and do not invalidate the technical products produced using the data, models, and software. However, the persistence of these problems could adversely affect the licensing process because DOE must demonstrate an effective quality assurance program as part of this process. Recent assessments show that management weaknesses remain despite DOE's actions. For example, one assessment notes that staff roles and responsibilities remain poorly defined, and that personnel are still not following procedures. Another assessment concluded that despite communication mechanisms, DOE had not established a climate of trust in the workplace. NRC has acknowledged the ability of DOE's quality assurance auditors to effectively identify quality problems; however, a recent NRC evaluation concluded that quality problems could adversely affect the licensing process. DOE cannot formally assess the overall effectiveness of its corrective actions because the plan's performance goals to assess management weaknesses lack objective measurements and time frames to determine whether corrective actions have been successful. Most of these goals fail to specify the amount of improvement expected, and none of them specify how quickly the improvement should be achieved or how long the improvement should be sustained before the problems can be considered corrected. For example, one goal calls for a decreasing trend in the average time needed to make revisions in procedures, but it does not specify the desired amount of the decrease, the length of time needed to achieve the decrease, or how long the decrease must be sustained. DOE has recently developed a project measurement tool that incorporates and revises some of the goals from the action plan, but most of the revised goals continue to lack the necessary time frames needed to determine whether the actions have corrected the recurring problems. A DOE independent review of the corrective action plan completed in March 2004 found that the corrective actions from the 2002 plan to address management weaknesses have been fully implemented. However, the review also noted the effectiveness of corrective actions under the plan could not be evaluated because many of the goals in the performance measurement tool that are linked to the 2002 plan lacked the level of objectivity and testing needed to measure effectiveness. To ensure proper assessment of the plan, we are recommending that DOE (1) revise the performance goals associated with the 2002 plan to ensure that they are measurable with specific time frames for achieving and maintaining success in each area of the plan and (2) close the plan after it develops evidence to show that the recurring quality assurance problems have been successfully corrected. In commenting on the report, DOE disagreed with the findings and recommendations, stating, among other things, that we mischaracterized the action plan and the results of several independent reviews. We disagree--the report correctly describes the plan and properly specifies the findings of the reviews. NRC agreed with our conclusions but suggested that DOE be given the flexibility to choose alternative approaches to achieve and measure performance. We agree, provided that any approach include objective measurements and time frames for reaching and sustaining desired performance and include an end point for closing out the 2002 plan. Background: In 2002, after more than 15 years of scientific investigation, Congress approved the Yucca Mountain site in Nevada as a suitable location for the development of a long-term permanent repository for high-level nuclear waste. DOE is responsible for developing and operating the repository, and NRC is responsible for licensing the repository. DOE is currently preparing an application to submit to NRC by December 2004 for a license to construct the repository. To obtain a license, DOE must, among other things, demonstrate to NRC that the repository will not exceed Environmental Protection Agency health and safety standards over a 10,000-year period. An ineffective quality assurance program runs the risk of introducing unknown errors into the design and construction of the repository that could lead to adverse health and safety consequences. To demonstrate compliance with the health standards over this 10,000- year period, DOE must rely primarily on a "performance assessment" computer model that incorporates over 1,000 data sources, approximately 60 scientific models, and more than 400 computer software codes to simulate the performance of the repository. Given the prominence of computer modeling in the licensing of the repository, one of DOE's most important tasks is to demonstrate the adequacy of the data, models, and software used to perform the simulation. In addition, as part of the licensing process, DOE must demonstrate that its quality assurance program can effectively identify and correct deficiencies in areas important to the safe operation and long-term performance of the repository, such as the natural and engineered barriers of the repository and the program's data, models, and software. See appendix I for more information on the role of quality assurance in the licensing process. DOE has a long-standing history of attempting to correct quality assurance problems. In 1988, we identified significant problems with the quality assurance program, noting that NRC had identified many specific concerns about the Yucca Mountain program, including: * DOE's heavy reliance on contractors and inadequate oversight would increase the likelihood that DOE might encounter quality-related problems; * the possibility that Nevada would contest the licensing proceedings, thereby increasing the probability that DOE would have to defend its quality assurance program; * additional expense and time-consuming delays to correct program weaknesses if DOE could not properly defend the quality of its work; and: * DOE staff's and contractors' negative attitude toward quality assurance.[Footnote 4] Since the late 1990s, DOE has attempted to correct continuing quality assurance problems in three areas critical to the repository's successful performance: the adequacy of the data sources, the validity of scientific models, and the reliability of computer software that have been developed at the site. These problems surfaced in 1998 when DOE began to run the initial versions of its performance assessment model. Specifically, DOE was unable to ensure that critical project data had been properly collected and tracked back to original sources. In addition, the department lacked a standardized process for developing scientific models used to simulate a variety of geologic events and an effective process for ensuring that computer software used to support the scientific models will work properly. DOE implemented actions in 1999 to correct these deficiencies and prevent their recurrence. In 2001, similar deficiencies associated with models and software resurfaced. DOE attributed the recurrence to ineffective procedures and corrective actions, improper implementation of quality procedures by line managers, and personnel who feared reprisal for expressing quality concerns. To ensure that it adequately addressed the problems to prevent future recurrence, DOE developed a more comprehensive corrective action plan in July 2002, concentrating on actions needed to address the causes of the recurring problems while improving the organizational culture and instilling a strong commitment to quality in all project personnel. The plan detailed specific actions for both DOE and its contractor, Bechtel/SAIC Company, LLC (Bechtel), to strengthen the roles, responsibilities, accountability, and authority of project personnel; develop clearer quality assurance requirements and processes; improve program procedures; create an improved programwide corrective action process; and improve processes for ensuring that employees can raise project concerns without fear of reprisals. Quality Assurance Problems Persist at the Yucca Mountain Project: DOE reports that it has implemented almost all of the actions identified in its 2002 corrective action plan; however, recent audits and assessments indicate that recurring quality assurance problems have not been corrected. In 2003, DOE conducted three audits to evaluate the effectiveness of the corrective actions taken to address recurring problems with data, models, and software. Because each audit identified additional quality assurance problems, DOE concluded that there was insufficient evidence to demonstrate that the recurring problems had been corrected. DOE recently closed the corrective action reports for data and software, but did so without determining whether corrective actions have been effective. To examine actions taken to correct some of the management weaknesses identified in the 2002 corrective action plan, DOE conducted four management assessments late in 2003. Collectively, these assessments found continuing management weaknesses that DOE had identified as root causes of the recurring problems. NRC also conducted an assessment that was issued in April 2004. NRC's assessment noted some improvements but also found continuing weaknesses and noted that quality assurance problems could hinder the licensing process. Audits Have Found Recurring Problems with Data, Models, and Software: In 2003, DOE's audits of data, models, and software identified continuing quality problems that could impede DOE's license application. As a result, DOE could not close corrective action reports for models and software for nearly 3 years. In a June 2003 audit, DOE found quality problems in developing and validating software. In September 2003, DOE quality assurance auditors found that some data sets were still not qualified or traceable to their sources. In October 2003, a DOE audit found continuing quality problems in model documentation and validation. DOE officials have stated that these findings represent problems with procedures and documentation and do not invalidate the technical products produced using the data, models, and software. In March 2004, DOE closed the corrective action reports for data and software but did so without evaluating the effectiveness of corrective actions--according to agency officials, they will evaluate effectiveness at a later date. DOE anticipates closing the corrective action report for models in August 2004 but also plans to do so without evaluating the effectiveness of corrective actions. Data Qualification and Traceability Problems Are Still Being Corrected: In April 2003, DOE again reported significant problems similar to those originally identified in 1998 with the qualification and traceability of data sets. At the time, DOE implemented corrective actions to recheck all of its data sets to confirm that they were traceable and qualified. However, a September 2003 audit identified similar data problems and new problems in addition to those noted in the corrective action report.[Footnote 5] The audit found that some data sets did not have the documentation needed to trace them back to their sources; the critical process of data control and management was not satisfactory; and, as in 1998, faulty definitions were developed for data procedures, which allowed unqualified data to be used. In addition, DOE found that overall compliance with procedures was unsatisfactory. Similarly, the April 2003 corrective action report also noted a lack of management leadership, accountability, and procedural compliance, issues which are closely related to the key improvement area of roles and responsibilities. DOE officials noted that these findings represented noncompliance with procedures, and that the procedures and processes were effective in producing defensible technical products if properly followed. As of February 2004, DOE had not finished rechecking all of its data sets or correcting problems in its data sets. However, DOE closed the corrective action report in March 2004 by making the rechecking process a continuing part of the Yucca Mountain repository's work. The corrective action report was closed without DOE evaluating the effectiveness of the rechecking process in correcting problems with data. DOE officials stated that they plan to evaluate effectiveness at a later date. Models Still Lack Proper Validation: An October 2003 DOE quality assurance audit found continuing problems with the documentation and validation of models that DOE plans to use in its license application.[Footnote 6] Although auditors reported that processes were effective in producing defensible models to support the license application, they found that for some models sampled, project personnel did not properly follow model validation procedures. These problems were similar to those identified by audits conducted in 2001. Auditors compared results from the 2003 audit with actions taken to correct problems identified in 2001 and found that procedures still were not being satisfactorily implemented in the areas of model documentation and traceability, model validation, and checking and review. For example, an action was taken in 2001 to improve the self- identification of problems before issuing new model reports by allowing for sufficient scheduling time for model checking and review. However, the 2003 audit concluded that instances of new errors in model reports were evidence that the previous actions may not have been fully implemented. As a result, DOE has been unable to close the May 2001 model corrective action report for almost 3 years. DOE recently directed a team of industry experts to review its models and revise them to ensure consistency, traceability, and procedural compliance. DOE anticipates closing the corrective action report in August 2004 but will do so without conducting another audit of models to determine if corrective actions have been effective. Software Development Problems Persist: In a June 2003 audit, DOE auditors discovered recurring software problems that could affect confidence in the adequacy of software codes.[Footnote 7] Specifically, the auditors found ineffective software processes in five areas: technical reviews, software classification, planning, design, and testing. The auditors found several of the software development problems to be similar to previously identified problems, indicating that previous actions were ineffective in correcting the problems. For example, auditors again noted instances of noncompliance with software procedures. They also concluded that technical reviews during software development were inadequate, even though documentation indicated that corrective actions for this condition had been completed 3 months before the 2003 audit. Auditors also noted poorly defined roles and responsibilities as a cause of problems identified in the technical review of software, even though DOE had taken actions under its 2002 corrective action plan to clarify roles and responsibilities. Because of these results, DOE was unable to close the June 2001 software corrective action report. DOE employed a team of industry professionals in the fall of 2003 to examine software quality problems identified from 1998 through 2003. The professionals' February 2004 report concluded that software problems recurred because DOE did not assess the effectiveness of its corrective actions and did not adequately identify the root causes of the problems. In a January 2004 follow-up audit of software, auditors verified that unqualified software was used to run approved models, and noted that procedural controls for determining the adequacy of software were inadequate. In March 2004, without evaluating the effectiveness of corrective actions, DOE closed the software corrective action report. DOE officials plan to evaluate the effectiveness of its corrective actions for software sometime in the future. Assessments Indicate Continuing Management Weaknesses: DOE reported in the fall of 2003 that it had implemented most of the actions identified in the plan focusing on management weaknesses, but four DOE management assessments of the Yucca Mountain project completed between September and November 2003 found that some of the identified management weaknesses had yet to be properly addressed. These assessments included one requested by project management comparing DOE's management practices at Yucca Mountain with external industry best practices,[Footnote 8] one required as an annual assessment of the adequacy and effectiveness of the quality assurance program,[Footnote 9] one requested by the project director that examined the effectiveness of selected DOE and contractor management systems,[Footnote 10] and one examining the project work environment.[Footnote 11] Collectively, these assessments identified continuing weaknesses in the areas of roles and responsibilities, quality assurance procedures, and a work environment that did not foster employee confidence in raising concerns without fear of reprisal. DOE officials stated that they are presently reviewing the findings of these assessments, and have recently initiated additional corrective actions. Unclear Roles and Responsibilities: Three of the four management assessments conducted late in 2003 identified significant continuing problems with the delineation and definition of roles and responsibilities for carrying out the quality assurance program. In its 2002 corrective action plan, DOE stated that it was not possible to build accountability into management without clearly and formally defining roles and responsibilities for DOE and its contractors. DOE's planned actions included clarification of roles and responsibilities within DOE and Bechtel through policy statements, communications, a new program manual, and realignment of the organization to support performance accountability. DOE reported that it had completed all corrective actions in this area by May 2003. The assessments noted that these actions had resulted in some improvements, but that some management weaknesses remained. The assessments found that the Yucca Mountain project: * lacked formal mechanisms for defining and communicating roles and responsibilities that meet both DOE and NRC requirements; * did not have a systematic process for assigning authorities to DOE and Bechtel organizations and individuals; * relied on program managers who had not fully assumed ownership and responsibility for quality assurance; * lacked formal control of documents outlining roles and responsibilities, ensuring that they reflect the organization; * lacked clear reporting relationships between the project and supporting national laboratories; * had not adequately established processes for reviewing procedures when needed; * had few systematic and effective approaches in place for assigning accountability to individuals and organizations; and: * did not effectively plan and communicate reorganizations and assign appropriate authority levels, in the opinion of many project employees. As a result of findings from these assessments, DOE is pursuing further corrective actions. For example, DOE plans to formally control the high-level document that defines its organizational structure. Also, Bechtel has initiated a management system improvement project, which includes issuing a new document defining roles and responsibilities. DOE officials expect that roles and responsibilities will continue to be a challenge in the future, but that efforts will continue. Ineffective Procedures: Three of the four management assessments identified continuing problems with project procedures, one of the areas of management weaknesses addressed by the 2002 corrective action plan. Although the assessments noted that DOE and Bechtel had made improvements in the procedure management system and DOE had reportedly reviewed existing procedures, issued new or revised procedures, and ensured that personnel using the procedures were properly trained, the assessments noted that: * procedures were overly prescriptive, * procedures did not cover all required processes, and: * continuing noncompliance with procedures remained a problem. Although DOE completed actions under the 2002 plan to revise project procedures, DOE has initiated further corrective actions, including a plan to again revise Yucca Mountain project procedures by June 2005. Inadequate Work Environment: Three of the four assessments identified continuing problems with efforts by DOE and Bechtel to ensure a work environment in which employees can freely raise concerns without fear of reprisal--one of the key areas of management weaknesses identified in the corrective action plan. DOE and Bechtel implemented corrective actions to improve the work environment by revising and expanding policies, modifying DOE contracts to require implementation of program requirements, decreasing the backlog of employee concerns, and providing programwide training that is based on industry practices. However, the assessments revealed continuing problems with the work environment, including both DOE's and Bechtel's employee concerns programs, which provide personnel with an opportunity to formally raise concerns about the project outside the normal chain-of-command without fear of reprisal. Appendix II describes the requirements of the Yucca Mountain employee concerns programs. Although the assessments noted ongoing management actions to strengthen the implementation of the concerns programs, they also noted that: * neither DOE nor Bechtel have effectively controlled corrective actions under the employee concerns programs, sometimes closing cases on the basis of anticipated actions; * both DOE and contractor employee concerns programs are not being utilized to their fullest; * there is a general lack of employee confidence in reporting safety issues to management; * DOE and Bechtel have not made effective resources available for determining the root causes of problems identified; * DOE and Bechtel have not established a climate of trust despite communication mechanisms and messages; and: * a majority of DOE and contractor employees either do not consider the project's corrective action process to be effective or are not sure of its effectiveness. Although the plan's actions to improve the work environment were completed in November 2003, DOE plans to take additional actions to improve employee confidence in raising issues without fear of reprisal. NRC Is Concerned That Recurring Problems Could Adversely Affect Licensing: NRC has commented on DOE's lack of progress in making improvements to the quality assurance program. At an April 2003 management meeting with DOE, an NRC official commented that the quality assurance program had not produced the outcomes necessary to ensure that the program is compliant with NRC requirements. In response, DOE outlined the steps it was taking to ensure that its license application would meet NRC expectations for completeness, accuracy, and compliance with quality assurance requirements. The steps included additional actions to improve performance in five areas: license application, procedural compliance, the corrective action program, the work environment, and accountability. In October 2003, DOE reported to NRC that it had completed some of the actions and was making progress in the remaining open action items. While NRC officials noted that DOE's actions might enhance performance, they found that significant implementation issues persist. NRC officials stated that they were seeking evidence of incremental DOE progress in the implementation of the quality assurance program in order to gain confidence in the adequacy of data, models, and software supporting the potential license application. In a November 2003 management meeting with DOE, NRC officials expressed encouragement with DOE's progress in implementing an improved corrective action process and the continued performance of effective audits and the identification of areas for improvement. However, the NRC staff continued to express concerns with DOE's lack of progress in correcting repetitive quality problems with models and software. NRC recently completed an evaluation of DOE's technical documents and supporting activities at Yucca Mountain. This prelicensing evaluation focused on an analysis of the technical information supporting three important repository models and the processes for developing and controlling the models. In addition, NRC evaluated the effectiveness of recent corrective actions in the areas of data, models, and software. The NRC report, released in April 2004, found that technical support for DOE's repository models was greatly improved, current models are more comprehensive and contain more data than those presented for site recommendation, software documentation was extensive, the management of databases was outstanding, and the trending program has been improved.[Footnote 12] However, the report noted concerns regarding the clarity and sufficiency of the technical information used to support the models. The NRC evaluation again found instances where data could not readily be traced back to their sources, unqualified data were used as direct inputs to the models, unqualified software was used to generate data supporting a model, and the model development process relied on inadequate checking and review procedures. In addition, NRC reported that DOE and Bechtel have not been fully successful in carrying out effective actions to eliminate recurring quality problems. The report states that DOE and Bechtel had not integrated human performance concerns in their root-cause and corrective action efforts in response to past quality problems. The NRC report concluded the following: "—if DOE continues to use its existing policies, procedures, methods, and practices at the same level of implementation and rigor, the license application may not contain information sufficient to support some technical positions in the application. This could result in a large volume of requests for additional information in some areas which could extend the review process, and could prevent NRC from making a decision regarding issuing a construction authorization to DOE within the time required by law.": Corrective Action Plan Lacks Measurable Goals: DOE cannot formally assess the overall effectiveness of its 2002 corrective action plan because the performance goals to assess management weaknesses in the plan lack objective measurements and time frames for determining success. For example, the goals do not specify the amount of improvement expected, how quickly the improvement should be achieved, or how long the improvement should be sustained before the problems can be considered corrected. For example, whereas 1 goal calls for a decreasing trend in the average time needed to make revisions in procedures, it does not specify the desired amount of the decrease, the length of time needed to achieve the decrease, or how long the decrease must be sustained. DOE recently developed a management tool to measure overall project performance that includes more than 200 performance indicators with supporting goals, including 17 goals linked to the 13 goals included in the 2002 corrective action plan. These 17 goals specify the desired amount of improvement, but most still lack the time frames needed for achieving and sustaining the goals. DOE officials told us they intend to use this performance measurement tool to track the progress of the project, including actions taken under the 2002 corrective action plan. A DOE independent review of the corrective action plan completed in March 2004 found that the corrective actions from the 2002 plan to address management weaknesses have been fully implemented. However, the review also noted the effectiveness of corrective actions under the plan could not be evaluated because many of the goals in the performance measurement tool that are linked to the 2002 plan lacked the level of objectivity and testing needed to measure effectiveness. Goals Are Not Objectively Measurable and Lack Specific Time Frames: DOE's 2002 plan included 13 goals to be used to determine the effectiveness of the corrective actions that addressed the five areas of management weaknesses. However, these goals were poorly defined, thus limiting DOE's ability to evaluate the effectiveness of actions taken. Both GAO[Footnote 13] and the Office of Management and Budget (OMB)[Footnote 14] have stated that performance goals need to be measurable, and time frames need to be established in order to track progress and demonstrate that deficiencies have been corrected. Of the 13 goals in the corrective action plan, 3 indicated how much improvement was expected. For example, 1 of the 3 goals specified that the number of significant quality problems self-identified by program managers should be at least 80 percent of all significant quality problems, including those identified by program managers, quality assurance auditors, or other employees. In contrast, 1 of the other 10 goals called for the achievement of a decreasing trend in the time needed for revising procedures, but did not specify how much of a decrease is expected. Further, none of the 13 goals specified the length of time needed to reach and maintain the desired goal to demonstrate that the actions taken were effective. For example, the goal calling for self-identified significant quality problems to be at least 80 percent of all significant quality problems did not indicate the length of time needed to achieve the goal or how long this goal should be sustained in order to demonstrate effectiveness. DOE does not intend to revise the goals of the 2002 corrective action plan to include quantifiable measures and time frames. Without such quantifiable measures to determine whether a goal has been met, and without a specified time for the goal to be maintained, DOE cannot use these goals to determine the effectiveness of the actions taken. Subsequent Efforts to Improve Goals Still Lack Time Frames: DOE's recent efforts to improve performance measurement have not allowed it to adequately measure the effectiveness of its corrective action plan. DOE has developed a projectwide performance measurement tool to assess project performance that includes over 200 performance indicators with supporting goals related to the project. At our request, Bechtel was able to link 17 of the supporting goals to 12 of the 13 goals of the 2002 corrective action plan. Although these linked goals improved quantifiable measurement for 11 of the plan's goals by specifying the amount of improvement expected, most did not include the necessary time frames for meeting the goals and sustaining the desired performance. DOE officials stated that this tool was not specifically tailored to evaluate the corrective action plan's effectiveness, but that they have decided to use it in lieu of the original 13 goals to monitor improvements and progress in correcting the management weaknesses identified in the plan. Table 1 provides a comparison of the supporting goals in the performance tool with the 2002 corrective action plan goals. Table 1: Comparison of Goals in the July 2002 Corrective Action Plan to Goals in the December 2003 Performance Tool: Key area of management weakness: Roles, responsibilities, accountability, authority; Original goals from corrective action plan, July 2002: (1) Improving trend in quality and schedule performance; Goals related to the corrective action plan in projectwide performance tool, December 2003: (1) Amount of actual work completed is 98 to 115 percent of the amount of work scheduled; GAO comments/observations: Quality performance is not included in new goal. Partial improvement for schedule - quantitative measure added, time frame to meet and sustain goal lacking. Key area of management weakness: Roles, responsibilities, accountability, authority; Original goals from corrective action plan, July 2002: (2) Decreasing trend in quality problems related to roles and responsibilities; Goals related to the corrective action plan in projectwide performance tool, December 2003: Goal is not covered in performance tool; GAO comments/ observations: Not applicable. Key area of management weakness: Quality assurance process; Original goals from corrective action plan, July 2002: (3) Numbers of high- priority (significant) quality problems that are self-identified are at least 80 percent of all significant quality problems; Goals related to the corrective action plan in projectwide performance tool, December 2003: (2) At least 80 percent of quality problems are self-identified; GAO comments/observations: Quantitative measure remains the same; goal is no longer focused on high-priority problems. Time frame to meet and sustain goal lacking. Key area of management weakness: Quality assurance process; Original goals from corrective action plan, July 2002: (4) Decreasing trend in average time to resolve significant quality problems and in number of delinquent corrective actions for significant quality problems; Goals related to the corrective action plan in projectwide performance tool, December 2003: (3) At least 90 percent of quality problems are closed in 60 days; (4) At least 90 percent of significant quality problems are closed in 100 days; (5) At least 80 percent of all problems are screened in 5 days; (6) At least 90 percent of all problems have corrective action plans in 30 days; (7) At least 80 percent of corrective action plans are on schedule; (8) 1 to 1.2 ratio of new problems to closed problems; GAO comments/observations: Partial improvement - quantitative measures added, time frames to meet and sustain goals lacking. Key area of management weakness: Written procedures; Original goals from corrective action plan, July 2002: (5) Decreasing number of quality problems related to ineffective procedures; Goals related to the corrective action plan in projectwide performance tool, December 2003: (9) 15 percent or less of all quality problems are based on ineffective procedures; GAO comments/observations: Partial improvement - quantitative measure added, time frame to meet and sustain goal lacking. Key area of management weakness: Written procedures; Original goals from corrective action plan, July 2002: (6) Decreasing trend in time needed to revise procedures; Goals related to the corrective action plan in projectwide performance tool, December 2003: (10) Procedure revisions are made in 75 days or less; GAO comments/ observations: Partial improvement - quantitative measure added, time frame to meet and sustain goal lacking. Key area of management weakness: Written procedures; Original goals from corrective action plan, July 2002: (7) Decreasing trend in average age of interim procedure changes; Goals related to the corrective action plan in projectwide performance tool, December 2003: (11) Interim procedure changes are made in less than 15 days; GAO comments/observations: Partial improvement - quantitative measure added, time frame to meet and sustain goal lacking. Key area of management weakness: Corrective action plans; Original goals from corrective action plan, July 2002: (8) Decreasing trend in number of repetitive quality problems; Goals related to the corrective action plan in projectwide performance tool, December 2003: (12) 5 percent or less of all corrective actions still have quality problems; (13) 5 percent or less of all quality problems are repeated; GAO comments/observations: Partial improvement - quantitative measures added, time frames to meet and sustain goals lacking. Key area of management weakness: Corrective action plans; Original goals from corrective action plan, July 2002: (9) Decreasing trend in average time to resolve significant quality problems; Goals related to the corrective action plan in projectwide performance tool, December 2003: (same as goals 3 and 4 in this column); GAO comments/ observations: Partial improvement - quantitative measure added, time frames to meet and sustain goals lacking. Key area of management weakness: Corrective action plans; Original goals from corrective action plan, July 2002: (10) Less than 10 percent of quality problems are resolved late; Goals related to the corrective action plan in projectwide performance tool, December 2003: (same as goals 3 and 4 in this column); GAO comments/observations: Partial improvement - quantitative measures added, time frames to meet and sustain goals lacking. Key area of management weakness: Work environment; Original goals from corrective action plan, July 2002: (11) Decreasing number of substantiated employee concerns for harassment, retaliation, intimidation, and discrimination; Goals related to the corrective action plan in projectwide performance tool, December 2003: (14) Zero concerns related to harassment, intimidation, retaliation, or discrimination are substantiated; GAO comments/observations: Partial improvement - quantitative measures added, time frame to meet and sustain goal lacking. Key area of management weakness: Work environment; Original goals from corrective action plan, July 2002: (12) Evaluation of routine employee concerns in less than 30 days, or 90 days for complex employee concerns involving harassment or intimidation; Goals related to the corrective action plan in projectwide performance tool, December 2003: (15) 25-day or less response time for routine concerns; (16) 80-day or less response time for complex concerns or harassment, retaliation, intimidation, or discrimination concerns; GAO comments/ observations: Partial improvement - quantitative measures changed, time frames to meet and sustain goals lacking. Key area of management weakness: Work environment; Original goals from corrective action plan, July 2002: (13) External evaluation of work environment shows positive changes; Goals related to the corrective action plan in projectwide performance tool, December 2003: (17) At least 80 percent favorable response rates to six employee survey questions; GAO comments/observations: Partial improvement - quantitative measure added, time frame to meet and sustain goal lacking. Source: GAO analysis of DOE data. Note: Performance goals in projectwide tool related to the corrective action plan represent a small fraction of the more than 200 goals being used for the project. [End of table] DOE Is Unable to Evaluate the Effectiveness of Corrective Actions: DOE has recently assessed the implementation of corrective actions, but it has not yet assessed the effectiveness of these actions in correcting recurring problems. In December 2003, DOE outlined the approach it used to determine whether corrective actions have been implemented.[Footnote 15] This approach is part of the overall process described in the 2002 action plan--appendix III provides an overview of the action plan and the status of the process. To determine if corrective actions had been implemented, DOE relied on the collective judgment of project managers regarding how effectively they have incorporated corrective actions into their regular project activities. A March 2004 DOE review analyzed the implementation of corrective actions for each of the management weaknesses but was not able to evaluate the effectiveness of the corrective actions.[Footnote 16] DOE's March 2004 review noted strong management commitment to improvement and described recent actions taken to ensure that work products meet quality objectives for a successful license application. However, the review identified continuing weaknesses in DOE's ability to determine the effectiveness of the actions it has taken. The review team attempted to measure how effectively DOE had met each of the plan's original 13 goals. The team was unable to measure whether 10 of the 13 goals had been met, but concluded that the project had met 2 of the goals and made progress toward another goal, based on an analysis of trends in quality problems identified. However, these conclusions were not based on an evaluation of quantifiable goals with time frames for meeting and sustaining the desired performance. The review also concluded that the performance indicators developed to evaluate the success of the actions lacked the level of objectivity and testing needed to measure effectiveness and that some lacked the data needed to assess effectiveness. The review recommended that DOE continue its corrective actions and refine performance indicators so that the effectiveness of corrective actions in meeting the plan's goals can be more readily measured. In April 2004, DOE notified NRC that it had completed, validated, and independently assessed the commitments it made in the 2002 corrective action plan, institutionalized the corrective actions, and established a baseline to foster and sustain continuous improvement. DOE officials stated they have achieved the initial goals of the 2002 plan through these actions. These officials indicated they would continue to refine and improve project tools used to evaluate the effectiveness of corrective actions. However, because of the limitations noted in its March 2004 review, DOE has not yet evaluated the effectiveness of corrective actions. Conclusions: Despite working nearly 3 years to address recurring quality assurance problems, recent audits and assessments have found that problems continue with data, models, and software, and that management weaknesses remain. As NRC has noted, quality assurance problems could delay the licensing process. Despite recurring quality problems, DOE has recently closed the corrective action reports for data and software and intends to close the corrective action report for models in August 2004 without first evaluating the effectiveness of the corrective actions taken to address the problems in these areas. DOE also does not intend to improve the goals of the 2002 plan associated with management weaknesses so that they can be adequately measured. Instead, DOE continues to plan and implement further actions to correct its quality problems and management weaknesses. This approach provides no indication regarding when DOE may be in a position to show that corrective actions have been successful. Entering into the licensing phase of the project without resolving the recurring problems could impede the application process, which at a minimum could lead to time- consuming and expensive delays while weaknesses are corrected and could ultimately prevent DOE from receiving authorization to construct a repository. Moreover, recurring problems could create the risk of introducing unknown errors into the design and construction of the repository that could lead to adverse health and safety consequences. Because of its lack of evidence that its actions have been successful, DOE is not yet in a position to demonstrate to NRC that its quality assurance program can ensure the safe construction and long-term operation of the repository. Recommendations for Executive Action: To better evaluate the effectiveness of management actions in correcting recurring quality problems, we recommend that the Secretary of Energy direct the Director, Office of Civilian Radioactive Waste Management, to: * revise the performance goals in the 2002 action plan to include quantifiable measures of the performance expected and time frames for achieving and maintaining this expected level of performance and: * close the 2002 plan once sufficient evidence shows that the recurring quality assurance problems and management weaknesses that are causing them have been successfully corrected. Agency Comments and Our Evaluation: We provided a draft of this report to DOE and NRC for their review and comments. DOE's written comments, which are reproduced in appendix IV, expressed disagreement with the report's findings and recommendations. DOE commented that the report did not properly acknowledge improvements the department has made in the quality assurance program; failed to properly characterize the 2002 Management Improvement Initiatives as a "springboard" to address management issues; did not consider DOE's use of the full range of performance indicators related to quality assurance; and mischaracterized the results of several independent, external reviews, taking a solely negative view of the findings. We disagree with most of DOE's comments. Our draft report acknowledged that DOE has taken a number of actions to address past problems in the quality assurance program, but to ensure clarity on this point, we have incorporated additional language to this effect in the report. However, our primary focus for this review was to evaluate the effectiveness of DOE's corrective actions in addressing the recurring quality problems. Despite the many actions taken to improve the quality assurance program, the management weaknesses and quality problems with data, models, and software have continued, indicating that the corrective actions have not been fully effective. Regarding DOE's views on our treatment of the 2002 Management Improvement Initiatives, DOE itself characterized the initiative as a "comprehensive corrective action plan." DOE stated that the implementation of the plan has been successful based on the evidence that responsible managers have taken agreed-upon action. This approach can be misleading, however, because it does not incorporate a determination of whether these actions have been effective. In fact, DOE has not evaluated the effectiveness of these actions in solving recurring problems. DOE further stated that we did not consider the full range of performance indicators related to quality assurance that DOE uses to manage the project. We agree. We asked DOE staff to compare their new performance indicators to the goals in the 2002 plan, and those are the goals that we presented for comparison in table 1 of our report. A discussion of the remainder of the hundreds of other goals was beyond the scope of our review and would not have added to an understanding of the overall problems with DOE's goals. Finally, we disagree with DOE's comment that we mischaracterized the results of recent independent reviews. We noted instances in these reports where improvements were found. However, we also devoted appropriate attention to evidence in these reports that address whether DOE's corrective actions have been effective. As our report states, these reports consistently found that these actions have not yet had their intended effect. In NRC's written comments, reproduced in appendix V, the agency agreed with our conclusions but suggested that DOE be given the flexibility to choose alternative approaches to achieve and measure quality assurance program performance. We agree that alternative approaches could be used to measure performance; however, to ensure the success of any approaches, DOE must include objective measurements and time frames for reaching and sustaining desired performance and include an end point for closing out the corrective action plan. Scope and Methodology: To assess the status of DOE's corrective actions to resolve recurring quality problems, we reviewed audits and deficiency reports written by the program over the past 5 years that identified problems with data, models, and software. We did not independently assess the adequacy of data, models, and software, but rather relied on the results of the project's quality assurance audits. In addition, we reviewed numerous documents that NRC prepared as part of its prelicensing activities at Yucca Mountain, including observations of quality assurance audits, NRC on-site representative reports, and correspondence between NRC and DOE on quality matters. We also observed an out-briefing of a quality assurance audit to obtain additional knowledge of how quality problems are identified and reported. To document the status of actions taken, we reviewed evidence used by DOE's Office of Civilian Radioactive Waste Management to prove corrective actions had been implemented and interviewed officials with DOE, at the Yucca site and in headquarters, and officials with Bechtel, the primary contractor. We also reviewed the results of four DOE assessments completed in the fall of 2003 that included the quality assurance program, interviewing the authors of the assessment reports to obtain a clear understanding of the problems identified. We attended quarterly meetings held between DOE and NRC to discuss actions taken under the plan and met with representatives of the State of Nevada Agency for Nuclear Projects and with representatives of the Nuclear Waste Technical Review Board, which was established to advise DOE on scientific and technical aspects of the Yucca Mountain project. To determine the adequacy of DOE's plan to measure the effectiveness of the actions it has taken, we examined the July 2002 corrective action plan and subsequent project performance measurement documents to determine how DOE intended to use goals and performance measures to evaluate the plan's effectiveness. We asked Bechtel officials to assist us in identifying and matching performance goals in the projectwide performance measurement tool with those in the 2002 corrective action plan. We compared DOE's approach in its corrective action plan and subsequent projectwide tool with GAO and OMB guidance on performance measurement. We discussed the implementation of the corrective action plan and methods for measuring its effectiveness with various DOE and NRC officials and DOE contractors in Washington, D.C., and at the Yucca Mountain project office in Las Vegas, Nevada. We also interviewed other GAO personnel familiar with performance measurement to more fully understand the key elements needed for effective assessments. We will send copies of this report to the appropriate congressional committees, the Secretary of Energy, and the Chairman of the Nuclear Regulatory Commission. We will also make copies available to others on request. In addition, this report will be available at no charge on the GAO Web site at [Hyperlink, http://www.gao.gov]. Signed by: If you or your staffs have any questions about this report, please call me on (202) 512-3841. Major contributors to this report are listed in appendix VI. Signed by: Robin M. Nazzaro Director, Natural Resources and Environment: [End of section] Appendixes: Appendix I: Role of Quality Assurance in the Licensing Process: After the Department of Energy (DOE) submits its license application to the Nuclear Regulatory Commission (NRC), NRC will review it to determine whether all NRC requirements have been met and whether the repository is likely to operate safely as designed. NRC's review will be guided by its Yucca Mountain Review Plan, which NRC developed to ensure the quality, uniformity, and consistency of NRC reviews of the license application and of any requested amendments.[Footnote 17] The review plan is not a regulation, but does contain the licensing criteria contained in federal regulations.[Footnote 18] DOE's application is to include general, scientific, and administrative information contained in two major sections: (1) a general information section that provides an overview of the engineering design concept for the repository and describes aspects of the Yucca Mountain site and its environs that influence repository design and performance, and (2) a detailed safety analysis section that provides a review of compliance with regulatory performance objectives that are based on permissible levels of radiation doses to workers and the public, established on the basis of acceptable levels of risk. The general information section covers such topics as proposed schedules for construction, receipt, and emplacement of waste; the physical protection plan; the material control and accounting program; and a description of site characterization work. The detailed safety analysis is the major portion of the application and includes DOE's detailed technical basis for the following areas: * the repository's safety performance before permanent closure in 100 to 300 years; * the repository's safety performance in the 10,000 years after permanent closure, on the basis of the "performance assessment" computer model; * a research and development program describing safety features or components for which further technical information is required to confirm the adequacy of design and engineered or natural barriers; * a performance confirmation program that includes tests, experiments, and analyses that evaluate the adequacy of information used to demonstrate the repository's safety over thousands of years; and: * administrative and programmatic information about the repository, such as the quality assurance program, records and reports, training and certification of personnel, plans for start-up activities, emergency planning, and control of access to the site. After DOE submits the license application (currently planned for December 2004), NRC plans to take 90 days to examine the application for completeness to determine whether DOE has addressed all NRC requirements in the application. One of the reviews for completeness will include an examination of DOE's documentation of the quality assurance program to determine if it addresses all NRC criteria. These criteria include, among other things, organization, design and document control, corrective actions, quality assurance records, and quality audits. If it deems any part of the application incomplete, NRC may either reject the application or require that DOE furnish the necessary documentation before proceeding with the detailed technical review of the application. If it deems the application complete, NRC will docket the application, indicating its readiness for a detailed technical review. Once the application is docketed, NRC will conduct a detailed technical review of the application over the next 18 months to determine if the application meets all NRC requirements, including the soundness of scientific analyses and preliminary facility design, and the NRC criteria established for quality assurance. If NRC discovers problems with the technical information used to support the license application, it may conduct specific inspections to determine the extent and effect of the problem. Because the data, models, and software used in modeling repository performance are integral parts of this technical review, quality assurance plays a key role since it is the mechanism used to verify the accuracy of the information DOE presents in the application. NRC may conduct inspections of the quality assurance program if technical problems are identified that are attributable to quality problems. According to NRC, any technical problems and subsequent inspections could delay the licensing of the repository or, in a rare instance, lead to ultimate rejection of the application. NRC will hold public hearings chaired by its Atomic Safety and Licensing Board to examine specific topics. Finally, within 3 to 4 years from the date that NRC dockets the application, NRC will make a decision to grant the application, reject the application, or grant it with conditions.[Footnote 19] Figure 1 presents the licensing process and timeline. Figure 1: License Application Review Process and Timeline: [See PDF for image] [End of figure] [End of section] Appendix II: Employee Concerns Programs at the Yucca Mountain Project: DOE and Bechtel/SAIC Company, LLC (Bechtel), have each established an employee concerns program to allow employees to raise concerns about the work environment without fear of reprisal. NRC requires licensees to establish a safe work environment where (1) employees are encouraged to raise concerns either to their own management or to NRC without fear of retaliation and (2) employees' concerns are resolved in a timely and appropriate manner according to their importance. DOE and contractor employees at Yucca Mountain have various means through which to raise concerns about safety, quality, or the work environment, including: * normal supervisory channels; * a corrective action program--a process in which any employee can formally cite a problem on the project, including the work environment, that needs to be investigated and corrective actions taken; * a DOE or contractor employee concerns program; or: * filing a concern directly with NRC. NRC encourages, but does not require, licensees to establish employee concerns programs. Both the DOE and Bechtel concerns programs at Yucca Mountain have three main steps: 1. An employee notifies concerns program staff about an issue that he/ she feels should be corrected, such as safety and health issues, free from harassment, retaliation, or quality assurance problems. 2. The concerns program staff documents and investigates the employee's concern. 3. The concerns program notifies the employee of the results of the investigation and notifies management of any need for corrective actions. DOE and Bechtel each have established a communication network to allow employees to register concerns. These networks include brochures and regular newsletters on the program and numerous computer links to the program on the contractor's intranet where employees can obtain concerns program forms on line. Recent statistics released by DOE show that most of the 97 concerns investigated by the DOE and Bechtel concerns programs in 2003 related to complaints against management. A summary of the concerns investigated in 2003 is shown in table 2. Table 2: Employee Concerns Investigated by DOE and Bechtel in 2003: Category of concern: Management problems or claims of mismanagement; Substantiated concerns: 26; Concerns not substantiated: 24; Total number of concerns: 50. Category of concern: Human resources; Substantiated concerns: 8; Concerns not substantiated: 6; Total number of concerns: 14. Category of concern: Harassment, intimidation, retaliation, or discrimination; Substantiated concerns: 4; Concerns not substantiated: 8; Total number of concerns: 12. Category of concern: Quality; Substantiated concerns: 5; Concerns not substantiated: 3; Total number of concerns: 8. Category of concern: Fraud, waste, and abuse; Substantiated concerns: 2; Concerns not substantiated: 1; Total number of concerns: 3. Category of concern: Safety; Substantiated concerns: 0; Concerns not substantiated: 2; Total number of concerns: 2. Category of concern: Equal employment opportunity; Substantiated concerns: 0; Concerns not substantiated: 1; Total number of concerns: 1. Category of concern: Security; Substantiated concerns: 0; Concerns not substantiated: 0; Total number of concerns: 0. Category of concern: Health; Substantiated concerns: 0; Concerns not substantiated: 0; Total number of concerns: 0. Category of concern: Environment; Substantiated concerns: 0; Concerns not substantiated: 0; Total number of concerns: 0. Category of concern: Workplace violence; Substantiated concerns: 0; Concerns not substantiated: 0; Total number of concerns: 0. Category of concern: Other; Substantiated concerns: 5; Concerns not substantiated: 2; Total number of concerns: 7. Category of concern: Total; Substantiated concerns: 50; Concerns not substantiated: 47; Total number of concerns: 97. Source: DOE. Note: Three concerns filed in 2003 were not included in this table. A concerns program official told us that two of these concerns were addressed by other organizations, and the resolution of the remaining concern was limited to providing information to management, as requested by the concerned individual. [End of table] [End of section] Appendix III: 2002 Corrective Action Plan Process and Status: DOE has established a process for completing corrective actions associated with the 2002 corrective action plan and evaluating their effectiveness. According to this process, after managers report they have taken actions to correct management weaknesses and specific problems with models and software, a confirmation team of DOE and contractor personnel verify that the actions have been completed. Once this step is completed, DOE conducts internal and external effectiveness reviews to determine if the actions have been effective in correcting the reported conditions. After the reviews of effectiveness, the results are assessed and reported to the Director of the Office of Civilian Radioactive Waste Management (OCRWM). The director then notifies NRC officials of the results of the effectiveness reviews, and the 2002 corrective action plan is closed. Figure 2 shows the corrective action plan process and the status of each step. Figure 2: 2002 Corrective Action Plan Process and Status: [See PDF for image] [End of figure] [End of section] Appendix IV: Comments from the Department of Energy: Department of Energy Washington, DC 20585: April 19, 2004: QA: N/A: Ms. Robin Nazzaro: Director, Natural Resources and Environment U.S. General Accounting Office: 441 G Street, NW Washington, D.C. 20548: Dear Ms. Nazzaro: Thank you for the opportunity to provide comments on the General Accounting Office (GAO) draft report, "Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation." Unfortunately, the Department must respectfully disagree with the report's findings and recommendations, and therefore with the conclusion that forms its title. I want to emphasize that the Office of Civilian Radioactive Waste Management (OCRWM) has shaped its quality assurance (QA) program to be consistent with Nuclear Regulatory Commission (NRC) requirements and standard industry practices. The role of QA is to verify that activities important to safety and waste isolation have been correctly performed. Detailed procedures address how technical work is documented so that the work and its results are reproducible, retrievable, transparent, and traceable. QA checks, audits, and inspections identify variances from procedures, and corrective actions are managed through a structured process. A nuclear QA program is effective if all personnel (not just those assigned to QA) proactively identify conditions that may affect quality or require attention, and if the organization can plan, implement, complete, and verify appropriate corrective actions in a timely manner. Evaluated against these criteria, I believe the OCRWM QA program has made significant progress and is operating effectively. In the Department's view, the major deficiencies of the draft report are as follows: The report authors did not acknowledge clear QA improvements we have made. By the measures of effectiveness that are important in the nuclear regulatory context, OCRWM has made substantial QA improvements. A few examples: In the last 15 months, self-identification by line management of conditions adverse to quality has increased by approximately 100 percent (Reference 1). This is a very positive development, showing that line managers are consistently reviewing their own work and ensuring that it is properly documented, reproducible, retrievable, transparent, and traceable, prior to being finalized. Some corrective actions require the modification of procedures (e.g., to better define requirements for documentation of scientific work), and the procedural change itself must follow strict QA processes. We have improved our ability to ensure that QA procedures are appropriate to the quality objectives they support, reducing the average time it takes to modify a procedure from a past average of seven months to a current average of three months (Reference 2). Line managers are increasingly effective in developing corrective actions for identified issues. Since October 2003, 75% of line managers' corrective action plans were accepted upon the first review by QA personnel (Reference 1). Our ability to address and close quality issues was demonstrated by the closure of two major, longstanding Corrective Action Reports on data management and software qualification in March 2004. Our safety conscious work environment - in the nuclear context, an environment where employees feel free to raise concerns about quality or safety without fear of reprisal - has been strengthened, as shown by internal surveys, performance indicators, and a comprehensive independent survey conducted in August 2003 by International Survey Research (Reference 3). That independent survey characterized OCRWM as "significantly and largely more positive" in the area of safety conscious work environment than other Federal agencies associated with research and technology. * The report authors failed to properly characterize the 2002 Management Improvement Initiatives (MII) effort, which is referred to inaccurately throughout the GAO draft report as the "2002 corrective actions plan," and its relationship to ongoing QA and management activities. The draft report asserts that although the 2002 MII is complete, there are "lingering quality problems with data, models, and software, and continuing management weaknesses." As noted above, the corrective actions associated with data management and software were verified and closed in March 2004. The corrective action report on validation of model reports is on track for closure within four months. With regard to alleged management weaknesses, the 2002 MII was initiated by the Department as an aggressive "springboard" effort to address management issues and transition improvements into day-to-day line management activities. The 2002 MII addressed five areas: roles, responsibilities, authority, and accountability; quality assurance programs and processes; program procedures; the Corrective Action Program; and safety conscious work environment. The focus on roles, responsibilities, authority, and accountability was important because lack of clarity in those areas had been cited in the past as a root cause of QA problems. Implementation of the 2002 MII has been successful: responsible managers demonstrated with objective evidence that they completed the commitments set out in MII action plans, and an independent assessment by Longenecker and Associates confirmed that the MII action statements had been appropriately completed (References 4, 5, and 6). On April 5, 2004, I wrote to the NRC to indicate that we have completed, validated, and independently assessed MII implementation and have transitioned the 2002 MII goals to ongoing line management activities (Reference 7). * The full range of performance indicators used by OCRWM to manage QA- related issues was not considered. The draft report suggests that the Department cannot assess the effectiveness of the 2002 MII because performance goals lack objective measures and timeframes. The effectiveness indicators that we defined as part of the 2002 MII were management metrics that supported improvement goals by setting high expectations and describing a desired future state to work toward. Work execution metrics, by contrast, quantify performance and set timeframes as appropriate - GAO included a small selection of these metrics in its chart on page 20 of the draft. OCRWM has in fact more than 300 performance indicators that we use to assess progress and identify issues on a continuous basis (Reference 8). Performance indicators are evaluated in detail at Monthly Operating Reviews (Reference 9). These elements of the OCRWM management tool inventory were not adequately addressed in the draft GAO report. * The draft report mischaracterizes the results of several independent, external reviews. The report does not acknowledge the positive findings that external evaluators have made in several independent assessments and seems to take a solely negative view of the recommendations made by those evaluators (References 6, 10, 11, 12). We view the identification of issues as positive opportunities that should be routinely sought, listened to, and acted upon. Where GAO sees "continuing problems," we see a measurable record of progress to date and a commitment to continuing improvement in the future. It is understood by the Department, by the NRC, and by knowledgeable outside observers that the repository program must meet rigorous quality assurance expectations for our license application to be acceptable to the Commission. The fact is, we are on schedule to submit our license application in December 2004, and we have an effective quality assurance program in place that will enable us to meet that objective. In summary, we have demonstrated steady and significant progress. We initiated the MII in 2002 to provide a special focus on specific improvement targets; we achieved the objectives of MII and have transitioned improvement initiatives to day-to-day management. Our continuous improvement culture means that we expect progress to continue, and our performance metrics enable us to assess that progress and direct management attention as needed. Based on these facts, we cannot concur with the findings and recommendations of the draft report. I urge GAO to further examine available information about our quality assurance program, performance indicators, safety conscious work environment, and other relevant aspects of the Program. Some highly pertinent information that was available during the time of GAO's audit, between April 2003 and March 2004, is not reflected in the draft report. More recent documentation - for instance, the MII Independent Review Report, which was published on March 19, 2004 - is also significant. The enclosed list of references identifies documents that, we believe, are critical for GAO to review and fully consider prior to working further on the draft report. Without full consideration of this information, GAO's findings on the Department's progress in addressing quality assurance issues are incomplete, and its conclusions are broadly inaccurate. I strongly urge you to review and incorporate additional information in your final report. You are welcome to revisit our offices, and we will provide any documentation you may require. Sincerely, Signed by: Margaret S.Y. Chu, Ph.D. Director: Office of Civilian Radioactive Waste Management: Enclosure: ENCLOSURE: LIST OF REFERENCES: 1. DOE/NRC Quarterly Quality Assurance Meeting report, February 18, 2004. 2. Metric Definition Sheet 2.5.1.2 (part of Yucca Mountain Project Performance Indicators Database), March 2004. 3. International Survey Research, Survey Summary Report, October 2003. 4. Memorandum, John Arthur to Margaret Chu, April 2, 2004. 5. Management Improvement Initiatives Transition Approach, December 2003. 6. Longenecker and Associates, Inc., Management Improvement Initiatives Independent Review Report, March 19, 2004. 7. Letter, Margaret Chu to Martin Virgilio, April 5, 2004. 8. Yucca Mountain Project Performance Indicators Database, ongoing internal management tool. 9. Office of Repository Development, Monthly Operating Review Annunciator Panel, ongoing internal management tool. 10. DOE Office of Independent Oversight and Performance Assurance, Management Assessment of the Office of Repository Development, November 2003. 11. D.L. English Consulting, Inc., FY 2003 Quality Assurance Management Assessment of the Office of Civilian Radioactive Waste Management, November 2003. 12. Booz Allen Hamilton, Performance Management Assessment: DOE Office of Civilian Radioactive Waste Management, September 30, 2003. The following are GAO's comments on the Department of Energy's letter dated April 19, 2004. GAO Comments: 1. We disagree. Our report states that the recent independent assessments have shown improvements in the key management areas identified in the 2002 corrective action plan. However, the assessments also showed that problems remain in these areas and thus the corrective actions have not yet been successful in correcting these weaknesses. DOE's examples of progress illustrate our point regarding improperly specified goals. For example, DOE states in its comments that line management's self-identification of conditions adverse to quality has increased approximately 100 percent in the last 15 months (as opposed to the identification of such conditions by quality assurance auditors). However, despite this seemingly dramatic increase, DOE has yet to meet its goal of line management's self-identifying 80 percent of all quality problems. (DOE's 100 percent increase brought them up to about 50 percent of all quality problems being self-identified by line managers.) Further, the goal continues to lack a time frame for when the 80 percent goal should be attained and for how long it should be sustained before the corrective actions can be judged successful. As our report points out, without such specificity, improvements cannot be evaluated in terms of overall success. 2. We disagree. The 2002 Management Improvement Initiatives clearly state that it was a "comprehensive corrective action plan necessary to address weaknesses in the implementation of [DOE's] quality assurance requirements and attain a level of performance expected of an NRC license applicant." Contrary to DOE's assertion, the initiative does not indicate it was a "springboard effort to address management issues and transition improvements into day-to-day line management activities." Although the transitioning of improvements to the line is laudable, the initiative focused on implementing corrective actions and evaluating the effectiveness of the actions in correcting problems. This approach is consistent with DOE's criteria for correcting significant conditions adverse to quality, and it is the criteria we relied on to determine whether the corrective actions specified in the initiatives were successful. 3. We agree. We did not include the full range of performance indicators (goals) that have recently been developed, and continue to change, to assess the 2002 plan. Instead, of the hundreds of indicators that are being developed to manage the project, we relied on those few that Bechtel officials told us were connected to the goals of the 2002 plan. As table 1 shows, some improvements have been made in specifying the quantitative aspects of the goals, but weaknesses continue to exist in the new goals. In fact, table 1 shows that DOE no longer has a goal in its performance tool that specifically tracks the trend in problems related to roles and responsibilities. This omission is particularly important because the area of roles and responsibilities was noted in the 2002 plan as one of the biggest sources of problems in the quality assurance process, and, as the recent assessments have found, this is an area with continuing problems. 4. We disagree. We acknowledge that these reviews found positive improvements in a number of management areas. However, we also note that continuing problems were found with management weaknesses despite all corrective actions having been implemented in 2003. 5. While DOE believes that it has achieved the objectives of the 2002 plan, it lacks evidence that its actions have been effective in addressing the management weaknesses and correcting the recurring quality problems with data, models, and software. Evaluating performance against measurable goals with time frames for meeting and sustaining the goals would provide the needed evidence. 6. The draft report that we sent to DOE for review included reviews of 9 of the 12 documents listed in the enclosure of DOE's letter. We have since reviewed the 3 remaining documents. The information in the 3 documents did not change our assessment of DOE's efforts to correct its quality assurance program. After full consideration of the information included in DOE's comments, we believe that our findings are complete and our conclusions are accurate. [End of section] Appendix V: Comments from the Nuclear Regulatory Commission: UNITED STATES NUCLEAR REGULATORY COMMISSION WASHINGTON, D.C. 20555-0001: April 16, 2004: Ms. Robin M. Nazarro Director, Science Issues Natural Resources and Environment United States General Accounting Office 441 G Street, NW: Washington, D.C. 20345: Dear Ms. Nazarro; I would like to thank you for the opportunity to review end submit comments on the draft report, "YUCCA MOUNTAIN; Persistent Quality Assurance Problems Could Delay Repository Licensing and Operation" (GAO-04-460). The U.S. Nuclear Regulatory Commission (NRC) appreciates the time and effort that you and your staff have taken to review this important topic. The NRC agrees with the GAO conclusion that the Department of Energy (DOE) should continue to improve the Quality Assurance Program for the proposed Yucca Mountain Repository. With respect to the specific GAO recommendations contained in the draft report, NRC suggests that DOE be given the flexibility to choose alternative approaches to achieve and measure improved Quality Assurance Program performance, since alternatives may be more suitable for the situation as DOE nears, then moves beyond, submittal of the license application. Two minor clarifying comments ore the draft report are enclosed. If you have any questions, please contact Mr. Tom Matuia at (301) 415-6700 or Mr. Ted Carter at (301) 415-6684, of my staff. Sincerely, Signed by: William D. Travers Executive Director for Operations: Enclosure: Specific Comments on Draft Report GAO-04-460: The following is GAO's comment on the U.S. Nuclear Regulatory Commission's letter dated April 16, 2004. GAO Comment: 1. We agree that alternative approaches could be used to measure performance; however, to ensure the success of any approaches, DOE must include objective measurements and time frames for reaching and sustaining desired performance and include an end point for closing out the corrective action plan. [End of section] Appendix VI: GAO Contact and Staff Acknowledgments: GAO Contact: Daniel Feehan, (303) 572-7352: Staff Acknowledgments: In addition to the individual named above, Robert Baney, Lee Carroll, Thomas Kingham, Chalane Lechuga, Jonathan McMurray, Judy Pagano, Katherine Raheb, Anne Rhodes-Kline, and Barbara Timmerman made key contributions to this report. (360268): FOOTNOTES [1] DOE and its subcontractor, Navarro Quality Services, which is a division of Navarro Research and Engineering, Inc., are responsible for carrying out quality assurance oversight activities, including conducting audits. DOE's primary contractor at the site, Bechtel/SAIC Company, LLC, is responsible for implementing DOE's quality assurance requirements related to ongoing project activities and for conducting audits of line activities. [2] Department of Energy, Office of Civilian Radioactive Waste Management, Management Improvement Initiatives (Washington, D.C.: July 19, 2002). [3] U.S. General Accounting Office, Nuclear Waste: Preliminary Observations on the Quality Assurance Program at the Yucca Mountain Repository, GAO-03-826T (Washington, D.C.: May 28, 2003). [4] U.S. General Accounting Office, Nuclear Waste: Repository Work Should Not Proceed Until Quality Assurance Is Adequate, GAO/RCED-88-159 (Washington, D.C.: Sept. 29, 1988). [5] Department of Energy, Office of Civilian Radioactive Waste Management, Report for Performance-Based Audit OQAP-BSC-03-14 of Technical Product Inputs at Bechtel SAIC Company, LLC, September 8-19, 2003 (Las Vegas, NV: Nov. 6, 2003). [6] Department of Energy, Office of Civilian Radioactive Waste Management, Report for Performance-Based Audit OQAP-BSC-03-10 of Analysis Model Report Processes and Products at Bechtel SAIC Company, LLC, October 21-31, 2003 (Las Vegas, NV: Jan. 20, 2004). [7] Department of Energy, Office of Civilian Radioactive Waste Management, Report for Audit OQAP-BSC-03-07 of Software and Software Activities at Bechtel SAIC Company, LLC, Lawrence Berkeley National Laboratory, and Lawrence Livermore National Laboratory, June 3-13, 2003 (Las Vegas, NV: Aug. 13, 2003). [8] Booz, Allen, Hamilton, Inc., Performance Management Assessment: DOE Office of Civilian Radioactive Waste Management (Las Vegas, NV: Sept. 30, 2003). [9] D.L. English Consulting, Inc., FY 2003 Quality Assurance Management Assessment of the Office of Civilian Radioactive Waste Management (South Dartmouth, MA: October 2003). [10] Department of Energy, Office of Independent Oversight and Performance Assurance, Management Assessment: Office of Repository Development (Washington, D.C.: November 2003). [11] International Survey Research, OCRWM 2003 Safety Conscious Work Environment Survey (Walnut Creek, CA: Nov. 7, 2003). [12] U.S. Nuclear Regulatory Commission, U.S. Nuclear Regulatory Commission Staff Evaluation of U.S. Department of Energy Analysis Model Reports, Process Controls, and Corrective Actions (Washington, D.C.: Apr. 7, 2004). [13] U.S. General Accounting Office, Internal Control Standards: Internal Control Management and Evaluation Tool, GAO-01-1008G (Washington, D.C.: August 2001). [14] Executive Office of the President, Office of Management and Budget, Circular No. A-123 (Washington, D.C.: June 24, 1995). [15] Department of Energy, Office of Civilian Radioactive Waste Management, Management Improvement Initiatives Transition Approach, Revision 1 (Washington, D.C.: December 2003). [16] Longenecker & Associates, Inc., under contract to Booz Allen Hamilton, Inc., OCRWM Management Improvement Initiatives (MII) Independent Review Report (Las Vegas, NV: Mar. 19, 2004). [17] Nuclear Regulatory Commission, Office of Nuclear Material Safety and Safeguards, Yucca Mountain Review Plan Final Report, NUREG-1804, Revision 2 (Washington, D.C.: July 2003). [18] U.S. Code of Federal Regulations, Title 10, Part 63. [19] A 4th year can be added to the process if NRC decides that the additional time is needed for hearings. GAO's Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. General Accounting Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.