Information Technology

Census Bureau Needs to Improve Its Risk Management of Decennial Systems Gao ID: GAO-08-259T December 11, 2007

For Census 2010, automation and information technology (IT) are expected to play a critical role. The Census Bureau plans to spend about $3 billion on automation and technology that are to improve the accuracy and efficiency of census collection, processing, and dissemination. From February 2006 through June 2009, the Bureau is holding a ''Dress Rehearsal'' during which it plans to conduct operational testing that includes decennial systems acquisitions. In October 2007, GAO reported on its review of four key 2010 Census IT acquisitions to (1) determine the status and plans, including schedule and cost, and (2) assess whether the Bureau is adequately managing associated risks. This testimony summarizes GAO's report on these key acquisitions and describes GAO's preliminary observations on the performance of handheld mobile computing devices used during the Dress Rehearsal.

As of October 2007, three key systems acquisitions for the 2010 Census were in process, and a fourth contract had recently been awarded. The ongoing acquisitions showed mixed progress in meeting schedule and cost estimates. Two of the projects were not on schedule. The award of the fourth contract, originally scheduled for 2005, was awarded in September 2007. In addition, one project had incurred cost overruns and increases to its projected life-cycle cost. As a result of the schedule changes, the full complement of systems and functionality that were originally planned will not be available for upcoming Dress Rehearsal operational testing. This limitation increases the importance of further system testing to ensure that the decennial systems work as intended. The Bureau's project teams for each of the four IT acquisitions had performed many practices associated with establishing sound and capable risk management processes, but critical weaknesses remained. Three project teams had developed a risk management strategy that identified the scope of the risk management effort. However, not all project teams had identified risks, established mitigation plans, or reported risks to executive-level officials. For example, one project team did not adequately identify risks associated with performance issues experienced by handheld mobile computing devices, even though Census field staff reported slow and inconsistent data transmissions with the device during the spring Dress Rehearsal operations. The magnitude of these difficulties is not clear, and the Bureau has not fully specified how it plans to measure the performance of the devices. Until the project teams implement key risk management activities, they face an increased probability that decennial systems will not be delivered on schedule and within budget or perform as expected.



GAO-08-259T, Information Technology: Census Bureau Needs to Improve Its Risk Management of Decennial Systems This is the accessible text file for GAO report number GAO-08-259T entitled 'Information Technology: Census Bureau Needs to Improve Its Risk Management of Decennial Systems' which was released on December 11, 2007. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Testimony: Testimony Before the Subcommittee on Information Policy, Census, and National Archives, Committee on Oversight and Government Reform, U.S. House of Representatives: United States Government Accountability Office: GAO: For Release on Delivery Expected at 2:00 p.m. EST: Tuesday, December 11, 2007: Information Technology: Census Bureau Needs to Improve Its Risk Management of Decennial Systems: Statement of David A. Powner: Director, Information Technology Management Issues: Mathew J. Scirè: Director, Strategic Issues: GAO-08-259T: GAO Highlights: Highlights of GAO-08-259T, a testimony before the Subcommittee on Information Policy, Census, and National Archives, Committee on Oversight and Government Reform, U.S. House of Representatives. Why GAO Did This Study: For Census 2010, automation and information technology (IT) are expected to play a critical role. The Census Bureau plans to spend about $3 billion on automation and technology that are to improve the accuracy and efficiency of census collection, processing, and dissemination. From February 2006 through June 2009, the Bureau is holding a ’Dress Rehearsal“ during which it plans to conduct operational testing that includes decennial systems acquisitions. In October 2007, GAO reported on its review of four key 2010 Census IT acquisitions to (1) determine the status and plans, including schedule and cost, and (2) assess whether the Bureau is adequately managing associated risks. This testimony summarizes GAO‘s report on these key acquisitions and describes GAO‘s preliminary observations on the performance of handheld mobile computing devices used during the Dress Rehearsal. What GAO Found: As of October 2007, three key systems acquisitions for the 2010 Census were in process, and a fourth contract had recently been awarded. The ongoing acquisitions showed mixed progress in meeting schedule and cost estimates. Two of the projects were not on schedule. The award of the fourth contract, originally scheduled for 2005, was awarded in September 2007. In addition, one project had incurred cost overruns and increases to its projected life-cycle cost. As a result of the schedule changes, the full complement of systems and functionality that were originally planned will not be available for upcoming Dress Rehearsal operational testing. This limitation increases the importance of further system testing to ensure that the decennial systems work as intended. The Bureau‘s project teams for each of the four IT acquisitions had performed many practices associated with establishing sound and capable risk management processes, but critical weaknesses remained. Three project teams had developed a risk management strategy that identified the scope of the risk management effort. However, not all project teams had identified risks, established mitigation plans, or reported risks to executive-level officials. For example, one project team did not adequately identify risks associated with performance issues experienced by handheld mobile computing devices, even though Census field staff reported slow and inconsistent data transmissions with the device during the spring Dress Rehearsal operations. The magnitude of these difficulties is not clear, and the Bureau has not fully specified how it plans to measure the performance of the devices. Until the project teams implement key risk management activities, they face an increased probability that decennial systems will not be delivered on schedule and within budget or perform as expected. Performance of Risk Management Activities by Key Census Acquisition Projects: Specific practices: Preparing for risk management: Determine risk sources and categories; Acquisition project 1: practice not implemented; Acquisition project 2: practice fully implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice fully implemented. Specific practices: Preparing for risk management: Define risk parameters; Acquisition project 1: practice fully implemented; Acquisition project 2: practice fully implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice fully implemented. Specific practices: Preparing for risk management: Establish and maintain a risk management strategy; Acquisition project 1: practice partially implemented; Acquisition project 2: practice fully implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice fully implemented. Specific practices: Preparing for risk management: Identify and involve the relevant stakeholders; Acquisition project 1: practice partially implemented; Acquisition project 2: practice partially implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice partially implemented. Specific practices: Identify and analyze risks: Identify and document the risks; Acquisition project 1: practice fully implemented; Acquisition project 2: practice partially implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice partially implemented. Specific practices: Identify and analyze risks: Evaluate, categorize, and prioritize risks; Acquisition project 1: practice partially implemented; Acquisition project 2: practice fully implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice fully implemented. Specific practices: Mitigate risks: Develop risk mitigation plans; Acquisition project 1: practice partially implemented; Acquisition project 2: practice partially implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice not implemented. Specific practices: Mitigate risks: Monitor status and implement risk mitigation plans; Acquisition project 1: practice partially implemented; Acquisition project 2: practice partially implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice partially implemented. Specific practices: Executive oversight: Review status with executive- level management; Acquisition project 1: practice not implemented; Acquisition project 2: practice not implemented; Acquisition project 3: practice fully implemented; Acquisition project 4: practice fully implemented. Source: GAO analysis of Census project data against industry standards. [End of table] What GAO Recommends: In its report, GAO made recommendations that the Bureau strengthen its systems testing and risk management activities, including risk identification and oversight. The Bureau agreed to examine additional ways to manage risks, but disagreed with the view that a full complement of systems would not be tested, stating it planned to do so during the Dress Rehearsal or later; however, the test plans have not been finalized and it remains unclear whether this testing will be done. To view the full product, including the scope and methodology, click on [hyperlink, http://www.GAO-08-259T]. For more information, contact David A. Powner at (202) 512-9286 or pownerd@gao.gov. [End of section] Mr. Chairman and Members of the Subcommittee: Thank you for the opportunity to participate in today‘s hearing on the 2010 Decennial Census Information Technology (IT) acquisitions that are an integral part of the reengineered census. As you know, the decennial census is mandated by the U.S. Constitution and provides data that are vital to the nation. These data are used to reapportion the seats of the U.S. House of Representatives, realign the boundaries of the legislative districts of each state, allocate billions of dollars in federal financial assistance, and provide a social, demographic, and economic profile of the nation‘s people to guide policy decisions at each level of government. Carrying out the census is the responsibility of the Department of Commerce‘s Census Bureau, which is now preparing for the 2010 Census. The Bureau is required to count the population on April 1, 2010, and the Secretary of Commerce is required to report to the President on the tabulation of total population by state within 9 months of that date. [Footnote 1] The Bureau plans to rely on automation and technology to improve the coverage, accuracy, and efficiency of the 2010 Census, and has awarded four key IT contracts to that end. It is also holding what it refers to as a Dress Rehearsal, from February 2006 through June 2009, a period centering around a mock Census Day on April 1, 2008. [Footnote 2] Planned Dress Rehearsal activities include operational testing of the 2010 Census systems in a census-like environment. The Bureau estimates that its IT acquisitions will absorb about $3 billion of the total $11.5 billion cost of the entire census. As requested, our testimony today will summarize our report on the four key IT acquisitions. In the report, we (1) determined the status and plans, including schedule and costs, for four key IT acquisitions; and (2) assessed whether the Bureau is adequately managing the risks facing these key system acquisitions. [Footnote 3] The report contains a detailed overview of the scope and methodology we used. As you also requested, our testimony today describes GAO‘s preliminary observations on the performance of handheld mobile computing devices used during address canvassing activities in the Dress Rehearsal. [Footnote 4] The preliminary observations presented in this report are based on field work we have conducted at the two Dress Rehearsal sites (Stockton, CA and Fayetteville, NC), as well as a review of Bureau documentation of its own observations of the Dress Rehearsal. The work on which this testimony is based was performed in accordance with generally accepted government auditing standards. Results in Brief: As of October 2007, three key systems acquisitions for the 2010 Census were in process, and a fourth contract had recently been awarded: * In one project, the Bureau is modernizing the database that provides address lists, maps, and other geographic support services for the census. This project is on schedule to complete improvements by the end of fiscal year 2008 and is meeting cost estimates. * In a second project, the Bureau is acquiring systems, equipment, and infrastructure for field staff to use in collecting census data. Deliverables provided to date include handheld mobile computing devices and installation of key support infrastructure. However, the schedule for this acquisition has been revised, resulting in delays in system development and testing of interfaces. Also, the life-cycle cost estimates for this program have increased, and we projected an $18 million cost overrun by December 2008. According to the contractor, the overrun is due primarily to an increase in the number of system requirements. * In a third project, the Bureau is acquiring a system for integrating paper, telephone responses, and field operations. The software development and testing are on schedule to provide (by December 2007) an initial system to process the major census forms during the Dress Rehearsal activities. However, the system development schedule was revised in October 2005, which is delaying some functionality. For example, a telephone-assistance system that was originally intended to be completed by fiscal year 2008 has been delayed. This acquisition is meeting current cost estimates. * Finally, a contract to replace the current system used to tabulate and disseminate census data was recently delayed by about a year (it was ultimately awarded in September 2007). As a result, of the 1-year delay, the Dress Rehearsal activities will use the current tabulation and dissemination system rather than a modernized version. The delays mean that the Dress Rehearsal operational testing will take place without the full complement of systems and functionality that was originally planned. As a result, further system testing will be necessary to ensure that the decennial systems work as intended. However, as of October 2007, Bureau officials had not finalized their plans for testing all the systems, and it is not clear whether these plans would include testing to address all interrelated systems and functionality, such as end-to-end testing. [Footnote 5] According to officials, these plans will not be finalized until February 2008. Without sufficient testing of all systems and their functionality, the Bureau increases the risk that costs will increase further, that decennial systems will not perform as expected, or both. As of October 2007, the four project teams managing the acquisitions had performed many practices associated with establishing sound and capable risk management processes. However, critical weaknesses remained. Specifically, three of the four project teams had developed risk management strategies identifying the scope of their risk management efforts; however, three project teams had weaknesses in identifying risks, establishing mitigation plans that identified planned actions and milestones, and reporting risk status to executive- level officials. For example, one project team did not adequately identify risks associated with performance issues experienced by handheld mobile computing devices. Further, in May and June 2007, both we and the Census Bureau observed the use of the handheld mobile computing device in Census-like conditions and these observations revealed a number of performance issues with the devices, such as slow and inconsistent data processing. The magnitude of these performance issues remains unclear. The Field Data Collection Automation (FDCA) contract anticipates the Bureau‘s need for data on the performance of the handheld mobile computing device; however, the Bureau has not fully specified the performance data it will use for the devices. As we have previously reported, a root cause of weaknesses in completing key risk management activities is the lack of policies for managing major acquisitions at the Bureau. [Footnote 6] Until the project teams implement key risk management activities, they face an increased probability that decennial systems will not be delivered on schedule and within budget or perform as expected. Because the entire complement of systems will not be available for Dress Rehearsal activities as originally planned, we recommended that the Census Bureau plan for and perform end-to-end testing so that all systems are tested in a census-like environment. Further, to help ensure that the three key acquisitions for the 2010 Census operate as intended, we recommended that the project teams strengthen risk management activities, including those associated with risk identification, mitigation, and oversight. In written comments on a draft of our report, the department agreed to examine additional ways to manage risks and prepare a formal action plan in response to our final report. However, the department said it had a major disagreement with our findings with regard to not conducting operational testing on a full complement of the key decennial systems, stating it plans to test all critical systems and interfaces during the Dress Rehearsal or later. Nonetheless, the Bureau‘s test plans have not been finalized, and it remains unclear whether testing will address all interrelated systems and functionality in a census-like environment, as would be provided by end-to-end testing. Consistent with our recommendation, following up with documented test plans to do end-to-end testing will help ensure that decennial systems will work as intended. Background: Conducting the decennial census is a major undertaking involving many interrelated steps including: * identifying and correcting addresses for all known living quarters in the United States (known as ’address canvassing“); * sending questionnaires to housing units; * following up with nonrespondents through personal interviews; * identifying people with nontraditional living arrangements; * managing a voluminous workforce responsible for follow-up activities; * collecting census data by means of questionnaires, calls, and personal interviews; * tabulating and summarizing census data; and: * disseminating census analytical results to the public. Role of IT in the Decennial Census: The Bureau estimates that it will spend about $3 billion on automation and IT for the 2010 Census, including four major systems acquisitions that are expected to play a critical role in improving coverage, accuracy, and efficiency. Figure 1 shows the key systems and interfaces supporting the 2010 Census, and highlights the four major IT systems we discuss today. As the figure shows, these four systems are to play important roles with regard to different aspects of the process. Figure 1: Key 2010 Census Systems and Interfaces: [See PDF for image] This figure is a schematic illustration of Key 2010 Census Systems and Interfaces. Systems and interfaces are depicted in three areas, with relationships indicated by connections through either one-way or two- way arrows. Four systems are shaded to indicate that the system is discussed in the report. The following data is depicted: Establish where to count: * Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Universe Control and Management; * Printing. Collect respondent information: * Decennial Response Integration System [system discussed in this report]; * Field Data Collection Automation; * Postal; * Respondent; * National Processing Center; * Decennial Applicant Personnel and Payroll System; * Census Evaluation and Experimentation (overlaps into Collect respondent information section). Provide results: * Data Access and Dissemination System [system discussed in this report]; * Response Processing System; * Archive System; * National Archives and Records Administration; * Census Evaluation and Experimentation (overlaps into Provide Results section). Relationships as indicated by connecting arrows (input and output): Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]: * Output arrow to Archive System; * Output arrow to Data Access and Dissemination System [system discussed in this report]; * Output arrow to Decennial Response Integration System [system discussed in this report]; * Output and Input arrow with Census Evaluation and Experimentation; * Output and Input arrow with Response Processing System; * Output and Input arrow with Universe Control and Management; * Output and Input arrow with Field Data Collection Automation [system discussed in this report]; * Input arrow from National Processing Center. Universe Control and Management: * Output and Input arrow with Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Output and Input arrow with Field Data Collection Automation [system discussed in this report]; * Output and Input arrow with Response Processing System; * Output arrow to Printing; * Output and Input arrow to Decennial Response Integration System [system discussed in this report]; * Output arrow to Census Evaluation and Experimentation; Printing: * Output arrow to Postal; * Input arrow from Decennial Response Integration System [system discussed in this report]; Decennial Response Integration System [system discussed in this report]: * Output arrow to Printing; * Output and Input arrow with Universe Control and Management; * Input arrow from Postal; * Input arrow from Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Output and Input arrow with Field Data Collection Automation [system discussed in this report]; * Output arrow to Response Processing System; * Output arrow to Archive System. Field Data Collection Automation [system discussed in this report]: * Output and Input arrow with Universe Control and Management; * Output and Input arrow with Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Output and Input arrow with Decennial Response Integration System [system discussed in this report]; * Input arrow from Respondent; * Input arrow from Decennial Applicant Personnel and Payroll System; * Output arrow to Census Evaluation and Experimentation (overlaps into Provide Results section); * Output arrow to National Processing Center. Decennial Applicant Personnel and Payroll System: * Output arrow to Field Data Collection Automation [system discussed in this report]. Postal: * Output arrow to Decennial Response Integration System [system discussed in this report]; * Input arrow from Printing; * Output and Input arrow with Respondent. Respondent: * Output and Input arrow with Postal; * Output arrow to Field Data Collection Automation [system discussed in this report]. National Processing Center: * Output arrow to Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Output arrow to Census Evaluation and Experimentation; * Input arrow from Field Data Collection Automation [system discussed in this report]. Census Evaluation and Experimentation: * Output and Input arrow with Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Input arrow from Universe Control and Management; * Input arrow from Field Data Collection Automation [system discussed in this report]; * Input arrow from National Processing Center; * Output and Input arrow with Response Processing System. Data Access and Dissemination System [system discussed in this report]: * Input arrow from Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Input arrow from Response Processing System. Archive System: * Input arrow from Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Input arrow from Decennial Response Integration System [system discussed in this report]; * Input arrow from Response Processing System; * Output arrow to National Archives and Records Administration. Response Processing System: * Output and Input arrow with Master Address File/Topologically Integrated Geographic Encoding and Referencing [system discussed in this report]; * Output and Input arrow with Universe Control and Management; * Output and Input arrow with Census Evaluation and Experimentation; * Input arrow from Decennial Response Integration System [system discussed in this report]; * Output arrow to Archive System; * Output arrow to Data Access and Dissemination System [system discussed in this report]. National Archives and Records Administration: * Input arrow from Archive System. Source: U.S. Census Bureau. [End of figure] To establish where to count (as shown in the top section of fig. 1), the Bureau will depend heavily on a database that provides address lists, maps, and other geographic support services. The Bureau‘s address list, known as the Master Address File (MAF), is associated with a geographic information system containing street maps known as the Topologically Integrated Geographic Encoding and Referencing (TIGER®) database. [Footnote 7] The MAF/TIGER database is the object of the first major IT acquisition”the MAF/TIGER Accuracy Improvement Project (MTAIP). To collect respondent information (a process depicted in the middle section of fig. 1), the Bureau is pursuing two initiatives. First, the Field Data Collection Automation (FDCA) program is expected to provide automation support for field data collection operations as well as reduce costs and improve data quality and operational efficiency. This acquisition includes the systems, equipment, and infrastructure that field staff will use to collect census data, such as handheld mobile computing devices. [Footnote 8] Second, the Decennial Response Integration System (DRIS) is to provide a system for collecting and integrating census responses from all sources, including forms, telephone interviews, and handheld mobile computing devices in the field. DRIS is expected to improve accuracy and timeliness by standardizing the response data and providing it to other Bureau systems for analysis and processing. To provide results (see the bottom section of fig. 1), the Data Access and Dissemination System II (DADS II) acquisition is to replace legacy systems for tabulating and publicly disseminating data. The DADS II program is expected to provide comprehensive support to DADS. Replacement of the legacy systems is expected to: * maximize the efficiency, timeliness, and accuracy of tabulation and dissemination products and services; * minimize the cost of tabulation and dissemination; and: * increase user satisfaction with related services. Table 1 provides a brief overview of the four acquisitions. Table 1: Four Key IT Acquisitions Supporting Census 2010: IT acquisition: MAF/TIGER Accuracy Improvement Project (MTAIP); Purpose: Modernize the system that provides the address list, maps, and other geographic support services for the Census and other Bureau surveys. IT acquisition: Field Data Collection Automation (FDCA); Purpose: Provide automated resources for supporting field data collection, including the provision of handheld mobile computing devices to collect data in the field, including address and map data. IT acquisition: Decennial Response Integration System (DRIS); Purpose: Provide a solution for data capture and respondent assistance. IT acquisition: Data Access and Dissemination System (DADS II); Purpose: Develop a replacement for the DADS legacy tabulation and dissemination systems. Source: GAO analysis of Census Bureau data. [End of table] Responsibility for these acquisitions lies with the Bureau‘s Decennial Management Division and the Geography Division. Each of the four acquisitions is managed by an individual project team staffed by Bureau personnel. Additional information on the contracts for these four systems is provided in appendix I of the report. In preparation for the 2010 Census, the Bureau plans a series of tests of its (new and existing) operations and systems in different environments, as well as to conduct what it refers to as the Dress Rehearsal. During the Dress Rehearsal period, which runs from February 2006 through June 2009, the Bureau plans to conduct development and testing of systems, run a mock Census Day, and prepare for Census 2010, which will include opening offices and hiring staff. These Dress Rehearsal activities are to provide an operational test of the available system functionalities in a census-like environment, as well as other operational and procedural activities. Decennial IT Acquisitions Were at Various Stages of Development and Showed Mixed Progress against Schedule and Cost Baselines: As of October 2007, three key decennial systems acquisitions were in process and a fourth contract had recently been awarded. The ongoing acquisitions (FDCA, DRIS) showed mixed progress in providing deliverables while adhering to planned schedules and cost estimates. The two ongoing projects had experienced schedule delays; the date for awarding the fourth contract was postponed several times. In addition, we estimated that one of the ongoing projects (FDCA) will incur about $18 million in cost overruns. In response to schedule delays as well as other factors, including cost, the Bureau made schedule adjustments and planned to delay certain system functionality. As a result, Dress Rehearsal operational testing will not address the full complement of systems and functionality that was originally planned, and the Bureau has not yet finalized its plans for further system tests. Delaying functionality increases the importance of operational testing after the Dress Rehearsal to ensure that the decennial systems work as intended. MTAIP Was Completing Improvements on Schedule and at Estimated Cost: MTAIP is a project to improve the accuracy of the MAF/TIGER database, which contains information on street locations, housing units, rivers, railroads, and other geographic features. We reported that MTAIP was on schedule to complete improvements by the end of fiscal year 2008 and was meeting cost estimates. As of October 2007, the acquisition was in the second and final phase of its life cycle. In Phase II, which began in January 2003 and is ongoing, the contractor is developing improved maps for all 3,037 counties in the United States. We reported that the contractor had delivered more than 75 percent of these maps, which are due by September 2008. Beginning in fiscal year 2008, maintenance for the contract will begin. The contract closeout activities are scheduled for fiscal year 2009. FDCA Had Provided Deliverables but Had Delayed Functionality and Was Experiencing Cost Increases: FDCA is to provide the systems, equipment, and infrastructure that field staff will use to collect census data. At the peak of the 2010 Census, about 4,000 field operations supervisors, 40,000 crew leaders, 500,000 enumerators and address listers, and several thousand office employees are expected to use or access FDCA. As of October 2007, the contractor was in the process of developing and testing FDCA software for the Dress Rehearsal Census Day, and had delivered 1,388 handheld mobile computing devices to be used in address canvassing for the Dress Rehearsal. Also, key FDCA support infrastructure had been installed, including the Security Operation Center. In future contract phases, the project will continue development, deploy systems and hardware, support census operations, and perform operational and contract closeout activities. However, the Bureau revised FDCA‘s original schedule and delayed or eliminated some of its key functionality from the Dress Rehearsal, including the automated software distribution system. According to the Bureau, it revised the schedule because it realized it had underestimated the costs for the early stages of the contract, and that it could not meet the contractor‘s estimated level of first-year funding because the fiscal year 2006 budget was already in place. According to the Bureau, this initial underestimate led to schedule changes and overall cost increases. According to the Bureau, FDCA was meeting all planned milestones on the revised schedule. For example, all sites for Regional Census Centers and Puerto Rico Area Offices had been identified. According to the Bureau, it is on schedule to open all these offices in January 2008. The project life-cycle costs had increased. At contract award in March 2006, the total cost of FDCA was estimated not to exceed $596 million. In May 2007, the life-cycle cost rose by a further $23 million because of increasing system requirements, which resulted in an estimated life- cycle cost of about $647 million. Table 2 shows the life-cycle cost estimates for FDCA as of October 2007. Table 2: FDCA Life-Cycle Cost Estimates: Execution period: Baseline planning period; Start date: March 31, 2006; End date: June 30, 2006; Cost estimates (in millions) September 2006: $11; Cost estimates (in millions) May 2007: $11. Execution period: Execution Period 1; Start date: July 1, 2006; End date: December 31, 2008; Cost estimates (in millions) September 2006: $200; Cost estimates (in millions) May 2007: $225. Execution period: Execution Period 2; Start date: January 1, 2009; End date: September 30, 2011; Cost estimates (in millions) September 2006: $319; Cost estimates (in millions) May 2007: $318. Execution period: Execution Period 3; Start date: August 1, 2010; End date: End of contract; Cost estimates (in millions) September 2006: $10; Cost estimates (in millions) May 2007: $10. Execution period: Leased equipment; Start date: N/A; End date: N/A; Cost estimates (in millions) September 2006: $12; Cost estimates (in millions) May 2007: $12. Execution period: Management reserve; Start date: N/A; End date: N/A; Cost estimates (in millions) September 2006: $7; Cost estimates (in millions) May 2007: $5. Execution period: Award fee; Start date: N/A; End date: N/A; Cost estimates (in millions) September 2006: $65; Cost estimates (in millions) May 2007: $65. Execution period: Total; Cost estimates (in millions) September 2006: $624; Cost estimates (in millions) May 2007: $647. Source: GAO analysis of Census Bureau data. Note: Total may not add due to rounding. [End of table] In addition, FDCA had already experienced $6 million in cost overruns, and both our analysis and the contractor‘s analysis expected FDCA to experience additional cost overruns. Based on our analysis of cost performance reports (from July 2006 to May 2007), we projected that the FDCA project will experience further cost overruns by December 2008. The FDCA cost overrun was estimated between $15 million and $19 million, with the most likely overrun to be about $18 million. The contractor, in contrast, estimated about a $6 million overrun by December 2008. According to the contractor, the major cause of projected cost overruns was the system requirements definition process. For example, in December 2006, the contractor noted a significant increase in the requirements for the Dress Rehearsal Paper Based Operations in Execution Period 1. According to the cost performance reports, this increase has meant that more work must be conducted and more staffing assigned to meet the Dress Rehearsal schedule. The Bureau agreed that cost increases occurred in some cases because of the addition of new requirements, most of which related to the security of IT systems, but added that in other cases, increases occurred from the process of the contractor converting high-level functional requirements into more detailed specific requirements. However, the process of developing detailed requirements from high-level functional requirements does not inevitably lead to cost increases if the functional requirements were initially well-defined. The FDCA schedule changes have increased the likelihood that the systems testing at the Dress Rehearsal will not be as comprehensive as planned. The inability to perform comprehensive operational testing of all interrelated systems increases the risk that further cost overruns will occur and that decennial systems will experience performance shortfalls. After a Schedule Revision, DRIS Was Delivering Reduced Functionality at Projected Cost: DRIS is to provide a system for collecting and integrating census responses, standardizing the response data, and providing it to other systems for analysis and processing. The DRIS functionality is critical for providing assistance to the public via telephone and for monitoring the quality and status of data capture operations. Although DRIS was currently on schedule to meet its December 2007 milestone, the Bureau revised the original DRIS schedule after the contract was awarded in October 2005. Under the revised schedule, the Bureau delayed or eliminated some functionality that was expected to be ready for the Dress Rehearsal mock Census Day. According to Bureau officials, they delayed the schedule and eliminated functionality for DRIS when they realized they had underestimated the fiscal years 2006 through 2008 costs for development. As shown in table 3, the government‘s funding estimates for DRIS Phase I were significantly lower than the contractor‘s. Table 3: DRIS Cost Estimates for Phase I (as of March 2006): Fiscal year: 2006; Cost estimates (in millions) Contractor: $18.6; Cost estimates (in millions) Government: $11.2. Fiscal year: 2007; Cost estimates (in millions) Contractor: $53.3; Cost estimates (in millions) Government: $23.8. Fiscal year: 2008; Cost estimates (in millions) Contractor: $48.7; Cost estimates (in millions) Government: $31.5. Fiscal Year: Total; Cost estimates (in millions) Contractor: $120.6; Cost estimates (in millions) Government: $66.5. Source: GAO analysis of Census Bureau data. [End of table] Originally, the DRIS solution was to include paper, telephone, Internet, and field data collection processing; selection of data capture sites; and preparation and processing of 2010 Census forms. However, the Bureau reduced the scope of the solution by eliminating the Internet functionality. In addition, the Bureau has stated that it will not have a robust telephone questionnaire assistance system in place for the Dress Rehearsal. As of October 2007, the Bureau was also delaying selecting sites for data capture centers, preparing data capture facilities, and recruiting and hiring data capture staff. Although Bureau officials told us that the revisions to the schedule should not affect meeting milestones for the 2010 Census, the delays mean that more systems development and testing will need to be accomplished later. Given the immovable deadline of the decennial census, the Bureau is at risk of reducing functionality or increasing costs to meet its schedule. The DRIS project was not experiencing cost overruns, and our analysis of cost performance reports from April 2006 to May 2007 projected no cost overruns by December 2008. As of May 2007, the DRIS contract value had not increased. DADS II Contract Had Recently Been Awarded after a Delay: The DADS II acquisition is to replace the legacy DADS systems, which tabulate and publicly disseminate data from the decennial census and other Bureau surveys. [Footnote 9] The DADS II contractor is also expected to provide comprehensive support to the Census 2000 legacy DADS systems. The DADS II contract award date had been delayed multiple times. The award date was originally planned for the fourth quarter of 2005, but the date changed to August 2006. On March 8, 2006, the Bureau estimated it would delay the award of the DADS II contract from August to October 2006 to gain a clearer sense of budget priorities before initiating the request for proposal process. The Bureau then delayed the contract award again by about another year. In January 2007, the Bureau released the DADS II request for proposal, and the contract was finally awarded in September 2007. Because of these delays, DADS II will not be developed in time for the Dress Rehearsal. Instead, the Bureau will use the legacy DADS system for tabulation during the Dress Rehearsal. Nonetheless, the Bureau plans to have the DADS II system available for the 2010 Census. The Bureau Was Making Progress in Risk Management Activities but Critical Weaknesses Remained: The project teams varied in the extent to which they followed disciplined risk management practices. For example, three of the four project teams had developed strategies to identify the scope of the risk management effort. However, three project teams had weaknesses in identifying risks, establishing adequate mitigation plans, and reporting risk status to executive-level officials. These weaknesses in completing key risk management activities can be attributed in part to the absence of Bureau policies for managing major acquisitions, as we described in an earlier report. [Footnote 10] Without effective risk management practices, the likelihood of project success is decreased. According to the Software Engineering Institute (SEI), the purpose of risk management is to identify potential problems before they occur. When problems are identified, risk-handling activities can be planned and invoked as needed across the life of a project in order to mitigate adverse impacts on objectives. Effective risk management involves early and aggressive risk identification through the collaboration and involvement of relevant stakeholders. Based on SEI‘s Capability Maturity Model® Integration (CMMI®), risk management activities can be divided into four key areas: * preparing for risk management; * identifying and analyzing risks; * mitigating risks, and: * executive oversight. The discipline of risk management is important to help ensure that projects are delivered on time, within budget, and with the promised functionality. It is especially important for the 2010 Census, given the immovable deadline. Project Teams Had Usually Established Risk Preparation Activities, but Some Improvements in These Activities Were Needed: Risk preparation involves establishing and maintaining a strategy for identifying, analyzing, and mitigating risks. The risk management strategy addresses the specific actions and management approach used to perform and control the risk management program. It also includes identifying and involving relevant stakeholders in the risk management process. Table 4 shows the status of the four project teams‘ implementation of key risk preparation activities as of October 2007. [Footnote 11] Table 4: Risk Management Preparation Activities Completed for the Key 2010 Census Systems: Specific practices: Determine risk sources and categories; MTAIP: practice not implemented; FDCA: practice fully implemented; DRIS: practice fully implemented; DADS: practice fully implemented. Specific practices: Define parameters used to analyze and categorize risks and parameters used to control risk management efforts; MTAIP: practice fully implemented; FDCA: practice fully implemented; DRIS: practice fully implemented; DADS: practice fully implemented. Specific practices: Establish and maintain the strategy to be used for risk management; MTAIP: practice partially implemented; FDCA: practice fully implemented; DRIS: practice fully implemented; DADS: practice fully implemented. Specific practices: Identify and involve the relevant stakeholders of the risk management process as planned; MTAIP: practice partially implemented; FDCA: practice partially implemented; DRIS: practice fully implemented; DADS: practice partially implemented. Source: GAO analysis of project data. [End of table] As the table shows, three project teams had established most of the risk management preparation activities. However, the MTAIP project team had implemented the fewest practices. The team did not adequately determine risk sources and categories or adequately develop a strategy for risk management. As a result, the project‘s risk management strategy was not comprehensive and did not fully address the scope of the risk management effort, including discussing techniques for risk mitigation and defining adequate risk sources and categories. In addition, three project teams (MTAIP, FDCA, and DADS II) had weaknesses regarding stakeholder involvement. The three teams did not provide sufficient evidence that the relevant stakeholders were involved in risk identification, analysis, and mitigation activities; reviewing the risk management strategy and risk mitigation plans; or communicating and reporting risk management status. These weaknesses can be attributed in part to the absence of Bureau policies for managing major acquisitions, as we described in our earlier reports. [Footnote 12] Without adequate preparation for risk management, including establishing an effective risk management strategy and identifying and involving relevant stakeholders, project teams cannot properly control the risk management process. The Project Teams Had Identified and Analyzed Risks but Not All Key Risks Were Identified: Risks must be identified and described in an understandable way before they can be analyzed and managed properly. This includes identifying risks from both internal and external sources and evaluating each risk to determine its likelihood and consequences. Table 5 shows the status of the four project teams‘ implementation of key risk identification and evaluation activities at the time of our October 2007 report. Table 5: Risk Identification and Evaluation Activities Completed for the Key 2010 Census Systems: Specific practices: Identify and document the risks; MTAIP: practice fully implemented; FDCA: practice partially implemented; DRIS: practice fully implemented; DADS: practice partially implemented. Specific practices: Evaluate and categorize each identified risk using the defined risk categories and parameters, and determine its relative priority; MTAIP: practice partially implemented; FDCA: practice fully implemented; DRIS: practice fully implemented; DADS: practice fully implemented. Source: GAO analysis of project data. [End of table] As of July 2007, the MTAIP and DRIS project teams were adequately identifying and documenting risks, including system interface risks. For example, the MTAIP project team identified significant risks regarding potential changes in funding and the turnover of contractor personnel as the program nears maturity, and the DRIS project team identified significant risks regarding new system security regulations, changes or increases to Phase II baseline requirements, and new interfaces after Dress Rehearsal. In contrast, the FDCA project team had not identified or documented any significant risks related to the handheld computers that will be used in the 2010 Census, despite problems arising during the Dress Rehearsal. The computers are designed to automate operations for field staff and eliminate the need to print millions of paper questionnaires and maps used by temporary field staff to conduct address canvassing and nonresponse follow-up. Automating operations may allow the Bureau to reduce the cost of operations; thus, it is critical that the risks surrounding the use of the handheld computers be closely monitored and effectively managed to ensure their success. However, the Bureau has not identified or documented risks associated with a variety of handheld computers performance problems that we identified through field work conducted at your request. Specifically, we found that during Dress Rehearsal activities between May 2007 and June 2007, as the Bureau tested a prototype of the handheld computers, field staff experienced multiple problems. For example, the field staff told us that they experienced slow and inconsistent data transmissions from the handheld computers to the central data processing center. The field staff reported the device was slow to process addresses that were a part of a large assignment area. Bureau staff reported similar problems with the handheld computers in observation reports, help desk calls, and debriefing reports. In addition, our own analysis of Bureau documentation revealed problems with the handheld computers: * Bureau observation reports revealed that the Bureau most frequently observed problems with slow processing of addresses, large assignment areas, and transmission. * The help desk call log revealed that field staff most frequently reported issues with transmission, the device freezing, map spotting and assignment areas. * Debriefing reports illustrated the impact of the handheld mobile computing problems on address canvassing. For example, one participant commented that the field staff struggled to find solutions to problems and wasted precious time in replacing the devices. * A time-and-motion study conducted by the Census Bureau indicated that field staff reported significant downtime in two test locations”about 23 percent in one location and about 27 percent in another location. The study, which is a draft that is subject to change, also described occurrences of failed transmissions and field staff attempts to resolve transmission problems. Collectively, the observation reports, help desk calls, debriefing reports, and time-and-motion study raised serious questions about the performance of the handheld computers during the address canvassing operation. According to the Bureau, the contractor has used these indicators to identify and address underlying problems during the Dress Rehearsal. Still, the magnitude of handheld computers performance issues throughout the Dress Rehearsal remains unclear. For example, the Bureau received analyses from the contractor on average transmission times. However, the contractor has not provided analyses that show the full range of transmission times, nor how this may have changed throughout the entire operation. In addition, the Bureau has not fully specified how it will measure performance of the handheld computers, even though the FDCA contract anticipates the Bureau‘s need for data on the performance of the handheld computers. The FDCA contract outlines the type of data the contractor will provide the Bureau on the performance of the handheld computers. Specifically, sections of the FDCA contract require the handheld computers to have a transmission log with what was transmitted, the date, time, user, destination, content/data type, and the outcome status. Another section of the Bureau‘s FDCA contract states that the FDCA contractor shall provide near real time reporting and monitoring of performance metrics and a ’control panel/dash board“ application to visually report those metrics from any Internet enabled PC. However, the contractor and the Bureau are not using a dashboard for Dress Rehearsal activities. Rather, during the Dress Rehearsal, the Bureau plans to identify what data and performance they would need for tracking the performance of the handheld computers in 2010 operations. In order for the Bureau to ensure that the FDCA handheld computers are ready for full scale operations, it will have to identify risks on a tight time frame. We recommended in a report on the Bureau‘s earlier version of the handheld computers that the Bureau define specific, measurable performance requirements for the handheld computer and other census-taking activities that address such important measures as productivity, cost savings, reliability, durability, and that the Bureau test the device‘s ability to meet those requirements in 2006. [Footnote 13] We also recommended in a March 2006 testimony that the Bureau validate and approve FDCA baseline requirements. [Footnote 14] The Bureau is working within a compressed time frame. By law, the decennial census must occur on April 1, 2010, and the results must be submitted to the President in December 2010. These dates cannot be altered, even if preparations are delayed. Access to real-time performance metrics via a ’control panel/dash board“ would assist Bureau management in assessing the handheld computer‘s performance and maximize the amount of time the Bureau and the contractor would have to remedy any problems identified during operations. Further, the Bureau‘s tight 2010 Decennial Operations Schedule allows little time for fixing problems with the device, raising the importance of the Bureau‘s access to these performance indicators. Such data would help fully inform stakeholders of the risks associated with the handheld computer, and allow project teams to develop mitigation activities to help avoid, reduce, and control the probability of these risks occurring. Finally, the FDCA and DADSII project teams did not provide evidence that specific system interface risks are being adequately identified to ensure that risk handling activities will be invoked should the systems fail during 2010 Census. For example, although the DADS II will not be available for the Dress Rehearsal, the project team did not identify any significant interface risks associated with this system. One reason for these weaknesses, as mentioned earlier, is the lack of Bureau policies for managing major acquisitions. If risks are not adequately identified and analyzed, management may be prevented from monitoring and tracking risks, and taking the appropriate mitigation actions, increasing the probability that the risks will materialize and magnifying the extent of damage incurred in such an event. Three of Four Project Teams‘ Risk Mitigation Plans and Monitoring Activities Were Incomplete: Risk mitigation involves developing alternative courses of action, workarounds, and fallback positions, with a recommended course of action for the most important risks to the project. Mitigation includes techniques and methods used to avoid, reduce, and control the probability of occurrence of the risk; the extent of damage incurred should the risk occur; or both. Table 6 shows the status of the four project teams‘ implementation of key risk mitigation activities. Table 6: Risk Mitigation Activities Completed for Key 2010 Census Systems: Specific practices: Develop a risk mitigation plan for the most important risks to the project, as defined by the risk management strategy; MTAIP: practice partially implemented; FDCA: practice partially implemented; DRIS: practice fully implemented; DADS: practice not implemented. Specific practices: Monitor the status of each risk periodically and implement the risk mitigation plan as appropriate; MTAIP: practice partially implemented; FDCA: practice partially implemented; DRIS: practice fully implemented; DADS: practice partially implemented. Source: GAO analysis of project data. [End of table] Three project teams (MTAIP, FDCA, and DADS II) had developed mitigation plans that were often untimely or included incomplete activities and milestones for addressing the risks. Some of these untimely and incomplete activities and milestones included the following: * The FDCA project team had developed mitigation plans for the most significant risks, but the plans did not always identify milestones for implementing mitigation activities. Moreover, the plans did not identify any commitment of resources, several did not establish a period of performance, and the team did not always update the plans with the latest information on the status of the risk. In addition, the FDCA project team did not provide evidence of developing mitigation plans to handle the other significant risks as described in their risk mitigation strategy. (These risks included a lack of consistency in requirements definition and insufficient FDCA project office staffing levels). * The mitigation plans for DADS II were incomplete, with no associated future milestones and no evidence of continual progress in working towards mitigating a risk. In several instances, DADS II mitigation plans were listed as ’To Be Determined.“ With regard to the second practice in the table (periodically monitoring risk status and implementing mitigation plans), the MTAIP, FDCA, and DADS II project teams were not always implementing the mitigation plans as appropriate. For example, although the MTAIP project team has periodically monitored the status of risks, it mitigation plans do not include detailed action items with start dates and anticipated completion dates; thus, the plans do not ensure that mitigation activities are implemented appropriately and tracked to closure. The FDCA and DADS II project teams did not identify system interface risks nor prepare adequate mitigation plans to ensure that systems will operate as intended. Because they did not develop complete mitigation plans, the MTAIP, FDCA, and DADS II project teams cannot ensure that for a given risk, techniques and methods will be invoked to avoid, reduce, and control the probability of occurrence. Project Teams Were Inconsistent in Reporting Risk Status to Executive- Level Management: Reviews of the project teams‘ risk management activities, status, and results should be held on a periodic and event-driven basis. The reviews should include appropriate levels of management, such as key Bureau executives, who can provide visibility into the potential for project risk exposure and appropriate corrective actions. Table 7 shows the status of the four project teams‘ implementation of activities for senior-level risk oversight at the time of our prior report. Table 7: Executive-Level Risk Oversight Activities Completed for the Key 2010 Decennial Systems: Specific practices: Review the activities, status, and results of the risk management process with executive-level management, and resolve issues; MTAIP: practice not implemented; FDCA: practice not implemented; DRIS: practice fully implemented; DADS: practice fully implemented. Source: GAO analysis of project data. [End of table] The project teams were inconsistent in reporting the status of risks to executive-level officials. DRIS and DADS II did regularly report risks; however, the FDCA and MTAIP projects did not provide sufficient evidence to document that these discussions occurred or what they covered. Failure to report a project‘s risks to executive-level officials reduces the visibility of risks to executives who should be playing a role in mitigating them. Implementation of GAO Recommendations Should Help Improve the Bureau‘s Risk Management: To help ensure that the Bureau‘s four key acquisitions for the 2010 Census operate as intended, we made several recommendations in our report. First, to ensure that the Bureau‘s decennial systems are fully tested, we recommended that the Secretary of Commerce require the Director of the Census Bureau to direct the Decennial Management Division and Geography Division to plan for and perform end-to-end testing so that the full complement of systems are tested in a census- like environment. In written comments on a draft of our final report, the department disagreed with our findings that a full complement of systems would not be tested, stating it plans to do so during the Dress Rehearsal or later. Nonetheless, the Bureau‘s test plans have not been finalized, and it remains unclear whether testing will address all interrelated systems and functionality in a census-like environment, as would be provided by end-to-end testing. Consistent with our recommendation following up with documented test plans to do end-to-end testing will help ensure that decennial systems will work as intended. Further, we recommended that the Secretary direct the Director of the Census Bureau to ensure that project teams strengthen risk management activities associated with risk identification, mitigation, and oversight. The department agreed to examine additional ways to manage risks and is working on an action plan to strengthen risk management activities. In summary, the IT acquisitions planned for 2010 Census will require continued oversight to ensure that they are achieved on schedule and at planned cost levels. Although, as of October 2007, the MTAIP and DRIS acquisitions were currently meeting cost estimates, FDCA was not. In addition, while the Bureau was making progress developing systems for the Dress Rehearsal, it was deferring certain functionality, with the result that the Dress Rehearsal operational testing would address less than a full complement of systems. Delaying functionality increases the importance of later development and testing activities, which will have to occur closer to the census date. It also raises the risk of cost increases, given the immovable deadline for conducting the 2010 Census. Further, the Bureau‘s project teams for each of the four acquisitions had implemented many practices associated with establishing sound and capable risk management processes, but they were not always consistent: the teams had not always identified risks, developed complete risk mitigation plans, or briefed senior-level officials on risks and mitigation plans. At this stage, we are particularly concerned about managing the risks associated with the handheld mobile computing devices, the numerous systems interfaces, and the remaining systems testing. Regarding the handheld mobile computing devices, it is critical that performance of these devices is clearly specified, measured, and that deficiencies in performance is effectively addressed. Until the project teams and the Decennial Management Division implement appropriate risk management activities, they face an increased probability that decennial systems will not be delivered on schedule and within budget or perform as expected. Mr. Chairman and members of the subcommittee, this concludes our statement. We would be happy to respond to any questions that you or members of the subcommittee may have at this time. If you have any questions on matters discussed in this testimony, please contact David A. Powner at (202) 512-9286 or Mathew Scirè at (202) 512-6806 or by e-mail at pownerd@gao.gov or sciremj@gao.gov. Other key contributors to this testimony include Mathew Bader, Thomas Beall, Jeffrey DeMarco, Richard Hung, Barbara Lancaster, Andrea Levine, Signora May, Cynthia Scott, Niti Tandon, Amos Tevelow, Jonathan Ticehurst, and Timothy Wexler. [End of section] Appendix I: Key 2010 Census Information Technology Acquisitions: IT acquisition: MAF/TIGER Accuracy Improvement Project (MTAIP); Contractor: Harris Corporation; Purpose: Modernize the system that provides the address list, maps, and other geographic support services for the Census and other Bureau surveys; Contract type: Cost plus award fee; Contract award: June 2002. IT acquisition: Field Data Collection Automation (FDCA); Contractor: Harris Corporation; Purpose: Provide automated resources for supporting field data collection, including the provision of handheld mobile computing devices to collect data in the field, including address and map data; Contract type: Cost plus award fee with some firm fixed price elements; Contract award: March 2006. IT acquisition: Decennial Response Integration System (DRIS); Contractor: Lockheed Martin Corporation; Purpose: Provide a solution for data capture and respondent assistance; Contract type: Cost plus award fee with some firm fixed price elements; Contract award: October 2005. IT acquisition: Data Access and Dissemination System (DADS II); Contractor: IBM; Purpose: Develop a replacement for the DADS legacy tabulation and dissemination systems; Contract type: To be determined; Contract award: September 2007. Source: GAO analysis of Census Bureau data. [End of table] [End of section] Footnotes: [1] 13 U.S.C. 141 (a) and (b). [2] Since issuance of our report in October 2007, the Bureau has tentatively moved the mock Census Day from April 1, 2008 to May 1, 2008. [3] GAO, Information Technology: Census Bureau Needs to Improve Its Risk Management of Decennial Systems, GAO-08-79 (Washington, D.C.: Oct. 5, 2007). [4] Address canvassing is a field operation to build a complete and accurate address list. In this operation, census field workers go door to door verifying and correcting addresses for all households and street features contained on decennial maps. [5] End-to-end testing is a form of operational testing that is performed to verify that a defined set of interrelated systems that collectively support an organizational core business function interoperate as intended in an operational environment. The interrelated systems include not only those owned and managed by the organization, but also the external systems with which they interface. [6] GAO, Census Bureau: Important Activities for Improving Management of Key 2010 Decennial Acquisitions Remain to be Done, GAO-06-444T (Washington, D.C.: Mar. 1, 2006). [7] TIGER is a registered trademark of the U.S. Census Bureau. [8] Handheld mobile computing devices will be used to update the Bureau‘s address list, to perform follow-up at addresses for which no questionnaire was returned, and to perform activities to measure census coverage. [9] The DADS II contract was originally planned to establish a new Web- based system that would serve as a single point for public access to all census data and integrate many dissemination functions currently spread across multiple Bureau organizations. [10] GAO-06-444T. [11] This analysis primarily addresses project teams‘ implementation of risk management processes. According to our analysis, the contractors for the three contracts awarded (MTAIP, FDCA, and DRIS) had implemented adequate risk management processes involving risk preparation, risk identification and analysis, and risk mitigation. [12] GAO, Information Technology Management: Census Bureau Has Implemented Many Key Practices, but Additional Actions Are Needed, GAO- 05-661 (Washington, D.C.: June 16, 2005) and GAO, Census Bureau: Important Activities for Improving Management of Key 2010 Decennial Acquisitions Remain to be Done, GAO-06-444T (Washington, D.C.: Mar. 1, 2006). [13] GAO, 2010 Census: Basic Design Has Potential, but Remaining Challenges Need Prompt Resolution, GAO-05-9 (Washington, D.C.: January12, 2005). [14] GAO-06-444T. [End of section] GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "Subscribe to Updates." Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office: 441 G Street NW, Room LM: Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Gloria Jarmon, Managing Director, JarmonG@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.