Census 2000

Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates Gao ID: GAO-05-71 November 12, 2004

Evaluations of past censuses show that certain groups were undercounted compared to other groups, a problem known as "coverage error." To address this, the Census Bureau included in its 2000 Census design the Accuracy and Coverage Evaluation Program (A.C.E.) to (1) measure coverage error and (2) use the results to adjust the census, if warranted. However, the Bureau found the A.C.E. results inaccurate and decided not to adjust or plan for adjustment in 2010. Congress asked GAO to determine (1) factors contributing to A.C.E.'s reported failure to accurately estimate census coverage error, and (2) the reliability of the revised coverage error estimates the Bureau subsequently produced. To do this, GAO examined three sets of Bureau research published in March 2001, October 2001, and March 2003 and interviewed Bureau officials.

According to senior Bureau officials, increasingly complicated social factors, such as extended families and population mobility, presented challenges for A.C.E., making it difficult to determine exactly where certain individuals should have been counted thus contributing to the inaccuracy of the coverage error estimates. For example, parents in custody disputes both may have an incentive to claim their child as a resident, but the Bureau used rules for determining where people should be countedresidence rules--that did not account for many of these kinds of circumstances. Other design decisions concerning both A.C.E. and the census also may have created "blind spots" that contributed to the inaccuracy of the estimates. The Bureau has not accounted for the effects of these or other key design decisions on the coverage error estimates, which could hamper the Bureau's efforts to craft a program that better measures coverage error for the next national census. Despite having twice revised A.C.E.'s original coverage error estimates, the Bureau has no reliable estimates of the extent of coverage error for the 2000 census. While both revisions suggested that the original estimates were inaccurate, in the course of thoroughly reviewing the revisions, the Bureau documented (1) extensive limitations in the revision methodology and (2) an unexpected pattern between the revised estimates and other A.C.E. data, both of which indicated that the revised coverage error estimates may be questionable themselves. Furthermore, when the Bureau published the revised estimates, it did not clearly quantify the impact of these limitations for readers, thus preventing readers from accurately judging the overall reliability of the estimates. It is therefore unclear how A.C.E. information will be useful to the public or policymakers, or how the Bureau can use it to make better decisions in the future.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-05-71, Census 2000: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates This is the accessible text file for GAO report number GAO-05-71 entitled 'Census 2000: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates' which was released on December 13, 2004. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: November 2004: CENSUS 2000: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates: GAO-05-71: GAO Highlights: Highlights of GAO-05-71, a report to congressional requesters Why GAO Did This Study: Evaluations of past censuses show that certain groups were undercounted compared to other groups, a problem known as ’coverage error.“ To address this, the Census Bureau included in its 2000 Census design the Accuracy and Coverage Evaluation Program (A.C.E.) to (1) measure coverage error and (2) use the results to adjust the census, if warranted. However, the Bureau found the A.C.E. results inaccurate and decided not to adjust or plan for adjustment in 2010. Congress asked GAO to determine (1) factors contributing to A.C.E.‘s reported failure to accurately estimate census coverage error, and (2) the reliability of the revised coverage error estimates the Bureau subsequently produced. To do this, GAO examined three sets of Bureau research published in March 2001, October 2001, and March 2003 and interviewed Bureau officials. What GAO Found: According to senior Bureau officials, increasingly complicated social factors, such as extended families and population mobility, presented challenges for A.C.E., making it difficult to determine exactly where certain individuals should have been counted thus contributing to the inaccuracy of the coverage error estimates. For example, parents in custody disputes both may have an incentive to claim their child as a resident, but the Bureau used rules for determining where people should be counted–residence rules--that did not account for many of these kinds of circumstances. Other design decisions concerning both A.C.E. and the census also may have created ’blind spots“ that contributed to the inaccuracy of the estimates (see figure). The Bureau has not accounted for the effects of these or other key design decisions on the coverage error estimates, which could hamper the Bureau‘s efforts to craft a program that better measures coverage error for the next national census. Factors Potentially Affecting Accuracy of Coverage Error Estimates at Different Points in the A.C.E. Program: [See PDF for image] [End of figure] Despite having twice revised A.C.E.‘s original coverage error estimates, the Bureau has no reliable estimates of the extent of coverage error for the 2000 census. While both revisions suggested that the original estimates were inaccurate, in the course of thoroughly reviewing the revisions, the Bureau documented (1) extensive limitations in the revision methodology and (2) an unexpected pattern between the revised estimates and other A.C.E. data, both of which indicated that the revised coverage error estimates may be questionable themselves. Furthermore, when the Bureau published the revised estimates, it did not clearly quantify the impact of these limitations for readers, thus preventing readers from accurately judging the overall reliability of the estimates. It is therefore unclear how A.C.E. information will be useful to the public or policymakers, or how the Bureau can use it to make better decisions in the future. What GAO Recommends: GAO is making recommendations related to how the Bureau handles design decisions and reports results to improve planning and reporting of 2010 census and coverage evaluation. The Department of Commerce concurred with the recommendations but raised several concerns regarding our analyses, which we have discussed in the report and used to revise it. www.gao.gov/cgi-bin/getrpt?GAO-05-71. To view the full product, including the scope and methodology, click on the link above. For more information, contact Patricia A. Dalton at (202) 512-6806 or daltonp@gao.gov. [End of section] Contents: Letter: Results in Brief: Background: Scope and Methodology: Design and Other Bureau Decisions Created Difficulties and Blind Spots in Census Coverage Evaluation: The Bureau Has Not Produced Reliable Revised Estimates of Coverage Error for the 2000 Census: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendixes: Appendix I: Comments from the Department of Commerce: Appendix II: Bureau Design Decisions Increased Imputations: Appendix III: Bureau Estimates of Population and Housing Unit Net Undercounts: Glossary: Related GAO Products: Tables: Table 1: Estimated Percent Population Undercounts for Selected Race/ Origin Groups: Table 2: Estimated Percent Occupied Housing Undercounts Differ from Population Undercounts for Selected Race/Origin Groups: Figures: Figure 1: A.C.E. Sample Operations Paralleled the Census: Figure 2: A.C.E. Sample Excluded Individuals and Records from Scope of Evaluation: Figure 3: A.C.E. Matching Results Lacked Data on Suspected Duplicates That Were in the Census: Figure 4: Reported Undercount Estimates Generally Decreased with A.C.E. Revisions: Figure 5: Population and Housing Undercounts: Abbreviations: A.C.E.: Accuracy and Coverage Evaluation: NAS: National Academy of Sciences: Letter November 12, 2004: The Honorable Henry A. Waxman: Ranking Minority Member: Committee on Government Reform: House of Representatives: The Honorable Danny K. Davis: Ranking Minority Member: Subcommittee on Civil Service and Agency Organization: Committee on Government Reform: House of Representatives: The Honorable Wm. Lacy Clay, Jr.: Ranking Minority Member: Subcommittee on Technology, Information Policy, Intergovernmental Relations and the Census: Committee on Government Reform: House of Representatives: The Honorable Charles A. Gonzalez: The Honorable Carolyn B. Maloney: House of Representatives: A decennial census must be as accurate as possible, because census results are used to, among other purposes, apportion congressional seats, redraw congressional districts, and allocate federal aid to state and local governments. However, given the nation's size and demographic complexity, some amount of error is inevitable. Unfortunately, evaluations of past censuses have shown that certain groups, for example African-Americans and Hispanics, have been undercounted in comparison to other groups. To estimate the extent that some groups were over-or undercounted in 2000--what the Bureau refers to as "coverage error"--the Census Bureau (Bureau) planned and implemented the Accuracy and Coverage Evaluation (A.C.E.) program. The primary goals of A.C.E. were to: * more accurately estimate the rate of coverage error via a sample survey of select areas nationwide, and: * if warranted, to use the results of this survey to adjust census estimates of the population for nonapportionment purposes. In March 2003, after much deliberation and research, the Bureau decided not to use any A.C.E. estimates of coverage error to adjust the 2000 Census, because it judged these estimates no more accurate than the official census data. The Bureau found that A.C.E. did not account for at least 3 million erroneously counted persons (mostly duplicates, or people counted twice) in the census, which raised questions about the reliability of the coverage error estimates. Furthermore, because of the difficulties the Bureau experienced in trying to produce reliable coverage error estimates, it announced officially in January 2004 that it did not plan to develop a procedure for adjusting the 2010 Census results for redistricting purposes. Agency officials said that in light of their past experiences, they do not think they can produce reliable coverage error estimates in time to meet deadlines for adjusting the census. This report responds to your request that we review why A.C.E. coverage error estimates were reportedly not sufficiently reliable to adjust or validate the 2000 Census. Specifically, our review examined (1) factors contributing to A.C.E.'s reported failure to accurately estimate census coverage error, and (2) the reliability of the revised coverage error estimates the Bureau subsequently produced. To meet the objectives of this report, we reviewed and analyzed the Bureau's publicly available research data and reports on the 2000 Census. We also reviewed methodology documents and other available information such as the minutes and supporting documents of the Executive Steering Committees for Adjustment Policy. Finally, we discussed the results of our analysis with senior Bureau officials and interviewed Bureau officials and committee members to obtain their views on the process. Since our focus was on the process and decisions that led to the results rather than on determining the underlying numbers themselves, we did not audit the Bureau's research, the underlying data, or its conclusions. Our work was performed in Washington D.C. and at the U.S. Census Bureau headquarters in Suitland, Maryland, from December 2002 through July 2004 in accordance with generally accepted government auditing standards. Results in Brief: The Bureau attributes the inaccuracy of A.C.E. estimates primarily to the rules it used for determining where people should be counted-- residence rules--which officials say did not fully capture the increasing complexity of American society. For example, parents in custody disputes both may have an incentive to claim their child as a member of their respective household, but the Bureau's residence rules did not always account for these kinds of circumstances, thus accurately counting such individuals was difficult. While the Bureau emphasizes residence rules, our research indicates that some of the Bureau's other design decisions concerning both A.C.E. and the census created several "blind spots" in A.C.E. that also may have contributed to the unreliability of coverage error estimates. For example, the Bureau decided to exclude from A.C.E. people living in group quarters such as college dorms, because their high mobility made them difficult to count during the 1990 coverage evaluation, and this weakened the 1990 coverage error estimates. In addition, the Bureau removed 2.4 million records it suspected were duplicates from the census population before A.C.E. had a chance to measure them, and then reinstated them after A.C.E. was finished, so that A.C.E. was blind to over 1 million duplicates (according to Bureau research) in that reinstated group. However, because the Bureau has not accounted for and clearly reported how these and other key design decisions affected the outcome of A.C.E., the full range of reasons why its estimates were unreliable remains obscure. This in turn could hamper the Bureau's efforts to craft a more successful coverage measurement program that better measures the accuracy of the next national census. The Bureau's revised coverage error estimates cannot necessarily be interpreted as any more reliable than the original estimates. In the course of an extensive review of A.C.E.'s results over a 3-year period, the Bureau revised its estimates of A.C.E. coverage error twice--first in October 2001 and second in March 2003--and both revisions suggested that the original estimates were inaccurate. However, in reviewing the revised estimates, the Bureau documented (1) extensive limitations in the revision methodology and (2) an unexpected pattern between the revised estimates and other data, both of which indicated that the revised estimates may be questionable themselves. Furthermore, while the Bureau described the aforementioned limitations in detail in its review documents, when it published the revised estimates of coverage error, it did not clearly quantify their impact for readers. It is therefore unclear how these revised estimates will be useful to the public or policymakers, or how the Bureau will use them to make better decisions in the future. As a result, we are recommending that the Bureau's future evaluation planning take into account the potential effect of future decisions relating to census or coverage evaluation design to avoid similar or other problems in the 2010 measure of coverage error. We also recommend that the Bureau clearly report how any significant changes in the design of the census and/or A.C.E. might affect the accuracy of the published coverage error estimates for 2010. Similarly, GAO recommends that the Bureau not only identify methodological limitations--as it currently does--but also report the potential range of impacts that methodological limitations and design changes may have on the census coverage error estimates it publishes in the future. The Under Secretary for Economic Affairs forwarded written comments from the Department of Commerce on a draft of this report (see app. I). The Department concurred with our recommendations but took exception to several statements in our draft report, writing that it thought some of our conclusions were misleading. In response, we revised the report, as appropriate, including additional information provided by the Department and clarifying the presentation of our analyses. We discuss the Department's comments more fully in the section "Agency Comments and Our Evaluation." Background: Since 1980, the Bureau has used statistical methods to generate detailed estimates of census undercounts and overcounts, including those of particular ethnic, racial, and other groups. To carry out the 2000 Census's Accuracy and Coverage Evaluation program (A.C.E.), the Bureau conducted a separate and independent sample survey that, when matched to the census data, was to enable the Bureau to use statistical estimates of net coverage errors to adjust final census tabulations according to the measured undercounts, if necessary. The Bureau obligated about $207 million to its coverage evaluation program from fiscal years 1996 through 2001, which was about 3 percent of the $6.5 billion total estimated cost of the 2000 Census. While the A.C.E. sample survey of people was conducted several weeks after Census Day, April 1, the "as of" date on which the total population is to be counted, many of the processes were the same as the 2000 Census. For the census, the Bureau tried to count everybody in the nation, regardless of their dwelling, and certain kinds of dwellings, including single-family homes, apartments, and mobile homes, along with demographic information on the inhabitants. For A.C.E., the Bureau surveyed about 314,000 housing units in a representative sample of "clusters"--geographic areas each with about 30 housing units. The sample comprised roughly 12,000 of the about 3 million "clusters" nationwide. As illustrated in figure 1, the Bureau used a similar process to develop address lists, collect response data, and tabulate and disseminate data--one for the decennial census and one for A.C.E. sample areas. For the Census, the Bureau mailed out forms for mail-back to most of the housing units in the country; hand-delivered mail-back forms to most of the rest of the country; and then carried out a number of follow-up operations designed to count nonrespondents and improve data quality. A.C.E. collected response data through interviewing from April 24 through September 11, 2000. Figure 1: A.C.E. Sample Operations Paralleled the Census: [See PDF for image] [End of figure] After the census and A.C.E. data collection operations were completed, the Bureau attempted to match each person counted on the A.C.E. list to the list of persons counted by the 2000 Census in the A.C.E. sample areas to determine exactly which persons had been missed or counted more than once by either A.C.E. or the Census. The results of the matching process, along with data on the racial/ethnic and other characteristics of persons compared, were to provide the basis for A.C.E. to estimate the extent of coverage error in the census and population subgroups and enable the Bureau to adjust the final decennial census tabulations accordingly. The matching process needed to be as precise and complete as possible, since A.C.E. collected data on only a sample of the nation's population, and small percentages of matching errors could significantly affect the estimates of under-and overcounts generalized to the entire nation. Since the 2000 Census, we have issued four other reports on A.C.E., addressing its cost and implementation as part of our ongoing series on the results of the 2000 Census, as well as a report on the lessons learned for planning a more cost-effective census in 2010. (See the Related GAO Products section at the end of this report for the assessments issued to date.) These reports concluded, among other things, that while (1) the address list the Bureau used for the A.C.E. program appeared much more accurate than the preliminary lists developed for the 2000 Census and (2) quality assurance procedures were used in the matching process, certain implementation problems had the potential to affect subsequent matching results and thus estimates of total census coverage error.[Footnote 1] In the end, the Bureau decided not to use A.C.E.'s matching process results to adjust the 2000 Census. In March 2001, a committee of senior career Bureau officials recommended against using A.C.E. estimates of census coverage error to adjust final census tabulations for purposes of redistricting Congress. In October 2001, the committee also recommended against adjusting census data used for allocating federal funds and other purposes, largely because Bureau research indicated that A.C.E. did not account for at least 3 million erroneously counted persons (mostly duplicates) in the census, raising questions about the reliability of coverage error estimates. In March 2003, after considerable additional research, the Bureau published revised coverage error estimates and again decided not to adjust official census data, this time for the purpose of estimating the population between the 2000 and 2010 censuses. In light of its 2000 experience, the Bureau officially announced in January 2004 that while it plans to fully evaluate the accuracy of the 2010 census, it will not develop plans for using these coverage error estimates to adjust the 2010 Decennial Census. Bureau officials have told us that there is insufficient time to carry out the necessary evaluations of the coverage estimates between census field data collection and the Bureau's legally mandated deadline (within 12 months of Census Day) for releasing redistricting data to the states. Furthermore, the Bureau does not believe adjustment is possible. In responding to an earlier GAO report recommending that the Bureau "determine the feasibility" of adjusting the 2010 Census, the Bureau wrote that the 2000 Census and A.C.E. was "a definitive test of this approach" which "provided more than ample evidence that this goal cannot be achieved."[Footnote 2] However, in March, the National Academy of Sciences (NAS) published a report that recommended that the Bureau and the administration request and Congress provide funding for an improved coverage evaluation program that could be used as a basis for adjusting the census, if warranted.[Footnote 3] The Academy agrees with the Bureau that 12 months is insufficient time for evaluation and possible adjustment; in the same publication, NAS recommended Congress consider extending the statutory deadline of 12 months for providing data for redistricting purposes, a suggestion which, if appropriate, could make adjustment possible. Scope and Methodology: To identify the factors that may have contributed to A.C.E. missing coverage errors in the census, we reviewed evaluations of A.C.E. and the Bureau's subsequent revisions to its estimation methodology, as well as changes made to the design from its 1990 attempts to estimate coverage. We interviewed Bureau officials responsible for A.C.E. decision making to obtain further context and clarification. We did not attempt to identify all factors contributing to the success or failure of A.C.E in estimating coverage error. Since our focus was on the process and decisions that led to the results rather than on determining the underlying numbers themselves, we did not audit the Bureau's research, the underlying data, or its conclusions. We relied on the Bureau's own reporting quality assurance processes to assure the validity and accuracy of its technical reporting, and thus we did not independently test or verify individual Bureau evaluations of their methodologies. To identify the extent of the census errors not accounted for by A.C.E., we reviewed the descriptive coverage error estimates and the limitations and context of these data as described in the Bureau reports published by the Bureau in March 2001, October 2001, and March 2003. On, August 9, 2004, we requested comments on the draft of this report from the Secretary of Commerce. On September 10, 2004, the Under Secretary for Economic Affairs, Department of Commerce forwarded written comments from the department (see app. I), which we address in the "Agency Comments and Our Evaluation" section at the end of this report. Design and Other Bureau Decisions Created Difficulties and Blind Spots in Census Coverage Evaluation: The following Bureau decisions concerning the design of the census and the A.C.E. program created difficulties and blind spots for the coverage evaluation, possibly preventing A.C.E. from reliably measuring coverage error: (1) using residence rules that were unable to capture the complexity of American society, (2) excluding the group quarters population from the A.C.E. sample survey, (3) making various decisions that led to an increase in the number of "imputed" records in the census, (4) removing 2.4 million suspected duplicate persons from the census but not the A.C.E. sample, and (5) reducing the sample area wherein A.C.E. searched for duplicates during matching. However, the Bureau has not accounted for how these design decisions have affected coverage error estimates, which has prevented it from pinpointing what went wrong with A.C.E., and this in turn could hamper its efforts to craft a more successful coverage measurement program for the next national head count. Bureau Attributes Inaccuracy of Coverage Error Estimates to Residence Rules That Did Not Capture Complexity of U.S. Society: Bureau officials attribute A.C.E.'s inaccuracy primarily to the fact that it used residence rules that do not fully capture the complexity of American society. According to senior Bureau officials, increasingly complicated social factors, such as extended families and population mobility, presented challenges for A.C.E., making it difficult to determine exactly where certain individuals should have been counted. Specifically, in developing A.C.E. methodology, Bureau officials assumed that each person in its sample could be definitively recorded at one known residence that the Bureau could determine via a set of rules. However, individuals' residency situations are often complicated: Bureau officials cite the example of children in custody disputes whose separated parents both may have strong incentives to claim the children as members of their household, despite census residence rules that attempt to resolve which parent should report the child(ren). In such situations, wherein the residence rules are either not understood, are not followed, or do not otherwise provide resolution, the Bureau has difficulty determining the correct location to count the children. Bureau officials cite similar difficulties counting college students living away from home, as well as people who live at multiple locations throughout the year, such as seasonal workers or retirees. A.C.E. design also assumed that follow-up interviews would clarify and improve residence data for people for whom vague, incomplete, or ambiguous data were provided and whose cases remained unresolved. However, the Bureau found it could not always rely on individuals to provide more accurate or complete information. In fact, in our earlier reporting on A.C.E. matching, we described several situations wherein conflicting information had been provided to the Bureau during follow- up interviews with individuals, and Bureau staff had to decide which information to use.[Footnote 4] More recently, the Associate Director for Decennial Census told us that returning to an A.C.E. household to try and resolve conflicting data sometimes yielded new or more information but not necessarily better information or information that would resolve the conflict. The Bureau plans to review and revise its census residence rules for 2010, which may clarify some of this confusion. Other Design Decisions Also Created Blind Spots in Census Coverage Evaluation: While the Bureau emphasizes residence rules as the primary cause of A.C.E. failure, our research indicates some of the Bureau's other design decisions created blind spots that also undermined the program's ability to accurately estimate census error. For example, the Bureau decided to leave people living in group quarters--such as dormitories and nursing homes--out of the A.C.E. sample survey, which effectively meant they were left out of the scope of A.C.E. coverage evaluation (see fig. 2). Figure 2: A.C.E. Sample Excluded Individuals and Records from Scope of Evaluation: [See PDF for image] Note: Figures are not drawn to proportion. [End of figure] As a result, the matching results could not provide coverage error information for the national group quarters population of 7.8 million. In addition, the Bureau did not design A.C.E. matching to search for duplicate records within the subset of this population counted by the census, though later Bureau research estimated that if it had, it would have measured over 600,000 additional duplicates there. In response to our draft report the Department wrote that coverage evaluation was designed to measure some of these duplicates since during its follow-up interviews at households during A.C.E. matching, the Bureau included questions intended to identify college students living away at college. While coverage evaluation in 1990 included some group quarters, such as college dormitories and nursing homes, within its sample, the Bureau reported that the high mobility of these people made it more difficult to count them, thus the 1990 estimates of coverage for this population were weak. The Bureau decided not to gather these data during 2000 A.C.E. data collection based in part on the difficulty of collecting and matching this information in the past, and in part as a result of a separate design decision to change the way to treat information for people who moved during the time period between the census and the coverage evaluation interviews. By excluding group quarters from the coverage evaluation sample, the Bureau had less information collected on a sample of this population that included some duplication, and the missing information may have enabled it to better detect and account for such duplication. In addition, by developing coverage error estimates that were not applicable to the group quarters population, the Bureau made the task of assessing the quality of the census as a whole more difficult. Figure 2 also shows that another blind spot emerged as the Bureau increased the number of "imputed"[Footnote 5] records in the final census, though they could not be included in the A.C.E. sample survey. The Bureau estimates a certain number of individuals--called imputations--that they have reason to believe exist, despite the fact that they have no personal information on them, and adds records to the census (along with certain characteristics such as age and race/ ethnicity) to account for them. For example, when the Bureau believes a household is occupied but does not have any information on the number of people living there, it will impute the number of people as well as their personal characteristics. The Bureau increased imputations in 2000 from about 2 million in 1990 to about 5.8 million records. Changes in census and coverage evaluation design from 1990 likely contributed to this increase. Specifically, the Bureau reduced the number of persons who could have their personal information recorded on the standard census form in 2000. In addition, the Bureau changed the way coverage evaluation accounted for people who moved between Census Day and the day of the coverage evaluation interview. These design changes resulted in less complete information on people and likely contributed to an increase in imputations. (These and other changes are explained in more detail in app. II.) Because imputed records are simply added to the census totals and do not have names attached to them, it was impossible for A.C.E. to either count imputed individuals using the A.C.E. sample survey or incorporate them into the matching process. At any rate, since the true number and characteristics of these persons are unknown, matching these records via A.C.E. would not have provided meaningful information on coverage error. A.C.E. was designed to measure the net census coverage error, in essence the net effect of people counted more than once minus people missed, and included measurement of the effects of imputation on coverage error. The Bureau generalizes its estimates of coverage error to cover imputations and also maintains that its imputation methods do not introduce any statistical bias in population counts. But the formulas used by the Bureau to calculate its estimates of coverage error account for imputations by subtracting them from the census count being evaluated, not by measuring the error in them. And the Bureau did not attempt to determine the accuracy of the individual imputations, that is although the Bureau imputed persons they have reason to believe existed, the Bureau does not know whether it over-or underestimated such persons. As more imputations are included in the census total, the generalization of coverage error estimates to that total population becomes less reliable. Similarly, the Bureau created an additional coverage error blind spot by including in the census 2.4 million records that it previously suspected were duplicates and thus were not included in the coverage evaluation. Prior to A.C.E. matching, the Bureau removed about 6 million persons from the census population, identifying them as likely duplicates. Then, after completing additional research on these possible duplicates, the Bureau decided to reinstate the records for 2.4 million of these persons it no longer suspected were duplicates. However, it did so after A.C.E. had completed matching and evaluating the census records from which the 2.4 million persons had been removed and for which coverage error estimation had begun (see fig. 3). Figure 3: A.C.E. Matching Results Lacked Data on Suspected Duplicates That Were in the Census: [See PDF for image] Note: Figures are not drawn to proportion. [End of figure] The Bureau documented in a memorandum that the exclusion of the records from A.C.E. processing was not statistically expected to affect A.C.E. results. However, later Bureau research concluded that over 1 million of these 2.4 million records were likely duplicates, none of which could have been detected by A.C.E. While the Bureau maintains that the reinstatement of the over 1 million reinstated likely duplicates did not affect the A.C.E. estimate in a statistically significant way, this suggests that the resulting A.C.E.-based estimate of national population itself is blind to the presence in the census of the over 1 million duplicates the Bureau reintroduced. For 2010, Bureau officials have chartered a planning group responsible for, among other things, proposing improvements to reduce duplication in the census, which may address some of the problem. In addition to excluding populations from the scope of evaluation, the Bureau further curtailed its ability to measure coverage error by reducing A.C.E.'s search area to only one geographic ring around selected A.C.E. sample areas during the matching process. For the 1990 Census, the Bureau's coverage evaluation program always searched at least one surrounding ring and an even larger ring in rural areas. However, in 1999, before a panel of the National Academy of Sciences, Bureau officials announced and defended the decision to not expand the search area except in targeted instances, saying that Bureau research indicated that the additional matches found in 1990 from the expanded search areas did not justify the additional effort. In its comments on our draft report, the Department writes that more important than the size of the search area is maintaining "balance"--i.e., search areas must be used consistently both to identify people who have been missed and to identify people who have been counted in error (including duplicates). The Department also justified the decision to reduce the search area in 2000 from 1990 in part by stating, "in an expected value sense, the reduced search area would have affected" [emphasis added] the extra missed people and the extra miscounted people equally, or been balanced. However, later research discovered large numbers of the missed duplicates in the actual census by matching A.C.E. persons to census persons nationwide--far beyond the areas searched during A.C.E. matching. A 2001 Bureau report presenting the results of computer rematching of the A.C.E. sample concluded, "Our analysis found an additional 1.2 million duplicate enumerations in units that were out- of-scope for the A.C.E. but would have been in-scope for [1990 coverage evaluation]."[Footnote 6] In other words, if the Bureau had continued its 1990 practice of searching in housing units in larger geographic areas in 2000, the A.C.E. process might have identified more duplicates and yielded better results.[Footnote 7] The Bureau research cited above appears to question the decision to reduce the search areas. In fact, after the 2000 Census was completed, again before the National Academy of Sciences, a Bureau official suggested that any coverage evaluation methods for 2010 should conduct a more thorough search, perhaps expanding the search area to two or more geographic rings everywhere. Bureau Has Not Fully Accounted for How Design Decisions Affected Coverage Error Estimates: This review has identified only some of the decisions that could have created problems in A.C.E. estimates. Because the Bureau has not attempted to account for how all of its design decisions relating to A.C.E. and the census affected the outcome of the program, the full range of reasons that A.C.E. estimates were not reliable remains obscure. Bureau research documented and this report describes the magnitude of the direct effects of most of these design decisions in terms of the size of the census population affected, and the Bureau's final reporting on the revised A.C.E. estimates mentions many design changes, but not together or comprehensively, and they do not explain how the changes might have affected the estimates of coverage error. Without clear documentation of how significant changes in the design of the census and A.C.E. might have affected the measurements of census accuracy, it is not apparent how problems that have arisen as a result of the Bureau's own decisions can be distinguished from problems that are less under the Bureau's control, i.e., difficulties inherent to conducting coverage evaluation. Thus the Bureau's plans to measure the coverage error for the 2010 Census are not based on a full understanding of the relationship between the separate decisions it makes about how to conduct A.C.E. and the census and the resulting performance of its coverage measurement program. This in turn could hamper the Bureau's efforts to craft a more successful coverage measurement program for the next national head count. The Bureau Has Not Produced Reliable Revised Estimates of Coverage Error for the 2000 Census: While the Bureau produced a vast body of research regarding the census and A.C.E., including multiple reassessments and revisions of earlier work, the revised estimates are not reliable. The initial A.C.E. estimates of coverage error suggested that while historical patterns of differences in undercounts between demographic groups persisted, the Bureau had succeeded in 2000 in reducing the population undercounts of most minorities, and the subsequent revised estimates showed even greater success in reducing population undercounts in the Census. However, the large number of limitations described in the Bureau's documentation of the methodology used to generate the revised estimates of coverage error suggest that these estimates are less reliable than reported and may not describe the true rate of coverage error. The Bureau, however, has not made the full impact of these methodological limitations on the data clear. Moreover, the final revised estimates of coverage error for the count of housing units and the count of people, which the Bureau expected to be similar if the estimates were reliable, differed, further raising questions about the reliability of the revised estimates. Bureau's Conclusion That A.C.E. Estimates Are Unreliable Is Based on Results That Are Unreliable Themselves: The Bureau undertook an extensive review of A.C.E.'s results over a 3- year period. In doing so, the Bureau revised its estimates of A.C.E. coverage error twice--first in October 2001 and again in March 2003. These revisions suggest that the original A.C.E. estimates were unreliable. Figure 4 illustrates how each of the revised A.C.E. estimates of coverage error reduced the undercount for most of the three major race/origin groups from the initial A.C.E. estimates. Note that the final revised estimate indicates that the total population was actually overcounted by one-half of 1 percent. Figure 4: Reported Undercount Estimates Generally Decreased with A.C.E. Revisions: [See PDF for image] [End of figure] The differences in the revised estimates presumably provide a measure of the error in the original A.C.E. estimates. (The estimated net population undercounts--and their standard errors--for these groups are provided in app. III.) However, the revised estimates of coverage error may not be reliable enough themselves to provide an adequate basis for such a comparison to measure error in the original A.C.E. estimates. First, the number of Bureau-documented limitations with respect to the methodologies used in generating A.C.E.'s revised estimates raises questions about the accuracy of the revised estimates. Within voluminous technical documentation of its process, the Bureau identified several methodological decisions wherein if the decisions had been made differently, they may have led to appreciably different results. Thus the methods the Bureau chose may have affected the estimates of census coverage error themselves and/or the measures of uncertainty associated with the estimates, limiting the general reliability of the revisions. The limitations in the methodologies for the revised estimates included the following: * Errors in the demographic data used to revise estimates may have contributed to additional error in the estimates. * Errors stemming from the Bureau's choice of model to resolve uncertain match cases were accounted for in initial March 2001 A.C.E. estimates but were not accounted for in the revised estimates in March 2003. * Alternative possible adjustments for known inefficiencies in computer matching algorithms would directly affect revised estimates. * The Bureau's evaluations of the quality of clerical matching were used to revise the initial A.C.E. coverage estimates, leaving the Bureau less reliable means to measure the uncertainty in the revised estimates. The Bureau Has Not Clearly Quantified and Reported the Full Impact of the Methodological Limitations on the Revised Estimates: For the earlier revision of coverage error estimates, the Bureau provided the range of impacts that could result from some different methodological decisions, enabling more informed judgments regarding the reliability of the data. For example, in support of its October 2001 decision to not adjust census data, the Bureau showed that different assumptions about how to treat coverage evaluation cases that the Bureau could not resolve could result in variations in the census count of about 6 million people. The Bureau also had reported previously the range of impacts on the estimates resulting from different assumptions and associated calculations to account for the inefficiency of computer matching. They found that different assumptions could result in estimates of census error differing by about 3.5 million people. However, with the final revision of the A.C.E. coverage error estimates, the Bureau did not clearly provide the ranges of impact resulting from different methodological decisions. While the Bureau did discuss major limitations and indicated their uncertain impact on the revised estimates of coverage error, the Bureau's primary document for reporting the latest estimates of coverage error did not report the possible quantitative impacts of all these limitations--either separately or together--on the estimates. Thus readers of the reported estimates do not have the information needed to accurately judge the overall reliability of the estimates, namely, the extent of the possible ranges of the estimates had different methodological decisions been made. Sampling errors were reported alongside estimates of census error, but these do not adequately convey the extent of uncertainty associated with either the reported quantitative estimates themselves or the conclusions to be drawn from them. For example, the Bureau decided to make no adjustment to account for the limitation of computer matching efficiency when calculating its latest revision of estimates of coverage error, unlike the adjustment it made when calculating its earlier revised estimates. When justifying its adjustment made in its earlier revised estimates, the Bureau demonstrated that the choice of adjustment mattered to the calculated results. But the potential significance to the calculated results of the Bureau's having made a different assumption was not reflected in the Bureau's primary presentation of its estimates and their errors. The absence of clear documentation on the possible significant impacts of such assumptions could lead readers of the Bureau's reporting to believe erroneously that all assumptions have been accounted for in the published statistics, or that the estimates of coverage error are more reliable than they are. Unexpected Differences in Patterns of Coverage Error between the Housing and the Population Count Further Call into Question Reliability of Revised Estimates: According to Bureau reporting, when it examined the validity of the revised coverage error estimates the Bureau expected to see across demographic groups similar patterns between the coverage error for the count of the population and the count of housing. That is, if a population was overcounted or undercounted, then the count of housing units for that population was expected to be overcounted or undercounted as well. The Bureau developed estimates of coverage error in the count of housing units from A.C.E. data. But the comparisons of non-hispanic blacks and hispanics to non-hispanic whites in figure 5 shows that the relative housing undercounts are opposite of what was expected by the Bureau. Figure 5: Population and Housing Undercounts: [See PDF for image] [End of figure] For example, the estimated population undercount for non-hispanic blacks is almost 3 percent greater than that of the majority group-- non-hispanic white or other--but the estimated housing unit undercount for non-hispanic blacks is about 0.8 percent less than that of the majority group. In addition, while the Bureau estimated that the non- hispanic white majority group had a net population overcount of over 1 percent, the Bureau estimated the majority group as having its housing units undercounted by about one-third of a percent. (The estimated net housing unit undercounts--and their standard errors--for these groups are provided in app. III.) Bureau officials told us that the problems A.C.E. and the census experienced with identifying coverage error in the population do not seem likely to have affected housing counts. However, when estimating coverage error for housing units for specific populations (e.g., by gender or race/ethnicity) errors in the population count can affect the reliability of housing tabulations. This is because when the Bureau tabulates housing data by characteristics like gender or race, it uses the personal characteristics of the person recorded as the head of the households living in each housing unit. So if there are problems with the Bureau's count of population for demographic groups, for example by gender or sex, they will affect the count of housing units for demographic groups. While the unexpected patterns in population and housing unit coverage error may be reconcilable, Bureau officials do admit that problems with the estimations of population coverage error may also adversely affect the reliability of other measures of housing count accuracy they rely upon, such as vacancy rates. Bureau officials have indicated the need to review this carefully for 2010. Bureau Reports Revised Estimates Resulted in Lessons Learned: While the multiple reassessments and revisions of earlier work did not result in reliable estimates, these efforts were not without value, according to the Bureau. Bureau officials stated that the revision process and results helped the Bureau focus for 2010 on detecting duplicates, revising residence rules, and improving the quality of enumeration data collected from sources outside the household, such as neighbors, as well as providing invaluable insights for its program of updating census population estimates throughout the decade. Conclusions: The volume and accessibility over the Internet of the Bureau's research may have made this the most transparent coverage evaluation exercise of a Decennial Census. However, as the Bureau has closed the book on Census 2000 and turned toward 2010, the reliability of the Bureau's coverage estimates remains unknown. The Bureau made extensive efforts to evaluate the census and its coverage error estimates resulting from A.C.E., but these efforts have not been sufficient to provide reliable revised estimates of coverage error. So while much is known about operational performance of the 2000 Census, one of the key performance measures for the 2000 census remains unknown. Moreover, neither Congress nor the public know why the coverage evaluation program did not work as intended, because the Bureau has not provided a clear accounting of how census and A.C.E. design decisions and/or limitations in the A.C.E. revision methodology discussed in this report accounted for the apparent weakness--or strengths--of A.C.E. Without such an accounting, the causes of problems and whether they can be addressed will remain obscure. And as the Bureau makes plans for coverage evaluation for the 2010 Census, whether that program approximates A.C.E.'s design or not, the Bureau will be missing valuable data that could help officials make better decisions about how to improve coverage evaluation. Finally, this lack of information calls into question the Bureau's claim (made in response to a prior GAO recommendation that the Bureau determine the feasibility of adjustment) that it has already established that using coverage evaluation for adjustment purposes is not feasible. Without clearly demonstrating what went wrong with its most recent coverage evaluation and why, the Bureau has not shown that coverage evaluation for the purpose of adjustment is not feasible. In fact, this report mentions two census improvements--to residence rules and to efforts to identify and reduce duplicates--that the Bureau is already considering that could make A.C.E. estimates more reliable, and perhaps even feasible. Furthermore, although the Bureau reports that its experience with revising A.C.E. estimates has provided lessons, it remains unclear how the Bureau will use its published coverage error estimates to make decisions leading to a more reliable measure of coverage error in 2010, or how the unreliable estimates can be of value to policymakers or the public. Recommendations for Executive Action: As the Bureau plans for its coverage evaluation of the next national head count in 2010, we recommend that the Secretary of Commerce direct that the Bureau take the following three actions to ensure that coverage evaluation results the Bureau disseminates are as useful as possible to Congress and other census stakeholders: * To avoid creating any unnecessary blind spots in the 2010 coverage evaluation, as the Bureau plans for its coverage evaluation in 2010, it should take into account how any significant future design decisions relating to census (for example, residence rules, efforts to detect and reduce duplicates, or other procedures) or A.C.E. (for example, scope of coverage, and changes in search areas, if applicable), or their interactions, could affect the accuracy of the program. * Furthermore, in the future, the Bureau should clearly report in its evaluation of A.C.E. how any significant changes in the design of census and/or A.C.E. might have affected the accuracy of the coverage error estimates. * In addition GAO recommends that in the future the Bureau plan to not only identify but report, where feasible, the potential range of impact of any significant methodological limitation on published census coverage error estimates. When the impact on accuracy is not readily quantifiable, the Bureau should include clear statements disclosing how it could potentially affect how people interpret the accuracy of the census or A.C.E. Agency Comments and Our Evaluation: The Under Secretary for Economic Affairs at the Department of Commerce provided us written comments from the Department on a draft of this report on September 10, 2004 (see appendix I). The Department concurred with our recommendations, but took exception to some of our analyses and conclusions and provided additional related context and technical information. In several places, we have revised the final report to reflect the additional information and provided further clarification on our analyses. The Department was concerned that our draft report implied that A.C.E. was inaccurate because it should have measured gross coverage error components, and that this was misleading because the Bureau designed A.C.E. to measure net coverage errors. While we have previously testified that the Bureau should measure gross error components, and the Department in its response states that this is now a Bureau goal for 2010, we clarified our report to reflect the fact that the Bureau designed A.C.E. to measure net coverage error. Further, although the Department agreed with our finding that the Bureau used residence rules that were unable to capture the complexity of American society thus creating difficulty for coverage evaluation, the Department disagreed with our characterization of the role four other census and A.C.E. design decisions played in affecting coverage evaluation. Specifically, the Bureau does not believe that any of the following four design decisions contributed significantly to the inaccuracy of the A.C.E. results: 1. The Treatment of the Group Quarters Population--The Department commented that we correctly noted that group quarter residents were excluded from the A.C.E. universe who would have been within the scope of A.C.E. under the 1990 coverage evaluation design, and that a large number of these people were counted more than once in 2000. The Department maintains that the Bureau designed A.C.E. to measure such duplicates. We believe this is misleading. As the Department noted, during its follow-up at housing units the Bureau included questions intended to identify the possible duplication of college students living away at college, and we have now included this in our final report. But as we stated in our draft report, A.C.E. did not provide coverage error information for the national group quarters population. Moreover, during A.C.E. the Bureau did not search for duplicate people within the group quarters population counted by the census, as it did within housing units counted by the census. In fact, later Bureau research estimated that if it had done so, the Bureau would have identified over 600,000 additional duplicates there. As such, our finding that this may have contributed to the unreliability of coverage error estimates still stands. 2. The Treatment of Census Imputations--The Department stated that A.C.E. was designed to include the effects of imputations on its measurement of coverage error and that there was no basis for our draft report stating that as more imputations were included in the census then coverage error estimates became less reliable. While we agree that the Bureau's estimates of coverage error accounted for the number of imputations, as we report, and as the Department's response reiterated, no attempt was made to determine the accuracy of the imputations included in the census. Thus any errors in either the number or demographic characteristics of the population imputed by the Bureau were not known within the coverage error processes or estimation. As a result, in generalizing the coverage error estimates to the imputed segment of the population, the Bureau assumed that the imputed population had coverage error identical to the population for which coverage error was actually measured. Furthermore, the larger the imputed segment of the population became the more this assumption had to be relied upon. Since the real people underlying any imputations are not observed by the census, the assumption is, in its strictest sense, untestable, thus we maintain that increasing the number of imputations included in the census may have made generalizing the coverage error estimates to the total census population less reliable. 3. The Treatment of Duplicate Enumerations in the Reinstated Housing Units--The Department wrote that our draft report incorrectly characterized the effects of reinstating duplicates into the census. The Department indicated that A.C.E., having been designed to measure net coverage error, treated the over 1 million likely duplicates "exactly correctly" and that including them in the census had no effect on the mathematical estimates of coverage error produced by A.C.E. We reported that, according to Bureau research, introducing the additional duplicates into the census appeared to have no impact on the A.C.E. estimates. But we view this fact as evidence of a limitation, or blind spot, in the Bureau's coverage evaluation. The fact that 2.4 million records, containing disproportionately over 1 million duplicate people could be added to the census without affecting the A.C.E. estimates demonstrates a practical limitation of those coverage error estimates. We maintain that the resultant measure of coverage error cannot be reliably generalized to the entire population count of which those 1 million duplicates are a part. 4. Size of the Search Area--The Department wrote that a search area like that used in 1990 would have done little to better measure the number of students and people with vacation homes who may have been duplicated in 2000. It described our conclusion regarding the reduction in search area from 1990 as not supported by the relative magnitudes of these situations. And finally, the Department offered additional support for the decision to reduce the search area by describing the reduced search area as balanced, or "in an expected value sense" [emphasis added] affecting the number of extra missed people and the extra miscounted people equally. In our final report we added a statement about the Department's concern over the importance of balance in its use of search areas. But we disagree that our conclusion is unsupported, since in our draft report we explicitly cited Bureau research that found an additional 1.2 million duplicate enumerations in units that were out-of-scope for 2000 A.C.E. but that would have been in-scope for 1990's coverage evaluation. In addition, the Department offered several other comments. Regarding our finding that the Bureau has not produced reliable revised estimates of coverage error for the 2000 Census, and, specifically, that the full impact of the Bureau's methodological limitations on the revised estimates has not been made clear, the Department wrote that the Census Bureau feels that further evaluations would not be a wise use of resources. We concur, which is why our recommendations look forward to the Bureau's preparation for 2010. The Department commented that it did not see how we could draw conclusions about the reliability of the Bureau's coverage evaluation estimates if we did not audit the underlying research, data, or conclusions. We maintain that the objectives and scope of our review did not require such an audit. As we described, and at times cited, throughout our draft report, we used the results of the Bureau's own assessment of the 2000 Census and its coverage evaluation. That information was sufficient to draw conclusions about the reliability of the A.C.E. estimates. As a result, there was no need to verify individual Bureau evaluations and methodologies. The Department expressed concern that our draft report implied that the unexpected differences in patterns of coverage error between the housing and the population count were irreconcilable. That was not our intent, and we have clarified that in the report. The Department expressed concern over the report's characterization of the 1990 coverage error estimates for group quarters as weak in part due to the high mobility of this population. However, the 1990 group quarters estimates are described as "weak" in a Bureau memorandum proposing that group quarters be excluded from the 2000 coverage evaluation. The memorandum also explains how the mobility within the group quarters population contributes to the resulting estimates. We have not revised the characterization of the group quarters coverage error estimates or the causal link due to the mobility of that population, but we have revised our text to state more clearly that the 1990 estimates being discussed are those for group quarters. As agreed with your offices, unless you release its contents earlier, we plan no further distribution of this report until 30 days from its issue date. At that time we will send copies to other interested congressional committees, the Secretary of Commerce, and the Director of the U.S. Census Bureau. Copies will be made available to others upon request. This report will also be available at no charge on GAO's Web site at [Hyperlink, http://www.gao.gov]. If you or your staff have any questions concerning this report, please contact me on (202) 512-6806 or by e-mail at daltonp@gao.gov or Robert Goldenkoff, Assistant Director, at (202) 512-2757 or goldenkoffr@gao.gov. Key contributors to this report were Ty Mitchell, Amy Rosewarne, and Elan Rohde. Signed by: Patricia A. Dalton: Director, Strategic Issues: [End of section] Appendixes: Appendix I: Comments from the Department of Commerce: UNITED STATES DEPARTMENT OF COMMERCE: The Under Secretary for Economic Affairs: Washington, D.C. 20230: SEP 10 2004: Ms. Patricia A. Dalton: Director, Strategic Issues: U.S. Government Accountability Office: Washington, DC 20548: Dear Ms. Dalton: The U.S. Department of Commerce appreciates the opportunity to comment on the U.S. Government Accountability Office draft report entitled 2000 Census: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates (GAO-04-468). The Department's comments on this report are enclosed. Sincerely, Signed by: Kathleen B. Cooper: Enclosure: Comments from the U.S. Department of Commerce Regarding the U.S. Government Accountability Office (GAO) Draft Report Entitled 2000 Census: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates (GAO-04-468): Thank you for providing us the opportunity to comment on your draft report 2000 Census: Design Choices Contributed to Inaccuracy of Coverage Evaluation Estimates (GAO-04-468). Response to GAO Recommendations: We concur with the GAO's recommendations regarding planning for the 2010 census. However, we have several comments on the analyses and conclusions in this report. Major Comments on GAO's Analyses and Conclusions: First, the Accuracy and Coverage Evaluation Program (A.C.E.) was designed explicitly and very publicly to measure net census coverage error. The GAO report implies that the A.C.E. should have had different goals-specifically, that the A.C.E. should have been designed to measure gross coverage error components, including providing a direct measure of census duplicates. While the U.S. Census Bureau now has made this a goal for 2010, it is misleading to imply that the 2000 census A.C.E. was inaccurate because it chose to measure net rather than gross coverage errors. Second, as the Census Bureau has stated and documented elsewhere, the estimates produced by the A.C.E. in March 2001 were not an accurate measure of census coverage. This inaccuracy occurred in spite of the outstanding execution of the design by the Census Bureau's field staff, clerical processing staff, and information technology staff. Therefore, we agree with the GAO's primary conclusion that the inaccuracy can only be attributed to the assumptions and decisions that underlay the design. However, it is important for future coverage measurement programs that the true design failures be understood. The GAO report points to five design decisions as critical to the lack of success of the A.C.E. The Census Bureau agrees that we had difficulties measuring residency status, but we cannot agree with the GAO's analysis of the other four design decisions nor with its conclusion. In fact, the Census Bureau views this report as misleading in its descriptions of the nature and source of the A.C.E. errors and believes that the GAO's conclusions are based on Census Bureau research results that have been taken out of context. The design of the A.C.E. was based in part on a key assumption-that through field interviewing, each person could be assigned to a correct (or at least consistently to a unique) location at which the person should have been enumerated. In the case of people enumerated at two different locations, the design relied in most cases upon information provided by respondents to determine which location should be considered the correct Census Day address and which should be considered an erroneous enumeration. It is important to note that, because of the statistical nature of the survey, each determination was made separately. This design, and that underlying assumption, seemed to have worked reasonably well in previous censuses. However, based on subsequent analysis, it is now clear that for the 2000 A.C.E., this assumption frequently did not hold. As a result, we found a number of cases where, for a person enumerated at two different locations, our separate determinations of census day address both concluded that the person was correctly enumerated and thus did not detect an erroneous enumeration. Please see the Executive Steering Committee for A.C.E. Policy II (ESCAP) Report 6 entitled "ESCAP II: Census Person Duplication and the Corresponding A.C.E. Enumeration Status." This report can be found at http://www.census.gov/dmd/www/pdf/Rgport6.pdf The Census Bureau remains convinced that the failure of this assumption (that we could gather accurate information about Census Day residence) was the principal cause of A.C.E. inaccuracy in 2000. The GAO report, however, lists four other design decisions (which they refer to as "blind spots") as the cause of A.C.E. inaccuracy: (1) the treatment of the group quarters population, (2) the treatment of census imputations, (3) the treatment of duplicate enumerations in the reinstated housing units, and (4) size of the search area. The Census Bureau, however, does not believe that any of these decisions contributed significantly to the inaccuracy of the A.C.E. (1) The Treatment of the Group Quarters Population-The GAO report correctly notes that non-institutional group quarters residents (e.g., students in college dorms) were included in the 1990 post-enumeration survey (PES) universe, but excluded from the 2000 A.C.E. universe. It is also true that analysis done after 2000 indicates a large number of people duplicated between the housing unit and the group quarters population-especially college students. However, what the report fails to make clear is that the A.C.E. was designed to measure such duplicates. Specifically, the design called for careful questioning of the student or his/her family to determine whether he/she was away at college. Evidence now indicates that this questioning failed to reveal that many students were counted at home, but should have been counted at college. There was never an intention to directly match between the home address and all possible college residences. This was certainly not done in 1990. In any case, including group quarters in the universe for A.C.E. would have had no effect on how persons responded to the A.C.E. questioning described above. Our collection and treatment of those responses-not the universe definition-are what led to the errors in the A.C.E. estimates. (2) The Treatment of Census Imputations-The Census Bureau agrees with the finding that there were more whole-person records imputed in Census 2000 than in the 1990 census, and that this is something that the Census Bureau must research. However, the A.C.E. was properly designed to account for this effect. The A.C.E. was explicitly designed to measure net census coverage error, including the effects on net error of the census imputations. There was no attempt to determine the accuracy of individual census imputations. There is no basis for GAO's claim that "as more imputations are included in the census total, the generalized coverage error estimates become less reliable." (3) The Treatment of Delicate Enumerations in the Reinstated Housing Units-The GAO report is particularly incorrect in its characterization of the effect of any duplicate enumeration reinstated to the census files after A.C.E. processing. Specifically, the following statement by the GAO cannot be sustained: "While the Bureau maintains that the reinstatement of the over 1 million reinstated likely duplicates did not affect the A.C.E. estimate in a statistically significant way, this suggests that the A.C.E. estimate itself is blind to the presence of the over 1 million duplicates the Bureau introduced." Again, the A.C.E. was designed explicitly to measure the net coverage error. Although not specifically identified as duplicates, these "over 1 million" cases were treated exactly correctly. Indeed, setting these duplicate cases aside for the purpose of dual system estimation of net error is exactly how they would and should have been treated even if they had been separately identified. The mathematical treatment is identical, with the exception that by setting aside these 1 million cases directly, sampling error actually was reduced. (4) Size of the Search Area-As noted by the GAO, the effective search area (for matching Census and A.C.E. interviews) was reduced from what it had been in 1990. However, this would have affected both the number of census omissions we measured and the number of census erroneous enumerations (including duplicates) we detected. Because, in an expected value sense, the reduced search area would have affected these components equally, this was a reasonable decision given our goal of measuring the net coverage error. Expanding the search area certainly would have resulted in a few more duplicate enumerations being detected and measured, but also would have resulted in a few less omissions being measured. In any case, the additional duplicates identified by expanding the search area only would be those situations where both enumerations were geographically near each other. An effective search area like that used in 1990 would have done little to better measure the number of students counted both at home and at college, or people counted both at their usual home and at a vacation home. The GAO's conclusion (that the size of the search area was a leading cause of A.C.E. inaccuracy in 2000) is not supported by the relative magnitudes of these various situations. Furthermore, the size of the search area is not the relevant issue. Rather, what matters is whether there was a consistent treatment of the search area used to identify missed people and erroneous enumerations. Any inconsistencies could result in balancing error, which was identified as a potential problem in the first Executive Steering Committee For A.C.E. Policy (ESCAP) report and subsequently was extensively evaluated. As noted in the second ESCAP report, these evaluations indicated that the dual-system estimates may have overstated the net undercount by 450,000 persons and that this was mainly a result of geocoding errors. In summary, while other-and equally reasonable-decisions could have been made with respect to these four "blind spots," none of these design choices we made for 2000 were the primary cause of A.C.E. inaccuracy. Other Comments: The GAO report states that with respect to the A.C.E. Revision II estimates, the Census Bureau "has not made the full impact of these methodological limitations on the data clear" [Page 16]. In response to this conclusion, the Census Bureau believes it is important to review the history of our efforts. The Census Bureau conducted Census 2000 and evaluated it through the A.C.E. The Census Bureau then conducted an evaluation of the A.C.E. When this evaluation raised questions, it itself was evaluated. Then, taking into account all previous evaluations, the Revision II estimates were prepared and their limitations carefully, thoroughly, and publicly documented. Although the Census Bureau may not have explicitly quantified any remaining non- sampling errors, the A.C.E. evaluations, research, and revision efforts have provided valuable information about erroneous census enumerations and have identified major issues with data collection for both the census and the coverage measurement survey. The Census Bureau feels that further evaluations would not be a wise use of resources. Although the GAO notes "we did not audit the Bureau's research, the underlying data or its conclusions," the report concludes that the A.C.E. Revision II estimates "are not reliable." We do not see how this conclusion can be supported, particularly if the GAO did not audit our research or data. The report raises some interesting questions concerning the differences between the person coverage estimates and the housing unit coverage estimates. These differences warrant further careful study. However, the report implies that these differences are, prima facie, unreconcilable, because "the Bureau expected to see across demographic groups similar patterns between the coverage error for the count of the population and the count of housing units." This ignores within- household error. Net undercount for the Black population can be fully consistent with net overcount for the Black housing unit count, possibly due to high within-household omissions. Net overcount of the non-Hispanic white and other population can be fully consistent with an undercount of the housing units, due to, for example, large duplication of college students. The report states that "the Bureau reported that the high mobility of these [Group Quarters] population made it more difficult to count them, thus the 1990 estimates of coverage were weak" [page 12]. It is true that the high mobility of this population made 1990 matching more difficult, and that there is more uncertainty in the coverage estimates for this segment of the population. It does not follow that having included a highly mobile and difficult-to-count group significantly weakened the overall 1990 estimates. The move to exclude this group in 2000 grew mainly from a decision made on other grounds to treat movers differently in 2000 than was done in 1990, although there were concerns about the difficulties associated with this group. [End of section] Appendix II: Bureau Design Decisions Increased Imputations: The Bureau made various design decisions that resulted in an increase in the number of "imputations"--or people guessed to exist--included in the census population that could not be included within the A.C.E. sample survey. The Bureau believes certain numbers of people exist despite the fact that the census records no personal information on them; thus it projects, via computer-executed algorithms, numbers and characteristics of people and includes them in the census. Such records are simply added to the census totals, and do not have names attached to them. Thus it was impossible for A.C.E. to either count imputed individuals using the A.C.E. sample survey or incorporate them into the matching process. Since the true number and the characteristics of these persons are unknown, matching nameless records via A.C.E. would not have provided any meaningful information on coverage evaluation. The number of people the Bureau imputed grew rapidly in 2000, from about 2 million in 1990 to about 5. One of the reasons for the large increase in imputations may be the decision by the Bureau to eliminate field edits--the last-minute follow-up operation to collect additional information from mail-back forms that had too little information on them to continue processing-- from field follow-up in 2000. While acknowledging that this decision may have increased imputations for 2000, a senior Bureau official justified the decision by describing the field edits in 1990 as providing at times a "clerical imputation" that introduced a subjective source of error, which computer-based imputation in 2000 lacked. The Bureau also reduced the number of household members for whom personal information could be provided on standard census forms, and this also contributed to the increase in imputations. Households reporting a household size greater than 6 in 2000--the number for whom personal characteristics could be provided--were to be automatically contacted by the Bureau to collect the additional information. Yet not all large households could be reached for the additional information, and the personal characteristics of the remaining household members needed to be imputed. Again, A.C.E. would have been unable to match people counted by its sample survey to imputations, so imputed people were excluded from A.C.E. calculations of coverage errors. An A.C.E. design choice by the Bureau that likely increased the amount of data imputed within the A.C.E. sample survey was how the Bureau decided to account for people who moved between Census Day and the day of the A.C.E. interview. Departing from how movers were dealt with in 1990, and partly to accommodate the initial design for the 2000 Census, which relied on sampling nonrespondents to the census, for 2000 the Bureau relied on the counts of the people moving into A.C.E. sample areas to estimate the number of matched people who had actually lived in the A.C.E. areas on Census Day but moved out. This decision resulted in the Bureau having less complete information about the Census Day residents in A.C.E. sample areas who had moved out, and likely increased the number of imputations that were later required, making it more difficult to match these moving persons to the census. A Bureau official also cited this decision as partial justification for not including group quarters in A.C.E. search areas. The extent that imputation affected the accuracy of the census is unknown. The National Academy of Sciences discussed in an interim report on the 2000 Census the possibility of a subset of about 1.2 million of these imputations being duplicates.[Footnote 8] That report stated that, for example, "it is possible that some of these cases-- perhaps a large proportion--were erroneous or duplicates," and described another subset of about 2.3 million that could include duplicates. However, this Academy report did not include any data to suggest the extent of duplicates within these groups, and it may similarly have been possible for the number of persons in this group to have been underestimated. The Bureau maintains that the imputations were necessary to account for the people its field operations led it to believe had been missed, and that its imputation methods do not introduce statistical bias. [End of section] Appendix III: Bureau Estimates of Population and Housing Unit Net Undercounts: As shown in Table 1, the initial A.C.E. results suggested that the differential population undercounts of non-Hispanic blacks and Hispanics--the difference between their undercount estimate and that of the majority groups--persisted from Bureau estimates from its coverage evaluation in 1990. Yet they also demonstrated that the Bureau had apparently succeeded in reducing the magnitude of those differences since its evaluation of the 1990 Census.[Footnote 9] Table 1: Estimated Percent Population Undercounts for Selected Race/ Origin Groups: Non-Hispanic black; Estimate in % (standard error in %): 1990 Estimates[A]: 4.57 (0.55); Estimate in % (standard error in %): A.C.E. (March 2001): 2.17 (0.35); Estimate in % (standard error in %): Revision I (October 2001): 0.78 (0.45); Estimate in % (standard error in %): Revision II (March 2003): 1.84 (0.43). Hispanic; Estimate in % (standard error in %): 1990 Estimates[A]: 4.99 (0.82); Estimate in % (standard error in %): A.C.E. (March 2001): 2.85 (0.38); Estimate in % (standard error in %): Revision I (October 2001): 1.25 (0.54); Estimate in % (standard error in %): Revision II (March 2003): 0.71 (0.44). Other; Estimate in % (standard error in %): 1990 Estimates[A]: --; Estimate in % (standard error in %): A.C.E. (March 2001): 0.73 (0.14); Estimate in % (standard error in %): Revision I (October 2001): -0.28 (0.20); Estimate in % (standard error in %): Revision II (March 2003): --. Non-Hispanic white; Estimate in % (standard error in %): 1990 Estimates[A]: 0.68 (0.22); Estimate in % (standard error in %): A.C.E. (March 2001): --; Estimate in % (standard error in %): Revision I (October 2001): --; Estimate in % (standard error in %): Revision II (March 2003): -1.13 (0.20). Total; Estimate in % (standard error in %): 1990 Estimates[A]: 1.61 (0.20); Estimate in % (standard error in %): A.C.E. (March 2001): 1.18 (0.13); Estimate in % (standard error in %): Revision I (October 2001): 0.06 (0.18); Estimate in % (standard error in %): Revision II (March 2003): -0.49 (0.20). Sources: A.C.E. and Revision I data are reported in Basis of "Revised Early Approximation" of Undercounts Released Oct. 17, 2001 (Oct. 26, 2001). 1990 and Revision II data are from PP-54 A.C.E. Revision II: Summary of Estimated Net Coverage (Dec. 31, 2002), Table 1. Notes: The reported net undercounts for 2000 Census are for household population only. The reported 1990 net undercount also covers the population living in noninstitutional, nonmilitary group quarters. [A] Tabulations by race/origin differed in 1990 from those for the 2000 Census. The estimates reported in this column were labeled in 1990 as "Black", "Hispanic", and "Non-Hispanic White & Other", which also included "American Indian off Reservation". [End of table] Subsequent revised results published in October 2001 for three race/ origin groups indicated that differential undercounts were generally lower than the initial A.C.E. estimates, but that only the undercount estimate for Hispanics was still statistically different from zero. Finally, the latest revised estimates of undercount reported in March 2003 that of these three major race/origin groups, only the non- Hispanic black and non-Hispanic white percentage undercounts were significantly different from zero, in addition to the national net overcount. Unlike the estimates of census population accuracy, which were revised twice since initial estimates, the census housing count accuracy estimates have not been revised and are based on the initial A.C.E. data. A subset of those results, including those provided here, were also published in October 2001. Table 2: Estimated Percent Occupied Housing Undercounts Differ from Population Undercounts for Selected Race/Origin Groups: Non-Hispanic black; Estimate in % (standard error in %): Housing undercount: -0.45 (0.29); Estimate in % (standard error in %): Population undercount: 1.84 (0.43). Hispanic; Estimate in % (standard error in %): Housing undercount: 0.06 (0.35); Estimate in % (standard error in %): Population undercount: 0.71 (0.44). Non-Hispanic white or "some other race"; Estimate in % (standard error in %): Housing undercount: 0.38 (0.14); Estimate in % (standard error in %): Population undercount: -1.13 (0.20). Total; Estimate in % (standard error in %): Housing undercount: 0.27 (0.13); Estimate in % (standard error in %): Population undercount: - 0.49 (0.20). Source: DSSD A.C.E. REVISION II MEMORANDUM SERIES #PP-50 Comparison of A.C.E. Revision II Population Coverage Results with HUCS Housing Coverage Results J. Gregory Robinson Population Division Glenn S. Wolfgang Decennial Statistical Studies Division (December 31, 2002). Notes: The reported net housing undercounts are for housing units only and do not include group quarters. Net undercounts reported here are referred to by the Bureau as "single cell dual system estimates". [End of table] [End of section] Glossary: This glossary is provided for reader convenience, not to provide authoritative or complete definitions. Accuracy and Coverage Evaluation: The Bureau's Accuracy and Coverage Evaluation (A.C.E.) program was intended to measure coverage error (see below) for the 2000 Decennial Census. The program was to enable the Bureau to more accurately estimate the rate of coverage error via a sample survey of select areas nationwide, and if warranted, to use the results of this survey to adjust census estimates of the population for nonapportionment purposes. Census Adjustment: The use of statistical information to adjust official census data. Count of Housing: A tally of certain kinds of dwellings, including single-family homes, apartments, and mobile homes, along with demographic information on the inhabitants. Count of Population: The headcount of everybody in the nation, regardless of their dwelling. Coverage Error: The extent that minority groups are over-or undercounted in comparison to other groups in the census. Coverage Evaluation: Statistical studies to evaluate the level and sources of coverage error in censuses and surveys. Duplicates: When the census erroneously counts a person more than once. Residence Rules: The rules the Bureau uses to determine where people should be counted. [End of section] Related GAO Products: 2010 Census: Cost and Design Issues Need to Be Addressed Soon. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-04-37] (Washington, D.C.: January 15, 2004). 2000 Census: Coverage Measurement Programs' Results, Costs, and Lessons Learned. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-287] (Washington, D.C.: January 29, 2003). 2000 Census: Complete Costs of Coverage Evaluation Programs Are Not Available. [Hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-03-41] (Washington, D.C.: October 31, 2002). 2000 Census: Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned. [Hyperlink, http:// www.gao.gov/cgi-bin/getrpt?GAO-02-297] (Washington, D.C.: March 14, 2002). 2000 Census: Coverage Evaluation Interviewing Overcame Challenges, but Further Research Needed. [Hyperlink, http://www.gao.gov/cgi-bin/ getrpt?GAO-02-26] (Washington, D.C.: December 31, 2001). (450176): FOOTNOTES [1] U.S. General Accounting Office, 2000 Census: Coverage Evaluation Interviewing Overcame Challenges, but Further Research Needed, GAO-02- 26 (Washington, D.C.: Dec. 31, 2001); and 2000 Census: Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned, GAO-02-297 (Washington, D.C.: Mar. 14, 2002). [2] GAO-04-37, pp. 43-44. [3] Constance F. Citro, Daniel L. Cork, and Janet L. Norwood, eds., The 2000 Census: Counting under Adversity (Washington, D.C.: The National Academies Press, 2004). [4] See GAO-02-297, pp. 9-12, for example. [5] The Bureau uses computer-executed algorithms to estimate imputations. [6] Executive Steering Committee For A.C.E. Policy II (ESCAP II) Report 20 October 11, 2001 ESCAP II: Person Duplication in Census 2000, Thomas Mule, p.iv. [7] Even so, it is not clear that all of these 1.2 million duplicates would have been identified had the search areas been expanded to completely include them, since the Bureau did not always identify duplicates within the search areas that it did use. [8] Constance F. Citro, Daniel L. Cork, and Janet L. Norwood, eds., The 2000 Census: Interim Assessment (Washington, D.C.: The National Academies Press, 2001). [9] More complete estimates of differential population undercounts are available on the Bureau's Internet Web site at www.census.gov. GAO's Mission: The Government Accountability Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO's Web site ( www.gao.gov ) contains abstracts and full-text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as "Today's Reports," on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to www.gao.gov and select "Subscribe to e-mail alerts" under the "Order GAO Products" heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office 441 G Street NW, Room LM Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: www.gao.gov/fraudnet/fraudnet.htm E-mail: fraudnet@gao.gov Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov (202) 512-4800 U.S. Government Accountability Office, 441 G Street NW, Room 7149 Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.