2000 Census

Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned Gao ID: GAO-02-297 March 14, 2002

The U.S. Census Bureau conducted the Accuracy and Coverage Evaluation (ACE) survey to estimate the number of people missed, counted more than once, or otherwise improperly counted in the 2000 Census. On the basis of uncertainty in the ACE results, the Bureau's acting director decided that the 2000 Census tabulations should not be adjusted in order to redraw the boundaries of congressional districts or to distribute billions of dollars in federal funding. Although ACE was generally implemented as planned, the Bureau found that it overstated census undercounts because of an error introduced during matching operations and other uncertainties. The Bureau concluded that additional review and analysis of these uncertainties would be needed before the data could be used. Matching more than 1.4 million census and ACE records involved the following four phases, each with its own matching procedures and multiple layers of review: computer matching, clerical matching, field follow-up, and clerical matching. The Bureau applied quality assurance procedures to each phase of person matching. Because the quality assurance procedures had failure rates of less than one percent, the Bureau reported that person matching quality assurance was successful at minimizing errors. Overall, the Bureau carried out person matching as planned, with few procedural deviations. GAO identified areas for improving future ACE efforts, including more complete documentation of computer matching decisions and better assurance that problems do not arise with the bureau's automated systems.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-02-297, 2000 Census: Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned This is the accessible text file for GAO report number GAO-02-297 entitled '2000 Census: Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned' which was released on March 14, 2002. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. United States General Accounting Office: GAO: Report to Congressional Requesters: March 2002: 2000 Census: Coverage Evaluation Matching Implemented as Planned, but Census Bureau Should Evaluate Lessons Learned: GAO-02-297: Contents: Letter: Results in Brief: Background: Matching Process Was Complex, and Application of Criteria Involved the Judgment of Trained Bureau Staff: Quality Assurance Results Suggest Person Matching Procedures Were Implemented as Planned: The Bureau Took Action to Address Some Deviations, but Effect on Matching Results Is Unknown: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendixes: Appendix I: Scope and Methodology: Appendix II: Comments from the Department of Commerce: Appendix III: GAO Contact and Staff Acknowledgments: Table: Table 1: Deviations from the Planned Person Matching Operation: Figures: Figure 1: A.C.E. Survey Followed Steps Similar to Census: Figure 2: Person Matching, Quality Assurance Coverage: Figure 3: Quality Assurance of Field Follow-up by A.C.E. Regional Office: [End of section] United States General Accounting Office: Washington, D.C. 20548: March 14, 2002: The Honorable Dave Weldon: Chairman: The Honorable Danny K Davis: Ranking Minority Member: Subcommittee on Civil Service, Census and Agency Organization: Committee on Government Reform: House of Representatives: The Honorable William Lacy Clay: The Honorable Carolyn B. Maloney: The Honorable Dan Miller: House of Representatives: To assess the quality of the population data collected in the 2000 Census, the U.S. Census Bureau conducted the Accuracy and Coverage Evaluation (A.C.E.) survey, a sample of persons designed to estimate the number of people missed, counted more than once, or otherwise improperly counted in the census. On the basis of uncertainty in the A.C.E. results, in separate decisions in March and October 2001, the acting director of the bureau decided that the 2000 Census tabulations should not be adjusted for purposes of redrawing the boundaries of congressional districts or for other purposes, such as distributing billions of dollars in federal funding. Although A.C.E. was generally implemented as planned, the bureau found that A.C.E. overstated census undercounts due in part to error introduced during matching operations and other remaining uncertainties. The bureau has reported that additional review and analysis on these remaining uncertainties would be necessary before any potential uses of these data can be considered. A critical component of the A.C.E. survey was the person matching operation, in which the bureau matched the persons counted in the A.C.E. survey to the persons counted in the census. The results of person matching formed the basis for statistical estimates of the proportions of the population missed or improperly counted by the census. This report, prepared at the request of the chairman and ranking minority member of the former House Subcommittee on the Census, reviews the person matching operation of A.C.E. We agreed to describe (1) the process and criteria involved in making an A.C.E. and census person match, (2) the quality assurance procedures used in the key person matching phases and the available results of those procedures, and (3) any deviations in the matching operation from what was planned. This report is the latest of several we have issued on lessons learned from the 2000 Census that can help inform the bureau's planning efforts for the 2010 Census. To address our three objectives, we examined relevant bureau program specifications, training manuals, office manuals, memorandums, and other progress and research documents. We also interviewed bureau officials at bureau headquarters in Suitland, Md., and the bureau's National Processing Center in Jeffersonville, Ind., which was responsible for the planning and implementation of the person matching operation. Further scope and methodological details are given in appendix I. We performed our audit work from September 2000 through April 2001 in accordance with generally accepted government auditing standards. On January 4, 2002, we requested comments on a draft of this report from the secretary of commerce. On February 13, 2002, the secretary of commerce forwarded written comments from the bureau (see appendix II), which we address in the "Agency Comments and Our Evaluation" section of this report. Results in Brief: Matching over 1.4 million census and A.C.E. records was a complex and often labor-intensive process that consisted of four phases, each with its own matching procedures and multiple layers of review. The four phases were as follows. * Computer matching, which took pairs of A.C.E. and census records and compared certain personal characteristics such as last name and age. The computer assigned a match score to each pair of records based on the extent to which the characteristics aligned. Experienced bureau staff then judgmentally determined cutoff scores to separate the groups of records that would be coded as a "match," "possible match," or one of a number of codes that defines them as not matched. However, bureau staff did not document the criteria they used to determine the cutoffs. As a result, future bureau staff may not benefit from the lessons learned by current staff about how cutoff scores are applied. * Clerical matching (first phase), in which over 250 trained bureau staff reviewed all records and attempted to link those records left unmatched in the previous phase, in part by matching records that contained abbreviations and spelling differences. * Field follow-up, in which bureau interviewers visited households where additional information was needed to assign match codes to a pair of records. * Clerical matching (second phase), in which clerks used information obtained from field follow-up to match and conduct a final review of records. The bureau coded as "unresolved" records without enough information to be coded otherwise. The bureau then used statistical imputation methods to assign a match code to records coded as "unresolved," based on an examination of the results of similar records for which the bureau was able to assign a match code. While some imputation is unavoidable, it introduces uncertainty into the estimates of census over- or undercount rates. The bureau applied quality assurance procedures to each phase of person matching. For example, during the field follow-up phase, supervisors and office staff were to review each questionnaire for legibility and completeness. In addition, A.C.E. regional offices were to reinterview a random sample of 5 percent of the households to ensure that enumerators had not falsified data. Because the quality assurance procedures had failure rates of less than 1 percent , the bureau reported that person matching quality assurance was successful at minimizing errors. Overall, the bureau carried out person matching as planned, with few procedural deviations. The operation deviated somewhat from what was planned as a result of programming errors, printing problems, and events that triggered delays. Although the bureau addressed these deviations and person matching continued, in some cases the effect the deviations had on person matching is unknown . For example, because of printing and other problems, pages and names were missing from some of the follow-up questionnaires, and a section that verified whether the person being matched was in the geographic sample area was incomplete in some others. The bureau was unable to document the extent, effect, or cause of the printing problems and coded incomplete questionnaires as "unresolved." Bureau officials believe that the effect of the deviations was small based on the timely actions taken to address them. Nevertheless, although the bureau has concluded that A.C.E. matching quality improved compared to that in 1990, the bureau has reported that matching error remained and contributed to an overstatement of the A.C.E. estimate of census undercounts. Furthermore, despite the improvement in matching reported by the bureau, A.C.E. results were not used to adjust the census because of these errors as well as other remaining uncertainties. Therefore, it will be important for the bureau to determine the impact of these operational deviations. Our review identified areas with opportunity for improving future A.C.E. efforts, including more complete documentation of computer matching decisions and better assurance that problems do not arise with the bureau's automated systems. Therefore, as part of the bureau's effort to isolate lessons learned from the 2000 Census and to prepare for the census in 2010, we recommend that the secretary of commerce direct the bureau to (1) document the criteria used during computer matching to determine the groups of matched, possibly matched, and nonmatched records, (2) determine why problems with some of its automated systems were not discovered prior to deployment, and (3) determine the effect that deviations from planned operations may have had on the matching results for affected records and thus the accuracy of A.C.E. estimates of census undercounts. The secretary of commerce forwarded written comments from the U.S. Census Bureau on a draft of this report. (See appendix II.) The bureau had no comments on the text of the report and agreed with, and is taking action on, two of our four recommendations. The bureau provided additional clarification on our other two recommendations. We comment further on the bureau's response in the "Agency Comments and Our Evaluation" section of this report. Background: From April 24 through September 11, 2000, the U.S. Census Bureau surveyed a sample of about 314,000 housing units (about 1.4 million census and A.C.E. records in various areas of the country, including Puerto Rico) to estimate the number of people and housing units missed or counted more than once in the census and to evaluate the final census counts. Temporary bureau staff conducted the surveys by telephone and in-person visits. The A.C.E. sample consisted of about 12,000 "clusters" or geographic areas that each contained about 20 to 30 housing units. The bureau selected sample clusters to be representative of the nation as a whole, relying on variables such as state, race and ethnicity, owner or renter, as well as the size of each cluster and whether the cluster was on an American Indian reservation. The bureau canvassed the A.C.E. sample area, developed an address list, and collected response data for persons living in the sample area on Census Day (April 1, 2000). Although the bureau's A.C.E. data and address list were collected and maintained separately from the bureau's census work, A.C.E. processes were similar to those of the census. Figure 1: A.C.E. Survey Followed Steps Similar to Census: [Refer to PDF for image: illustration] Census Operations: Develop Address List: * Field canvassing nationwide; * Receiving address files from U.S. Postal Service; * Soliciting feedback from local/tribal governments; (Census addresses in A.C.E. areas). Collect Response Data: * Mailing out mail-back of forms; * Hand-delivering mail-back forms; * Following up with non-respondents; * Following up on other types of cases. (Data for people found by Census in and around A.C.E. areas). Tabulate and Disseminate Data: To President to re-apportion seats in the U.S. House of Representatives; To states for redistricting and other purposes (13 USC 141); To federal government and other users for Federal funds allocation and other uses. A.C.E. Operations: Develop Address List: Field canvassing in A.C.E. sample areas. Housing unit matching (Census addresses in A.C.E. areas). Collect Response Data: Person interviewing; Person matching: * Computer matching; * Clerical matching (first phase); * Field follow-up; * Clerical matching (second phase). Estimate accuracy and coverage. Planning 2010 Census: Adjust? No: No adjustment; Yes: Go to next step. Tabulate and Disseminate Data: To President to re-apportion seats in the U.S. House of Representatives; To states for redistricting and other purposes (13 USC 141); To federal government and other users for Federal funds allocation and other uses. Source: U.S. Census Bureau documents. [End of figure] After the census and A.C.E. data collection operations were completed, the bureau attempted to match each person counted by A.C.E. to the list of persons counted by the census in the sample areas to determine the number of persons who lived in the sample area on Census Day. The results of the matching process, together with the characteristics of each person compared, provided the basis for statistical estimates of the number and characteristics of the population missed or improperly counted by the census. Correctly matching A.C.E. persons with census persons is important because errors in even a small percentage of records can significantly affect the undercount or overcount estimate. Matching Process Was Complex, and Application of Criteria Involved the Judgment of Trained Bureau Staff: Matching over 1.4 million census and A.C.E. records was a complex and often labor-intensive process. Although several key matching tasks were automated and used prespecified decision rules, other tasks were carried out by trained bureau staff who used their judgment to match and code records. The four phases of the person matching process were (1) computer matching, (2) clerical matching, (3) nationwide field follow-up on records requiring more information, and (4) a second phase of clerical matching after field follow-up.[Footnote 1] Each subsequent phase used additional information and matching rules in an attempt to match records that the previous phase could not link. Figure: Computer Matching: [Refer to PDF for image: illustration] Computer matching: * Record-linkage software; * Experienced bureau staff review. Clerical matching (first phase): Field follow-up: Clerical matching (second phase). [End of figure] Computer matching took pairs of census and A.C.E. records and compared various personal characteristics such as name, age, and gender. The computer then calculated a match score for the paired records based on the extent to which the personal characteristics were aligned. Experienced bureau staff reviewed the lists of paired records, sorted by their match scores, and judgmentally assigned cutoff scores. The cutoff scores were break points used to categorize the paired records into one of three groups so that the records could be coded as a "match," "possible match," or one of a number of codes that defines them as not matched. Computer matching successfully assigned a match score to nearly 1 million of the more than 1.4 million records reviewed (about 66 percent). Bureau staff documented the cutoff scores for each of the match groups. However, they did not document the criteria or rules used to determine cutoff scores, the logic of how they applied them, and examples of their application . As a result, the bureau may not benefit from the possible lessons learned on how to apply cutoff scores. When the computer links few records as possible matches, clerks will spend more time searching records and linking them. In contrast, when the computer links many records as possible matches, clerks will spend less time searching for records to link and more time unlinking them. Without documentation and knowledge of the effect of cutoff scores on clerical matching productivity, future bureau staff will be less able to determine whether to set cutoff scores to link few or many records together as possible matches. Figure: First Phase of Clerical Matching: [Refer to PDF for image: illustration] Computer matching: Clerical matching (first phase): * Automated matching tools; * Clerk review; * Technician review - Analyst review. Field follow-up: Clerical matching (second phase): [End of figure] During clerical matching, three levels of matchers”including over 200 clerks, about 40 technicians, and 10 experienced analysts or "expert matchers"”applied their expertise and judgment to manually match and code records. A computer software system managed the workflow of the clerical matching stages. The system also provided access to additional information, such as electronic images of census questionnaires that could assist matchers in applying criteria to match records. According to a bureau official, a benefit of clerical matching was that records of entire households could be reviewed together, rather than just individually as in computer matching. During this phase over a quarter million records (or about 19 percent) were assigned a final match code. The bureau taught clerks how to code records in situations in which the A.C.E. and census records differed because one record contained a nickname and the other contained the birth name. The bureau also taught clerks how to code records with abbreviations, spelling differences, middle names used as first names, and first and last names reversed. These criteria were well documented in both the bureau's procedures and operations memorandums and clerical matchers' training materials, but how the criteria were applied depended on the judgment of the matchers. The bureau trained clerks and technicians for this complex work using as examples some of the most challenging records from the 1998 Dress Rehearsal person matching operation. In addition, the analysts had extensive matching experience. For example, the 4 analysts that we interviewed had an average of 10 years of matching experience on other decennial census surveys and were directly involved in developing the training materials for the technicians and clerks. Figure: Field Follow-up: [Refer to PDF for image: illustration] Computer matching: Clerical matching (first phase): Field follow-up: * Questionnaires; * Temporary field staff interview; * Temporary field supervisory review; * A.C.E. regional office review. Clerical matching (second phase): [End of figure] The bureau conducted a nationwide field follow-up on over 213,000 records (or about 15 percent) for which the bureau needed additional information before it could accurately assign a match code. For example, sometimes matchers needed additional information to verify that possibly matched records were actually records of the same person, that a housing unit was located in the sample area on Census Day, or that a person lived in the sample area on Census Day. Field follow-up questionnaires were printed at the National Processing Center and sent to the appropriate A.C.E. regional office. Field follow-up interviewers from the bureau's regional offices were required to visit specified housing units and obtain information from a knowledgeable respondent. If the household member for the record in question still lived at the A.C.E. address at the time of the interview and was not available to be interviewed after six attempts, field follow-up interviewers were allowed to obtain information from one or more knowledgeable proxy respondents, such as a landlord or neighbor. Figure: Second Phase of Clerical Matching: [Refer to PDF for image: illustration] Computer matching: Clerical matching (first phase): Field follow-up: Clerical matching (second phase): * Automated matching tools; * Clerk review; * Technician review; * Analyst review. [End of figure] The second phase of clerical matching used the information obtained during field follow-up in an attempt to assign a final match code to records. As in the first phase of clerical matching, the criteria used to match and code records were well documented in both the bureau's procedures and operations memorandums and clerical matchers' training materials. Nevertheless, in applying those criteria, clerical matchers had to use their own judgment and expertise. This was particularly true when matching records that contained incomplete and inconsistent information, as noted in the following examples. * Different household members provided conflicting information. The census counted one person”the field follow-up respondent. A.C.E. recorded four persons”including the respondent and her daughter. The respondent, during field follow-up, reported that all four persons recorded by A. C.E. lived at the housing unit on Census Day. During the field follow-up interview, the respondent's daughter came to the house and disagreed with the respondent. The interviewer changed the answers on the field follow-up questionnaire to reflect what the daughter said”the respondent was the only person living at the household address on Census Day. The other three people were coded as not living at the household address on Census Day. According to bureau staff, the daughter's response seemed more reliable. * An interviewer's notes on the field follow-up questionnaire conflicted with recorded information. The census counted 13 people”including the respondent and 2 people not matched to A.C.E. records. A.C.E. recorded 12 people”including the respondent, 10 other matched people, and the respondent's daughter who was not matched to census records. The field follow-up interview attempted to resolve the unmatched census and A.C.E. people. Answers to questions on the field follow-up questionnaire verified that the daughter lived at the housing address on Census Day. However, the interviewer's notes indicated that the daughter and the respondent were living in a shelter on Census Day. The daughter was coded as not living at the household address on Census Day, while the respondent remained coded as matched and living at the household address on Census Day. According to bureau staff, the respondent should also have been coded as a person that did not live at the household address on Census Day, based on the notes on the field follow-up questionnaire. * A.C.E., census, or both counted people at the wrong address. The census counted two people”the respondent and her husband twice; once in an apartment and once in a business office that the husband worked in, both in the same apartment building. The A. C.E. did not record anyone at either location, as the residential apartment was not in the A.C.E. interview sample. The respondent, during field follow- up, reported that they lived at their apartment on Census Day and not at the business office. The couple had responded to the census on a questionnaire delivered to the business office. A census enumerator, following up on the "nonresponse" from the couple's apartment, had obtained census information from a neighbor about the couple. The couple, as recorded by the census at the business office address, was coded as correctly counted in the census. The couple, as recorded by the census at the apartment address, was coded as living outside the sample block. According to bureau staff, the couple recorded at the business office address were correctly coded, but the couple recorded at the apartment should have been coded as duplicates. * An uncooperative household respondent provided partial or no information. The census counted a family of four”the respondent, his wife, and two daughters. A.C.E. recorded a family of three”the same husband and wife, but a different daughter's name, "Buffy." The field follow-up interview covered the unmatched daughters”two from census and one from A.C.E. The respondent confirmed that the four people counted by the census were his family and that "Buffy" was a nickname for one of his two daughters, but he would not identify which one. The interviewer wrote in the notes that the respondent "was upset with the number of visits" to his house. "Buffy" was coded as a match to one of the daughters; the other daughter was coded as counted in the census but missed by A. C.E. According to bureau staff, since the respondent confirmed that "Buffy" was a match for one of his daughters”although not which one”and that four people lived at the household address on Census Day, they did not want one of the daughters coded so that she was possibly counted as a missed census person. Since each record had to have a code identifying whether it was a match by the end of the second clerical matching phase, records that did not contain enough information after field follow-up to be assigned any other code were coded as "unresolved." The bureau later imputed the match code results for these records using statistical methods. While imputation for some situations may be unavoidable, it introduces uncertainty into estimates of census over- or undercount rates. The following are examples of situations that resulted in records coded as "unresolved." * Conflicting information was provided for the same household. The census counted four people”a woman, an "unmarried partner," and two children. A.C.E. recorded three people”the same woman and two children. During field follow-up, the woman reported to the field follow-up interviewer that the "unmarried partner" did not really live at the household address, but just came around to baby-sit, and that she did not know where he lived on Census Day. According to bureau staff, probing questions during field follow-up determined that the "unmarried partner" should not have been coded as living at the housing unit on Census Day. Therefore, the "unmarried partner" was coded as "unresolved." * A proxy respondent provided conflicting or inaccurate information. The census counted one person”a female renter. A.C.E. did not record anyone. The apartment building manager, who was interviewed during field follow-up, reported that the woman had moved out of the household address sometime in February 2000, but the manager did not know the woman's Census Day address. The same manager had responded to an enumerator questionnaire for the census in June 2000 and had reported that the woman did live at the household address on Census Day. The woman was coded as "unresolved." Quality Assurance Results Suggest Person Matching Procedures Were Implemented as Planned: The bureau employed a series of quality assurance procedures for each phase of person matching. The bureau reported that person matching quality assurance was successful at minimizing errors because the quality assurance procedures found error rates of less than 1 percent. Computer Matching: Clerks were to review all of the match results to ensure, among other things, that the records linked by the computer were not duplicates and contained valid and complete names. Moreover, according to bureau officials, the software used to link records had proven itself during a similar operation conducted for the 1990 Census. The bureau did not report separately on the quality of computer matched records. Although there were no formal quality assurance results from computer matching, at our request the bureau tabulated the number of records that the computer had coded as "matched" that had subsequently been coded otherwise. According to the bureau, the subsequent matching process resulted in a different match code for about 0.6 percent of the almost 500,000 records initially coded as matched by the computer. Of those records having their codes changed by later matching phases, over half were eventually coded as duplicates and almost all of the remainder were rematched to someone else. Two Phases of Clerical Matching: Technicians reviewed the work of clerks and analysts reviewed the work of technicians primarily to find clerical errors that (1) would have prevented records from being sent to field follow-up, (2) could cause a record to be incorrectly coded as either properly or erroneously counted by the census, or (3) would cause a record to be incorrectly removed from the A.C.E. sample. Analysts' work was not reviewed. Clerks and technicians with error rates of less than 4 percent had a random sample of about 25 percent of their work reviewed, while clerks and technicians exceeding the error threshold had 100 percent of their work reviewed. About 98 percent of clerks in the first phase of matching had only a sample of their work reviewed. According to bureau data, less than 1 percent of match decisions were revised during quality assurance reviews, leading the bureau to conclude that clerical matching quality assurance was successful. Under certain circumstances, technicians and analysts performed additional reviews of clerks' and technicians' work. For example, if during the first phase of clerical matching a technician had reviewed and changed more than half of a clerk's match codes in a given geographic cluster, the cluster was flagged for an analyst to review all of the clerk and technician coding for that area. During the second phase, analysts were required to make similar reviews when only one of the records was flagged for their review. This is one of the reasons why, as illustrated in figure 2, these additional reviews were a much more substantial part of the clerks' and technicians' workload that was subsequently reviewed by more senior matchers. The total percentage of workload reviewed ranged from about 20 to 60 percent across phases of clerical matching, far in excess of the 11-percent quality assurance level for the bureau's person interviewing operation. Figure 2: Person Matching, Quality Assurance Coverage: [Refer to PDF for image: stacked vertical bar graph] Percentage of workload reviewed for both QA cases and Review of other cases is plotted for the following stage/phase of matching: First phase of clerical matching, clerk; First phase of clerical matching, technician; Second phase of clerical matching, clerk; Second phase of clerical matching, technician. Source: GAO analysis of U.S. Census Bureau data. [End of figure] Field Follow-up: The quality assurance plan for the field follow-up phase had two general purposes: (1) to ensure that questionnaires had been completed properly and legibly and (2) to detect falsification.[Footnote 2] Supervisors initially reviewed each questionnaire for legibility and completeness. These reviews also checked the responses for consistency. Office staff were to conduct similar reviews of each questionnaire. To detect falsification, the bureau was to review and edit each questionnaire at least twice and recontact a random sample of 5 percent of the respondents. As shown in figure 3, all 12 of the A.C.E. regional offices exceeded the 5 percent requirement by selecting more than 7 percent of their workload for quality assurance review, and the national rate of quality assurance review was about 10 percent. Figure 3: Quality Assurance of Field Follow-up by A.C.E. Regional Office: [Refer to PDF for image: combined vertical bar and line graph] Vertical bars depict eligible QA as a percentage of total work for the indicated regional offices. Line depicts percentage of random QA failing QA. Regional offices: Boston; New York; Philadelphia; Detroit; Chicago; Kansas City; Seattle; Charlotte; Atlanta; Dallas; Denver; Los Angeles. Source: GAO analysis of U.S. Census Bureau data. [End of figure] At the local level, however, there was greater variation. There are many reasons why the quality assurance coverage can appear to vary locally. For example, a local census area could have a low quality assurance coverage rate because interviewers in that area had their work reviewed in other areas, or the area could have had an extremely small field follow-up workload, making the difference of just one quality assurance questionnaire constitute a large percentage of the local workload. Seventeen local census office areas (out of 520 nationally, including Puerto Rico) had 20 percent or more of field follow-up interviews covered by the quality assurance program, and, at the other extreme, 5 local census areas had 5 percent or less of the work covered by the quality assurance program. Less than 1 percent of the randomly selected questionnaires failed quality assurance nationally, leading the bureau to report this quality assurance operation as successful. When recontacting respondents to detect falsification by interviewers, quality assurance supervisors were to determine whether the household had been contacted by an interviewer, and if it had not, the record of that household failed quality assurance. According to bureau data, about 0.8 percent of the randomly selected quality assurance questionnaires failed quality assurance nationally. This percentage varied between 0 and about 3 percent across regions. The Bureau Took Action to Address Some Deviations, but Effect on Matching Results Is Unknown: The bureau carried out person matching as planned, with only a few procedural deviations. Although the bureau took action to address these deviations, it has not determined how matching results were affected. As shown in table 1, these deviations included (1) census files that were delivered late, (2) a programming error in the clerical matching software, (3) printing errors in field follow-up forms, (4) regional offices that sent back incomplete questionnaires, and (5) the need for additional time to complete the second phase of clerical matching. It is unknown what, if any, cumulative effect these procedural deviations may have had on the quality of matching for these records or on the resultant A.C.E. estimates of census undercounts. However, bureau officials believe that the effect of the deviations was small based on the timely responses taken to address them. The bureau conducted reinterviewing and re-matching studies on samples of the 2000 A.C.E. sample and concluded that matching quality in 2000 was improved over that in 1990, but that error introduced during matching operations remained and contributed to an overstatement of A.C.E. estimates of the census undercounts. The studies provided some categorical descriptions of the types of matching errors measured, but did not identify the procedural causes, if any, for those errors. Furthermore, despite the improvement in matching reported by the bureau, A.C.E. results were not used to adjust the census due to these errors as well as other remaining uncertainties. The bureau has reported that additional review and analysis on these remaining uncertainties would be necessary before any potential uses of these data can be considered. Table 1: Deviations from the Planned Person Matching Operation: Deviation: Late delivery of census files; Corrective action taken: Bureau employees worked extra hours to make up the time; Effect on process: Computer matching was started 3 days later than scheduled and finished 1 day behind schedule. Deviation: Programming error in clerical matching software; Corrective action taken: The number of records to be completed between error rate calculations was modified twice in the software managing the quality assurance of clerical matching and the software problem was quickly fixed; Effect on process: Assignments of sampled or 100-percent review of clerks' and technicians' work were made manually for 2 days. Deviation: 1. Programming error caused errors in printing last names; 2. Other printing problems; Corrective action taken: 1. Printing of field follow-up questionnaires was suspended temporarily. The procedure was supplemented. 2. No action taken because bureau staff viewed it as insignificant; Effect on process: 1. Extra steps were taken during matching for 5 percent of records. This slowed each region's questionnaire processing for 1 to 4 days. 2. The effect is unknown, but bureau staff viewed it as insignificant. Deviation: Regional offices sent back incomplete field follow-up questionnaires that contained a section to verify whether a housing unit was in the A.C.E. sample; Corrective action taken: Forty-eight incomplete field follow-up questionnaires were returned to the regional offices during the first 6 days of the second clerical matching phase; Effect on process: The effect is unknown because the total number of questionnaires with this section incomplete is not known. Deviation: Extra time was needed to complete the second phase of clerical matching; Corrective action taken: The schedule for the second phase of clerical matching was extended; Effect on process: Subsequent A.C.E. operations had to make up the time. [End of table] Late Delivery of Census Files Delayed Computer Matching Start: The computer matching phase started 3 days later than scheduled and finished 1 day late due to the delayed delivery of census files. In response, bureau employees who conducted computer matching worked overtime hours to make up lost time. Furthermore, A.C.E. regional offices did not receive clusters in the prioritized order that they had requested. The reason for prioritizing the clusters was to provide as much time as possible for field follow-up on clusters in the most difficult areas. Examples of areas that were expected to need extra time were those with staffing difficulties, larger workloads, or expected weather problems. Based on the bureau's Master Activities Schedule, the delay did not affect the schedule of subsequent matching phases. Also, bureau officials stated that although clusters were not received in prioritized order, field follow-up was not greatly affected because the first clerical matching phase was well staffed and sent the work to regional offices quickly. Programming Error and Analyst Backlog Required Software Modifications during Clerical Matching: On the first full day of clerical matching, the bureau identified a programming error in the quality assurance management system, which made some clerks and technicians who had not passed quality assurance reviews appear to have passed. In response, bureau officials manually overrode the system. Bureau officials said the programming error was fixed within a couple of days, but could not explain how the programming error occurred. They stated that the software system used for clerical matching was thoroughly tested, although it was not used in any prior censuses or census tests, including the Dress Rehearsal. As we have previously noted, programming errors that occur during the operation of a system raise questions about the development and acquisition processes used for that system.[Footnote 3] Field Follow-up Questionnaires Contained Printing Errors: A programming error caused last names to be printed improperly on field follow-up forms for some households containing multiple last names. In situations in which regional office staff may not have caught the printing error and interviewers may have been unaware of the error”such as when those questionnaires were completed before the problem was discovered”interviews may have been conducted using the wrong last name, thus recording misleading information. According to bureau officials, in response, the bureau (1) stopped printing questionnaires on the date officials were notified about the misprinted questionnaires, (2) provided information to regional offices that listed all field follow-up housing units with multiple names that had been printed prior to the date the problem was resolved, and (3) developed procedures for clerical matchers to address any affected questionnaires being returned that had not been corrected by regional office staff. While resolving the problem, productivity was initially slowed in the A.C.E. regional offices for approximately 1 to 4 days, yet field follow-up was completed on time. Bureau officials inadvertently introduced this error when they addressed a separate programming problem in the software. Bureau officials stated that they tested this software system; however, the system was not given a trial run during the Census Dress Rehearsal in 1998. According to bureau officials, the problem did not affect data quality because it was caught early in the operation and follow-up forms were edited by regional staff. However, the bureau could not determine the exact day of printing for each questionnaire and thus did not know exactly which households had been affected by the problem. According to bureau data, the problem could have potentially affected over 56,000 persons, or about 5 percent of the A.C.E. sample. In addition to the problem printing last names, the bureau experienced other printing problems. According to bureau staff, field follow-up received printed questionnaires that were (1) missing pages, (2) missing reference notes written by clerical matchers, and (3) missing names and/or having some names printed more than once for some households of about nine or more people. According to bureau officials, these problems were not resolved during the operation because they were reported after field follow-up had started and the bureau was constrained by deadlines. Bureau officials stated that they believed that these problems would not significantly affect the quality of data collected or match code results, although bureau officials were unable to provide data that would document either the extent, effect, or cause of these problems. Regional Offices Sent Back Incomplete Field Follow-up Questionnaires: The bureau's regional offices submitted questionnaires containing an incomplete "geocoding" section. This section was to be used in instances when the bureau needed to verify whether a housing unit (1) existed on Census Day and (2) was correctly located in the A.C.E. sample area. Although the bureau returned 48 questionnaires during the first 6 days of the operation to the regional offices for completion, bureau officials stated that after that they no longer returned questionnaires to the regional offices because they did not want to delay the completion of field follow-up. A total of over 10,000 questionnaires with "geocoding" sections were initially sent to the regional offices. The bureau did not have data on the number, if any, of questionnaires that the regional offices submitted incomplete beyond the initial 48. The bureau would have coded as "unresolved" the persons covered by any incomplete questionnaires. As previously stated, the bureau later imputed the match code results for these records using statistical methods, which could introduce uncertainty into estimates of census over- or undercount rates. According to bureau officials, this problem was caused by (1) not printing a checklist of all sections that needed to be completed by interviewers, (2) no link from any other section of the questionnaire to refer interviewers to the "geocoding" section, and (3) field supervisors following the same instructions as interviewers to complete their reviews of field follow-up forms. However, bureau officials believed that the mistake should have been caught by regional office reviews before the questionnaires were sent back for processing. Extra Time Was Needed to Complete the Second Phase of Clerical Matching: About a week after the second clerical matching phase began, officials requested an extension, which was granted for 5 days, to complete the second clerical matching phase. According to bureau officials, the operation could have been completed by the November 30, 2000, deadline as planned, but they decided to take extra steps to improve data quality that required additional time. According to bureau officials, the delay in completing person matching had no effect on the final completion schedule, only the start of subsequent A.C.E. processing operations. Conclusions: Matching A.C.E. and census records was an inherently complex and labor- intensive process that often relied on the judgment of trained staff, and the bureau prepared itself accordingly. For example, the bureau provided extensive training for its clerical matchers, generally provided thorough documentation of the process and criteria to be used in carrying out their work, and developed quality assurance procedures to cover its critical matching operations. As a result, our review identified few significant operational or procedural deviations from what the bureau planned, and the bureau took timely action to address them. Nevertheless, our work identified opportunities for improvement. These opportunities include a lack of written documentation showing how cutoff scores were determined and programming errors in the clerical matching software and software used to print field follow-up forms. Without written documentation, the bureau will be less likely to capture lessons learned on how cutoff scores should be applied, in order to determine the impact on clerical matching productivity. Moreover, the discovery of programming errors so late in the operation raises questions about the development and acquisition processes used for the affected A.C.E. computer systems. In addition, one lapse in procedures may have resulted in incomplete geocoding sections verifying that the person being matched was in the geographic sample area. The collective effect that these deviations may have had on the accuracy of A.C.E. results is unknown. Although the bureau has concluded that A.C.E. matching quality improved compared to 1990, the bureau has reported that error introduced during matching operations remained and contributed to an overstatement of the A.C.E. estimate of census undercounts. To the extent that the bureau employs an operation similar to A.C.E. to measure the quality of the 2010 Census, it will be important for the bureau to determine the impact of the deviations and explore operational improvements, in addition to the research it might carry out on other uncertainties in the A.C.E. results. Recommendations for Executive Action: As the bureau documents its lessons learned from the 2000 Census and continues its planning efforts for 2010, we recommend that the secretary of commerce direct the bureau to take the following actions: 1. Document the criteria and the logic that bureau staff used during computer matching to determine the cutoff scores for matched, possibly matched, and unmatched record pairs. 2. Examine the bureau's system development and acquisition processes to determine why the problems with A.C.E. computer systems were not discovered prior to deployment of these systems. 3. Determine the effect that the printing problems may have had on the quality of data collected for affected records, and thus the accuracy of A.C.E. estimates of the population. 4. Determine the effect that the incomplete geocoding section of the questionnaires may have had on the quality of data collected for affected records, and thus the accuracy of A.C.E. estimates of census undercounts. The secretary of commerce forwarded written comments from the U.S. Census Bureau on a draft of this report. (See appendix II.) The bureau had no comments on the text of the report and agreed with, and is taking action on, two of our four recommendations. In responding to our recommendation to document the criteria and the logic that bureau staff used during computer matching to determine cutoff scores, the bureau acknowledged that such documentation may be informative and that such documentation is under preparation. We look forward to reviewing the documentation when it is complete. In responding to our recommendation to examine system development and acquisition processes to determine why problems with the A.C.E. computer systems were not discovered prior to deployment, the bureau responded that despite extensive testing of A.C.E. computer systems, a few problems may remain undetected. The bureau plans to review the process to avoid such problems in 2010, and we look forward to reviewing the results of their review. Finally, in response to our two recommendations to determine the effects that printing problems and incomplete questionnaires had on the quality of data collected and the accuracy of A.C.E. estimates, the bureau responded that it did not track the occurrence of these problems because the effects on the coding process and accuracy were considered to be minimal since all problems were identified early and corrective procedures were effectively implemented. In our draft report we recognized that the bureau took timely corrective action in response to these and other problems that arose during person matching. Yet we also reported that bureau studies of the 2000 matching process had concluded that matching error contributed to error in A.C.E. estimates without identifying procedural causes, if any. Again, to the extent that the bureau employs an operation similar to A.C.E. to measure the quality of the 2010 Census, it will be important for the bureau to determine the impact of the problems and explore operational improvements as we recommend. We are sending copies of this report to other interested congressional committees. Please contact me on (202) 512-6806 if you have any questions. Key contributors to this report are included in appendix III. Signed by: Patricia A. Dalton: Director: Strategic Issues: [End of section] Appendix I: Scope and Methodology: To address our three objectives, we examined relevant bureau program specifications, training manuals, office manuals, memorandums, and other progress and research documents. We also interviewed bureau officials at bureau headquarters in Suitland, Md., and the bureau's National Processing Center in Jeffersonville, Ind., which was responsible for the planning and implementation of the person matching operation. In addition, to review the process and criteria involved in making an A.C.E. and census person match, we observed the match clerk training at the National Processing Center and a field follow-up interviewer training session in Dallas, Tex. To identify the results of the quality assurance procedures used in key person matching phases, we analyzed operational data and reports provided to us by the bureau, as well as extracts from the bureau's management information system, which tracked the progress of quality assurance procedures. Other independent sources of the data were not available for us to use to test the data that we extracted, although we were able to corroborate data results with subsequent interviews of key staff. Finally, to examine how, if at all, the matching operation deviated from what was planned, we selected 11 locations in 7 of the 12 bureau census regions (Atlanta, Chicago, Dallas, Denver, Los Angeles, New York, and Seattle).[Footnote 4] At each location we interviewed A.C.E. workers from November through December 2000. The locations selected for field visits were chosen primarily for their geographic dispersion (i.e., urban or rural), variation in type of enumeration area (e.g., update/leave or list enumerate), and the progress of their field follow-up work. In addition, we reviewed the match code results and field follow-up questionnaires from 48 sample clusters. These clusters were chosen because they corresponded to the local census areas we visited and contained records reviewed during every phase of the person matching operation. The results of our field visits and our cluster review are not generalizable nationally to the person matching operation. We performed our audit work from September 2000 through September 2001 in accordance with generally accepted government auditing standards. [End of section] Appendix II: Comments from the Department of Commerce: The Secretary Of Commerce: Washington, D.C. 20230: February 13, 2002: Mr. J. Christopher Mihm: Director, Strategic Issues: U.S. General Accounting Office: Washington, DC 20548: Dear Mr. Mihm: The Department of Commerce appreciates the opportunity to comment on the General Accounting Office (GAO) draft document entitled 2000 Census: Coverage Evaluation Matching Implemented As Planned, but Census Bureau Should Evaluate Lessons Learned. The Department has no comment on the text of this report. Our responses to GAO's recommendations are enclosed. Warm regards, Signed by: Donald L. Evans: Enclosure: [End of letter] Comments from the U.S. Department of Commerce, U.S. Census Bureau: U.S. General Accounting Office draft report entitled 2000 Census: Coverage Evaluation Matching Implemented As Planned, but Census Bureau Should Evaluate Lessons Learned: Comments on the Text of the Report: The U.S. Census Bureau has no comments on the text of the report. Responses to GAO Recommendations: 1. Document the criteria and the logic that Bureau staff used during computer matching to determine the cutoff scores for matched, possibly matched, and unmatched record pairs. Census Bureau Response: The Census Bureau has acknowledged that such a document may be informative. As such, a document is under preparation. 2. Examine the Bureau's system development and acquisition processes to determine why the problems with A.C.E. computer systems were not discovered prior to deployment of these systems. Census Bureau Response: The Census Bureau conducted extensive systematic testing on the A.C.E. computer systems; however, these systems are inherently complex, and a few problems may have remained undetected in spite of extensive testing. The problem identified in the report was primarily related to a software program that was developed after the Dress Rehearsal, and its testing was not as extensive as what was done for the other components of the system. We plan to review our system development processes to avoid similar problems in the 2010 census. 3. Determine the effect that the printing problems may have had on the quality of data collected for affected records, and thus the accuracy of A.C.E. estimates of the population. Census Bureau Response: When the printing problems were identified, it was thought that they would not significantly affect the coding process; therefore, we did not track the incidence of the problems and cannot report on the effect of these problems. However, the effect on the accuracy of the A.C.E. estimates is believed to be minimal, because the problems were identified early and corrective procedures were effectively implemented. 4. Determine the effect that the incomplete geocoding section of the questionnaires may have had on the quality of data collected for affected records, and thus the accuracy of A.C.E. estimates of census undercounts. Census Bureau Response: As in Item 3, we did not track the incidence of such cases, because the effects on accuracy were believed to be minimal, given that the problem was identified early and corrective procedures were effectively implemented. [End of section] Appendix III: GAO Contact and Staff Acknowledgments: GAO Contact: Robert Goldenkoff, (202) 512-2757: Acknowledgments: In addition to those named above, Ty Mitchell, Lynn Wasielewski, Steven Boyles, Angela Pun, J. Christopher Mihm, and Richard Hung contributed to this report. [End of section] Footnotes: [1] A person record should have contained the following characteristics: first name, last name, middle name, gender, race, Hispanic origin, age, date of birth, and relationship to the respondent of the A.C.E. or the census. [2] According to the bureau, a questionnaire failed quality assurance if a respondent said that the original follow-up interviewer did not contact him or her for the original interview. [3] U.S. General Accounting Office, 2000 Census: Headquarters Processing System Status and Risks, [hyperlink, http://www.gao.gov/products/GAO-01-1] (Washington, D.C.: October 17, 2000). [4] The 11 locations we visited were Chicago, Ill.; Miami and Lakeland, Fla.; New York, N.Y.; McAllen, Beaumont, and Houston, Tex.; Los Angeles, Calif.; Seattle, Wash.; and Phoenix and Window Rock, Ariz. [End of section] GAO‘s Mission: The General Accounting Office, the investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO‘s commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through the Internet. GAO‘s Web site [hyperlink, http://www.gao.gov] contains abstracts and full text files of current reports and testimony and an expanding archive of older products. The Web site features a search engine to help you locate documents using key words and phrases. You can print these documents in their entirety, including charts and other graphics. Each day, GAO issues a list of newly released reports, testimony, and correspondence. GAO posts this list, known as ’Today‘s Reports,“ on its Web site daily. The list contains links to the full-text document files. To have GAO e-mail this list to you every afternoon, go to [hyperlink, http://www.gao.gov] and select ’Subscribe to daily E-mail alert for newly released products“ under the GAO Reports heading. Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. General Accounting Office: 441 G Street NW, Room LM: Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Public Affairs: Jeff Nelligan, managing director, NelliganJ@gao.gov: (202) 512-4800: U.S. General Accounting Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.