2010 Census

Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate Gao ID: GAO-08-1012 September 30, 2008

The U.S. Census Bureau (Bureau) estimates that it will spend at least $2 billion to enumerate households that did not return census forms during the 2010 Census. Increasing the response rate would reduce the number of households that Bureau field staff must visit. To address concerns about reducing the cost of enumerating these households, GAO (1) analyzed how the Bureau develops, supports, and updates the response rate estimate, and the extent to which the Bureau uses the estimate to inform its 2010 planning efforts; (2) described the methods the Bureau considered for increasing response in 2010 and how it tested these methods; and (3) assessed how the Bureau identifies and selects for testing methods to increase response rate, including considering other surveys' methods. To meet these objectives, GAO analyzed the Bureau's documentation for estimating the response rate and selecting for testing methods to increase response, and interviewed experts from other survey organizations.

The 2010 Census response rate estimate is not fully supported, systematically reevaluated, or clearly incorporated into the life cycle cost estimate and planning efforts for nonresponse follow-up--where census workers visit households that do not return their census forms. Specifically, the Bureau could not demonstrate support for one component underpinning the estimate--a general decline due to decreasing public participation in surveys--because it did not document its decisions or data sources when developing the estimate. The two other estimate components that affect responses are the short-form-only census and the replacement questionnaire. In 2001, the Bureau estimated the 2010 Census response rate to be 69 percent. However, from 2001 through 2008, the Bureau did not systematically reevaluate the estimate or consider test results from this decade to determine if the estimate should be updated. Although the Bureau revised the estimate to 64 percent after a major redesign of its nonresponse follow-up operation in 2008, the Bureau still lacks procedures for establishing when and how to reevaluate and, if necessary, update the estimate. To estimate costs and plan for nonresponse follow-up, the Bureau relies on response rate estimates for local census office types because these estimates reflect geographic differences. Officials said that the local estimates reflect components of the national estimate. However, only one of the three components from the national estimate--the replacement questionnaire--was clearly reflected in the local census office type estimates. Through various national and field tests and experiments, the Bureau tested nine methods to increase 2010 Census response and currently plans to implement two of these methods--the replacement questionnaire and two-column bilingual form. The Bureau also plans to use a communications campaign to increase response and plans to test campaign messages in 2009. In July 2006, the Bureau decided not to include an Internet response option in the 2010 Census. However, the Bureau recently announced that it is again considering including the Internet option in 2010, although it has not developed further plans for testing it. For 2010, the Bureau established test objectives and research questions to identify methods to test for increasing response. However, Bureau officials did not document the methods that they considered but decided not to test or the rationale behind those decisions. Although officials said that they considered cost, complexity of the test, and compatibility of experiments in their decisions, they did not specify how they weighed these factors to select and prioritize the nine methods they chose to test. Officials said that they consider the experiences of other survey organizations to identify potential methods to increase response, but they, along with some experts, noted that such methods may only be indirectly applicable to the decennial census. Nonetheless, testing modifications to methods the Bureau has previously considered or tested, such as testing a variety of telephone reminder messages, may yield additional opportunities for increasing response.

Recommendations

Our recommendations from this work are listed below with a Contact for more information. Status will change from "In process" to "Open," "Closed - implemented," or "Closed - not implemented" based on our follow up work.

Director: Team: Phone:


GAO-08-1012, 2010 Census: Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate This is the accessible text file for GAO report number GAO-08-1012 entitled '2010 Census: Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate' which was released on October 30, 2008. This text file was formatted by the U.S. Government Accountability Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Report to Congressional Requesters: United States Government Accountability Office: GAO: September 2008: 2010 Census: Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate: 2010 Census: GAO-08-1012: GAO Highlights: Highlights of GAO-08-1012, a report to congressional requesters. Why GAO Did This Study: The U.S. Census Bureau (Bureau) estimates that it will spend at least $2 billion to enumerate households that did not return census forms during the 2010 Census. Increasing the response rate would reduce the number of households that Bureau field staff must visit. To address concerns about reducing the cost of enumerating these households, GAO (1) analyzed how the Bureau develops, supports, and updates the response rate estimate, and the extent to which the Bureau uses the estimate to inform its 2010 planning efforts; (2) described the methods the Bureau considered for increasing response in 2010 and how it tested these methods; and (3) assessed how the Bureau identifies and selects for testing methods to increase response rate, including considering other surveys‘ methods. To meet these objectives, GAO analyzed the Bureau‘s documentation for estimating the response rate and selecting for testing methods to increase response, and interviewed experts from other survey organizations. What GAO Found: The 2010 Census response rate estimate is not fully supported, systematically reevaluated, or clearly incorporated into the life cycle cost estimate and planning efforts for nonresponse follow-up”where census workers visit households that do not return their census forms. Specifically, the Bureau could not demonstrate support for one component underpinning the estimate”a general decline due to decreasing public participation in surveys”because it did not document its decisions or data sources when developing the estimate. The two other estimate components that affect responses are the short-form-only census and the replacement questionnaire. In 2001, the Bureau estimated the 2010 Census response rate to be 69 percent. However, from 2001 through 2008, the Bureau did not systematically reevaluate the estimate or consider test results from this decade to determine if the estimate should be updated. Although the Bureau revised the estimate to 64 percent after a major redesign of its nonresponse follow-up operation in 2008, the Bureau still lacks procedures for establishing when and how to reevaluate and, if necessary, update the estimate. To estimate costs and plan for nonresponse follow-up, the Bureau relies on response rate estimates for local census office types because these estimates reflect geographic differences. Officials said that the local estimates reflect components of the national estimate. However, only one of the three components from the national estimate”the replacement questionnaire”was clearly reflected in the local census office type estimates. Through various national and field tests and experiments, the Bureau tested nine methods to increase 2010 Census response and currently plans to implement two of these methods”the replacement questionnaire and two-column bilingual form. The Bureau also plans to use a communications campaign to increase response and plans to test campaign messages in 2009. In July 2006, the Bureau decided not to include an Internet response option in the 2010 Census. However, the Bureau recently announced that it is again considering including the Internet option in 2010, although it has not developed further plans for testing it. For 2010, the Bureau established test objectives and research questions to identify methods to test for increasing response. However, Bureau officials did not document the methods that they considered but decided not to test or the rationale behind those decisions. Although officials said that they considered cost, complexity of the test, and compatibility of experiments in their decisions, they did not specify how they weighed these factors to select and prioritize the nine methods they chose to test. Officials said that they consider the experiences of other survey organizations to identify potential methods to increase response, but they, along with some experts, noted that such methods may only be indirectly applicable to the decennial census. Nonetheless, testing modifications to methods the Bureau has previously considered or tested, such as testing a variety of telephone reminder messages, may yield additional opportunities for increasing response. What GAO Recommends: GAO recommends that the Secretary of Commerce direct the Bureau to establish procedures for developing, documenting, and reevaluating the response rate estimate and for selecting for testing methods to increase the response rate. In commenting on a draft of this report, Commerce generally agreed with GAO‘s recommendations and committed to take action for the 2020 Census. To view the full product, including the scope and methodology, click on [hyperlink, http://www.gao.gov/cgi-bin/getrpt?GAO-08-1012]. For more information, contact Mathew J. Scirč at (202) 512-6806 or sciremj@gao.gov [End of section] Contents: Letter: Results in Brief: Background: The Response Rate Estimate Lacks Support, Is Not Systematically Reevaluated, and Is Not Clearly Incorporated into Planning Efforts: The Bureau Plans to Incorporate Additional Methods to Increase Mail Response in 2010: Bureau Lacks Procedures for Selecting for Testing Methods to Increase Response Rate: Conclusions: Recommendations for Executive Action: Agency Comments and Our Evaluation: Appendix I: Objectives, Scope, and Methodology: Appendix II: Testing of Methods to Increase Response to 2010 Census: Appendix III: Comments from the Department of Commerce: Appendix IV: GAO Contacts and Staff Acknowledgments: Table: Table 1: Self-Response Methods Tested in 2010 Testing Cycle and Effect on Response Rate in Tests: Figures: Figure 1: Decennial Census Costs from 1970 through 2010 (Projected) in Constant 2010 Dollars: Figure 2: Excerpt from Advance Letter from Census 2000: Figure 3: Reminder Postcard from 2003 Census Test: Figure 4: Components of the Census Bureau's 2001 National Response Rate Estimate for the 2010 Census: Figure 5: Excerpt from Two-Column Bilingual Form from 2008 Dress Rehearsal: Figure 6: Text for Telephone Reminder Call from 2003 Census Test: Figure 7: Message on Flap to Distinguish Replacement Questionnaire from Initial Questionnaire from 2005 Census Test: Figure 8: Due Date on Envelope from 2006 Test: Figure 9: Example of a Census 2000 Communications Campaign Poster: United States Government Accountability Office: Washington, DC 20548: September 30, 2008: The Honorable Tom Davis: Ranking Member: Committee on Oversight and Government Reform: House of Representatives: The Honorable Michael R. Turner: Ranking Member: Subcommittee on Information Policy, Census and National Archives: Committee on Oversight and Government Reform: House of Representatives: The U.S. Census Bureau (Bureau) estimates that even after adjusting for inflation, the 2010 decennial census will be the most expensive census in our nation's history, costing from $13.7 billion to $14.5 billion.[Footnote 1] The Bureau estimates that more than $2 billion will be used to employ temporary field staff for nonresponse follow-up- -its largest field operation where enumerators interview households that did not return census forms. Increasing the response rate would reduce the number of households that Bureau field staff must visit during this nationwide operation.[Footnote 2] According to Bureau officials, a 1 percent increase in the response rate can save $75 million. The Bureau expects to hire over 700,000 temporary workers to conduct nonresponse follow-up with about 47 million households over the course of 10 weeks in 2010. The Bureau initially based the schedule, staffing, and funding it needed for nonresponse follow-up on an estimated national response rate of 69 percent. However, in February 2008, the Director of the Bureau initiated a replanning of the Field Data Collection Automation program--a major acquisition that includes systems; equipment, including handheld computers; and infrastructure for field staff to use in collecting data for the 2010 Census. After analyzing several options to revise the design of the 2010 Decennial Census, on April 3, 2008, the Secretary of Commerce announced that the Bureau would no longer use handheld computers in nonresponse follow-up and revised the estimated national response rate from 69 percent to 64 percent. The Bureau estimated that this option would result in a cost increase of $2.2 billion to $3 billion over the previously reported cost estimate of $11.5 billion. To address your concerns about reducing the cost of nonresponse follow- up operations, we reviewed the Bureau's estimated response rate and plans for increasing response. Specifically, we (1) analyzed how the Bureau develops, supports, and updates the response rate estimate, and the extent to which the Bureau uses the estimate to inform its 2010 planning efforts; (2) described the methods the Bureau considered for increasing response rates in 2010 and what it did to test these methods; and (3) assessed how the Bureau identifies and selects for testing methods to increase response rate, including considering other surveys' methods for increasing response. To meet these objectives, we reviewed documentation to support the components of the response rate estimate and research literature on methodologies to increase response to mail surveys and efforts to estimate survey response rate. We analyzed Bureau documents related to 2000 Census evaluations, the 2010 research and testing program, the 2010 Census life cycle cost estimate, and the communications campaign. We also interviewed Bureau officials in the Decennial Management Division and other divisions about methods to increase self-response. Further, we interviewed experts on survey methodology from Statistics Canada, the U.S. Department of Labor's Bureau of Labor Statistics, and various academic institutions, as well as former Census Bureau officials and researchers. We asked the experts a common set of questions in order to compare responses and identify recurring themes. Appendix I includes a list of experts we interviewed and additional information on our scope and methodology. We conducted our review from July 2007 through September 2008 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. Results in Brief: The 2010 Census response rate estimate is not fully supported, systematically reevaluated, or clearly incorporated into cost and planning efforts. The Bureau developed a national response rate estimate of 69 percent for the 2010 Census in 2001, but could not demonstrate support for one of the three components underpinning it-- the general decline due to decreasing public participation in surveys. (The other two components are the use of a short-form-only census and the introduction of a replacement mailing, both of which the Bureau expects to increase response.) The Bureau does not have procedures for developing the estimate, nor could the Bureau provide detailed documentation on how the response rate estimate was developed. Clear, detailed documentation would allow the Bureau and others to better assess the reliability of the estimate, such as determining whether the estimate's assumptions are supported and realistic. Although the Bureau reevaluated and developed a revised estimate of 64 percent after the Bureau announced a major redesign of the nonresponse follow-up operation in April 2008, the Bureau still does not have procedures for systematically reevaluating the estimate throughout the decade based on test results or other triggering events. For example, the Bureau did not reevaluate the estimate prior to April 2008 or use test results conducted during the decade to determine if the estimate should be updated. Having procedures for when and how to reevaluate the estimate would help the Bureau to ensure that the estimate is current and reliable. Finally, for estimating cost and planning for nonresponse follow-up, the Bureau relies on response rate estimates from four primary types of local census offices (LCO), which help to account for expected differences in response patterns across geographic settings. According to the Bureau, the LCO-type response rate estimates are, in part, based on components of the national estimate. However, the Bureau could not explain why it applied the replacement mailing component equally across all LCO-type estimates or how it incorporated the short-form-only component and the general decline component into any of the LCO-type estimates. Establishing a quantitative basis for applying the components to the LCO-type estimates would better inform the Bureau's cost estimate and planning efforts for operations that are based on response, such as hiring enumerators for nonresponse follow-up. Overall, the Bureau lacks procedures for developing, documenting, and reevaluating the response rate estimate. Nonetheless, Bureau officials stated that they tried to conservatively estimate the response rate to ensure that they are adequately prepared during nonresponse follow-up and to avoid repeating what happened in 1990, when the Bureau overestimated the response rate, requiring a supplemental appropriation of $110 million and forcing it to extend nonresponse follow-up by up to 8 weeks for some areas. Conservatively estimating the response rate may be a reasonable approach. However, having clear documentation to support the estimate and establishing procedures for when or how to reevaluate the estimate to ensure that it reflects current information from testing would enable the Bureau and others to assess whether the estimate is reliable. An unreliable response rate estimate can produce an inaccurate cost estimate and can increase risk and uncertainty to operational plans that are based on the response rate estimate. Through various national and field tests and experiments, the Bureau tested nine methods to increase self-response to the 2010 Census and plans to implement two of these methods--the replacement questionnaire and two-column bilingual form. An additional method the Bureau plans to include in the 2010 Census to increase response, the Integrated Communications Campaign, has not been tested, although a similar communications campaign was part of the 2000 Census. The Bureau does not plan to include three methods it tested to increase mail response- -a telephone reminder call, messaging to distinguish initial and replacement questionnaires, and a due date on the initial mailing package--because they did not significantly increase response or required further testing. Bureau officials said that they will be conducting additional testing on the use of a due date as part of the 2010 Census Program for Evaluations and Experiments. The National Academy of Sciences also recommended such analysis. At the time of this report, the Bureau had not yet completed its plans for 2010 evaluations and experiments. The Bureau also does not plan to include four methods it tested to increase self-response through electronic systems--interactive voice response; a letter encouraging responding via the Internet, instead of sending a replacement questionnaire; computer-assisted telephone interviewing; and the Internet--because they did not increase response or were too costly. In July 2006, the Bureau decided not to include the Internet option in the design of the 2010 Census largely because it had underestimated contract costs for developing the system that included the Internet option, but also because test results indicated that the Internet did not always increase the response rate and that security over respondent data was a concern. However, the Bureau recently announced that it is again considering including the Internet option in 2010, even though it has not developed further plans for testing it. Finally, the Bureau has not yet developed detailed testing or evaluation plans for the communications campaign, though it plans to test campaign messages in 2009. To identify methods for increasing response in 2010 that it planned to test, the Bureau developed a research strategy that included establishing test objectives and research questions. However, the Bureau did not document the methods it considered but decided not to test during the 2010 testing cycle or the rationale behind those decisions. Further, for methods that the Bureau decided to test, officials could not provide support for how they selected or prioritized these methods. Although officials said that they considered cost, complexity of the test, and compatibility of experiments in their decisions, they did not specify how they defined or weighed these factors to select and prioritize the nine methods they chose to test. Documenting decisions about methods that were not selected for testing would help the Bureau more effectively build capacity and institutional knowledge about changes in methods to consider in the future. Bureau officials also said that they consider the experiences of other survey organizations to identify potential methods to increase response rate. However, both Bureau officials and some survey experts acknowledged that methods used to increase response in other surveys may only be indirectly applicable to the decennial census. Nonetheless, testing modifications to methods the Bureau has previously considered or tested, such as testing a variety of telephone reminder messages, may yield additional opportunities for increasing response. To enhance credibility of the response rate for determining cost and planning for future decennial censuses and to inform assumptions underlying the 2020 response rate estimate, we recommend that the Secretary of Commerce direct the Bureau to establish procedures for documenting the process for developing the response rate estimate, including analyzing 2010 data to assess the reasonableness of assumptions used in applying the national estimate's components to the LCO-type estimates, as well as establishing when and how to reevaluate the response rate estimate. Further, to improve the planning and transparency of the Bureau's research and testing for future censuses, we recommend that the Bureau develop procedures for selecting methods to test for increasing response. On September 22, 2008, the Secretary of Commerce provided written comments on a draft of this report (see app. III). Commerce generally agreed with our conclusions and recommendations and provided suggestions where additional context or clarification was needed. Where appropriate, we incorporated these changes. Background: In 1970, the Bureau moved away from conducting the census door-to-door and began mailing census questionnaires to households to be filled out and returned by mail. Since the 1970 Census, the Bureau has used mail as the primary method for collecting census data. For Census 2000, the mailout/mailback method was used for more than 80 percent of the population.[Footnote 3] Households that fail to mail back the census questionnaires are included in nonresponse follow-up workload, where enumerators follow up with door-to-door visits and telephone calls or solicit census data from knowledgeable people, such as neighbors. The census response rate declined dramatically from 1970 through 1990. The 1970 response rate was 78 percent. The rate decreased to 75 percent for the 1980 Census and then decreased again to 65 percent for the 1990 Census.[Footnote 4] Although the Bureau estimated that the 2000 Census response rate would continue to decline to 61 percent, actual response exceeded the estimate, reaching 65 percent for the mailout/mailback universe prior to nonresponse follow-up.[Footnote 5] The response rate is defined as the percentage of census forms completed and returned for all housing units that were on the Bureau's address file eligible to receive a census questionnaire delivered by mail or by a census enumerator. The denominator used in calculating the response rate includes vacant housing units and other addresses that were determined to be undeliverable or deleted through other census operations.[Footnote 6] Since 1970, per household census costs have increased (in constant 2010 dollars) from about $14 in 1970 to an estimated $88 in 2010--a figure that does not include the recent, major redesign of the Field Data Collection Automation program. According to the Bureau, factors contributing to the increased costs include an effort to accommodate more complex households, busier lifestyles, more languages and greater cultural diversity, and increased privacy concerns. In addition, the number of housing units--and hence, the Bureau's workload--has continued to increase. The Bureau estimated that the number of housing units for the 2010 Census will increase by almost 14 percent over 2000 Census levels (from 117.5 million to 133.8 million housing units). As a result of multiple factors, the total inflation-adjusted life cycle cost for the decennial census has increased more than tenfold since 1970, as shown in figure 1. Figure 1: Decennial Census Costs from 1970 through 2010 (Projected) in Constant 2010 Dollars: This figure is a bar graph showing decennial census costs from 1970 through 2010 (projected) in constant 2010 dollars. The X axis represents the census year, and the Y axis represents the cost in billion of dollars. Census year: 1970; Cost in billions of dollars: 1.0. Census year: 1980; Cost in billions of dollars: 2.6. Census year: 1990; Cost in billions of dollars: 4.1. Census year: 2000; Cost in billions of dollars: 8.2. Census year: 2010 (projected); Cost in billions of dollars: 11.8. [See PDF for image] Source: GAO analysis of U.S. Census Bureau figures. Note: This figure does not reflect the Bureau's estimated increase in the 2010 Census life cycle cost estimate ranging from $13.7 billion to $14.5 billion, which the Bureau announced on April 3, 2008, with the replan of the Field Data Collection Automation program. [End of figure] The Bureau conducts its testing, evaluation, and experimentation program for the decennial census primarily through the Decennial Statistical Studies Division. The division develops and coordinates the application of statistical techniques in the design and conduct of decennial census programs. The Bureau conducts various tests throughout the decade to determine the effect and feasibility of form or operational changes. For example, in the 1990s, the Bureau tested whether adding a mandatory response message and implementing a multiple mailing strategy, such as sending an advance letter and reminder postcard, as shown in figures 2 and 3, would increase the response rate. Based on positive test results, these methods were added to the Census 2000 design. Figure 2: Excerpt from Advance Letter from Census 2000: This figure is an excerpt from advance letter from census 2000. United States Department Of Commerce: Bureau of the Census: Washington, DC 20233-2000: Office Of The Director: March 6, 2000: About one week from now, you will receive a U.S. Census 2000 form in the mail. When you receive your form, please fill it out and mail it in promptly. Your response is very important. The United States Constitution requires a census of the United States every 10 years. Everyone living in the United States on April 1, 2000, must be counted. By completing your census form, you will make sure that you and members of your household are included in the official census count. Official census counts are used to distribute government funds to communities and states for highways, schools, health facilities, and many other programs you and your neighbors need. Without a complete, accurate census, your community may not receive its fair share. You can help in another way too. We are now hiring temporary workers throughout the United States to help complete the census. Call the Local Census Office near you for more information. The phone number is available from directory assistance or the Internet at [hyperlink, http://www.census.gov/jobs2000]. With your help, the census can count everyone. Please do your part. Thank you. Sincerely, Signed by: Kenneth Prewitt Director: Bureau of the Census: Enclosure: United States Census 2000: [See PDF for image] Source: U.S. Census Bureau. [End of figure] Figure 3: Reminder Postcard from 2003 Census Test: United States Department Of Commerce: Economics and Statistics Administration: U.S. Census Bureau: Washington, DC 20233-0001: Office Of The Director: February 7, 2008: A few days ago, you should have received a request for your participation in the 2003 National Census Test. The test is being conducted by the U.S. Census Bureau to help develop new methods for the next census in 2010. If you have already mailed back your completed census form, please accept our sincere thanks. If you have not responded, please mail it back as soon as possible. Taking an accurate census is important for all communities throughout the United States to get their fair share of federal funding. Your participation in this test is important to the success of the next census. Sincerely, Signed by: Charles Louis Kincannon: Director, U.S. Census Bureau: [See PDF for image] Source: U.S. Census Bureau. [End of figure] The Decennial Statistical Studies Division also conducts evaluations of census operations typically using data collected as part of the decennial process to determine whether individual steps in the census operated as expected. It is also responsible for conducting experiments during the census that may be instructive for future censuses. These experiments usually involve using alternative processes for a subset of the population. The Bureau is working with internal and external stakeholders on defining and developing its program for experiments and evaluations for the 2010 Census, known as Census Program for Evaluations and Experiments, with design work and implementation starting in 2008 and continuing through 2011. The final set of activities would include analysis, documentation, and presentation of the research, and these activities would start in 2009 and be completed by the end of fiscal year 2013. The Response Rate Estimate Lacks Support, Is Not Systematically Reevaluated, and Is Not Clearly Incorporated into Planning Efforts: 2010 Census Response Rate Estimate Is Not Fully Supported: The Bureau developed the response rate estimate for the 2010 Census in 2001 but could not demonstrate support for one of the three components underpinning the estimate. To establish the 2010 estimate, Bureau officials told us that they used Census 2000 mail response data as the baseline and then incorporated three other components to arrive at a response rate estimate of 69 percent: an increase in response related to eliminating the long form and moving to a short-form-only census, an increase in response related to sending a replacement questionnaire to nonresponding households, and a general decline in mail response due to decreasing public participation in surveys. The components of the estimate are outlined in figure 4, and a detailed explanation of the baseline and each component follows. Figure 4: Components of the Census Bureau's 2001 National Response Rate Estimate for the 2010 Census: This figure is a chart showing the components of the census bureau's 2001 national response rate estimate for the 2010 census. [See PDF for image] Source: GAO analysis of U.S. Census Bureau information. [End of figure] Baseline Rate: Included in the baseline national response rate of 65 percent[Footnote 7] are the effects of methods, such as the multiple mailing strategy and the communications campaign (called the Partnership and Marketing Program) that were implemented in Census 2000 and are planned to be implemented in the 2010 Census as well. Bureau officials stated that the communications campaign, which included paid advertising, had an impact on the response rate achieved in 2000, but were unable to quantify that effect and did not project the campaign's effect for 2010. Short-Form-Only Census: The Bureau estimated an increase of 1 percentage point from eliminating the long form and moving to a short-form-only census. The Bureau conducted an analysis comparing Census 2000 response rates for the short form and the long form. Even though the difference in response rates for the two form types was rather large--the short form had a response rate that was more than 10 percentage points higher than the long form--the overall effect on the response rate estimate was small because the long form was sent to only approximately 17 percent of housing units in 2000. Replacement Mailing: In 2001, the Bureau expected that sending a replacement questionnaire to households that had not responded by a certain date would increase the response rate by 7 percentage points.[Footnote 8] The magnitude of this component was based on test results from the 1990s, which showed an increase in the response rate related to the replacement mailing of at least 7 percentage points, and possibly 10 percentage points or more. Bureau officials stated that they estimated this effect conservatively because they assumed that a higher response rate from a replacement questionnaire is more likely to occur in tests than in the decennial census. Also, testing showed that not all the respondents who return replacement questionnaires would be removed from the nonresponse follow-up workload prior to being enumerated because of the timing of the enumerator's visit. General Decline: The Bureau estimated a decrease of 4 percentage points from the baseline level due to what it believes is a decline over time in public survey participation. However, this assumption is not supported by quantitative analysis or research studies but rather based on the opinion of subject matter experts and senior officials. Further, the Bureau could not demonstrate who was consulted, when they were consulted, or how they decided on the amount of the general decline. Best practices for cost estimation from our Cost Assessment Guide[Footnote 9] call for documenting assumptions and data sources used to develop cost estimates, such as the general decline component of the response rate estimate. In addition, several experts we interviewed agreed that the Bureau should have used quantitative analysis of mail response data from other surveys to support the general decline component when it developed the estimate in 2001. To help support the general decline component, we suggested that the Bureau use changes in participation in the American Community Survey, which uses a similar mailing strategy to the decennial census and for which response is also required by law, for comparison purposes. In May 2008, the Bureau completed an analysis of American Community Survey response rate data from 2000 to 2007, which demonstrated a decline of 6.6 percentage points in cooperation rates[Footnote 10] to the initial questionnaire. Although the decline in response to the American Community Survey may not be directly comparable to response behavior for the decennial census, it provides some support for the Bureau's assumption that mail survey participation may be declining. The Bureau does not have procedures for developing the response rate estimate. Specifically, the Bureau has no established policies for documenting deliberations related to or data sources used in developing the estimate, including no 2010 Census decision memorandum to document the original estimate prepared in 2001. Several experts also suggested that the Bureau should develop a model for developing the estimate and incorporate response rate data and demographic data from other surveys.[Footnote 11] Bureau officials, however, noted that they have not determined how to apply data from other surveys to estimate the census response rate because the census differs from other surveys, nor do they plan to develop a model of the response rate estimate. Instead, Bureau officials stated that they tried to conservatively estimate the response rate to ensure that they are adequately prepared during nonresponse follow-up and to avoid repeating what happened in 1990, when the Bureau overestimated the response rate, requiring a supplemental appropriation of $110 million and forcing it to extend nonresponse follow-up by up to 8 weeks for some areas. In contrast, in 2000 the response rate exceeded the Bureau's estimate. Conservatively estimating the response rate may be a reasonable approach. However, having clear, detailed documentation about decisions related to or data sources used in developing the estimate would enable the Bureau and others to better assess whether the estimate is reliable, such as whether its assumptions are supported and realistic. An unreliable response rate estimate can produce an inaccurate cost estimate and can increase risk and uncertainty to operational plans that are based on the response rate estimate. The Bureau Does Not Systematically Reevaluate the Response Rate Estimate: Although the Bureau updated the response rate estimate as a result of a major redesign of census operations undertaken in April 2008, the Bureau still lacks procedures for establishing when and how to reevaluate the response rate estimate. From 2001 through April 2008, the Bureau did not reevaluate the estimate to determine whether it should be updated, even though, after establishing the initial estimate in 2001, the Bureau completed several evaluations from Census 2000 and conducted several census tests that could have informed the estimate. During the first few years of the decade, the Bureau completed 12 evaluations that discussed aspects of the Census 2000 response rate. In addition, the Bureau conducted five tests from 2003 through 2007 that were designed, in part, to test methods to increase response. For example, in 2005 and 2007, the Bureau tested a two-column bilingual form, which includes identical questions and response options in English and Spanish. The 2005 test demonstrated a significant positive effect on mail response rates of 2.2 percentage points overall, and a 3.2 percentage point increase in areas with high concentrations of Hispanic and non-White populations, and the 2007 test revealed a similar impact. However, Bureau officials stated that they were concerned that the two-column bilingual form has not been tested in a census environment. In addition, they noted that the bilingual form will be sent to only a portion of the country--approximately 10 percent of census tracts--with high concentrations of Spanish-speaking populations, based on American Community Survey data. Best practices state that assumptions, such as the response rate estimate, should be revised as new information becomes available. Our Cost Assessment Guide recommends that preliminary information and assumptions be monitored to determine relevance and accuracy and be updated as changes occur to reflect the best information available. Further, according to several experts, the Bureau could have updated the estimate based on response rate data and demographic data from other surveys, and another expert suggested that the Bureau use triggering events, such as after tests, for reviewing the response rate estimate. Based on this expert's experience, the Bureau could establish a change control board chaired by senior Bureau officials to determine whether the response rate estimate should be revised. However, Bureau officials explained that they do not update the estimate based on results from tests because the tests cannot replicate the decennial census environment. Overall, the Bureau has not specified when or how it is to reevaluate its response rate estimate. Establishing these procedures would help the Bureau ensure that the estimate is current and reliable in order to better inform planning efforts. The Bureau revised the estimate down to 64 percent after announcing that nonresponse follow-up would be changed from an automated to a paper-based operation. Prior to the redesign, the Bureau had planned to reduce the nonresponse follow-up workload by using handheld computers to remove households that returned their forms late--including many replacement mailing returns--from the enumerator assignments on a daily basis. According to Bureau officials, they revised the response rate estimate based on the timing of mail returns in 2000 and replacement questionnaire returns in the 2003 test. However, the revised estimate does not fully reflect recently designed procedures to remove late mail returns in 2010. Specifically, for 2010 the Bureau now plans to mail the replacement questionnaire earlier, send a blanket replacement questionnaire to some areas and a targeted replacement questionnaire to others,[Footnote 12] and conduct three clerical removals of late mail returns immediately prior to and during the nonresponse follow-up operation. These operational plans are still being finalized pending further analysis. Bureau officials said that although they hope that these revised operations will increase response, they did not update the response rate estimate to reflect these current operational plans because they have not tested this approach under decennial census conditions and therefore have no basis for estimating the potential effect of these operational changes on response. Local Response Rate Estimates Used in Planning Efforts Lack Support: To estimate costs and to plan for activities such as nonresponse follow- up, questionnaire printing, and postage needs, the Bureau uses a life cycle cost model that relies on response rate estimates for four primary LCO types[Footnote 13] and not the national estimate. Using LCO- type estimates helps to account for expected differences in response patterns across geographic settings. For example, response rates for inner-city areas are estimated to be lower than in suburban areas. These estimates range from 55 percent to 72 percent, but lack support for how they were developed. In determining the 2010 response rate estimates for the four LCO types, the Bureau said that it equally applied two components from the national estimate--the replacement mailing component and the short-form-only component--to the LCO types in the cost model. However, the Bureau did not conduct quantitative analysis to determine whether the replacement mailing component in the national estimate should be applied equally for each LCO type. By applying an equal percentage point increase across all LCO types for the replacement mailing, the Bureau has, in effect, assumed that the LCO type with the lowest baseline--48 percent--would experience a higher relative increase in response--15 percent--due to the replacement mailing than the LCO type with the highest baseline--65 percent--which would be expected to experience an 11 percent increase in response. The Bureau has not demonstrated support for assuming that the LCO types will experience these different relative increases in response. Further, the Bureau could not demonstrate whether or how the short-form-only component and the general decline component of the national estimate were reflected in the life cycle cost model. Best practices for cost estimation from our Cost Assessment Guide call for assumptions to be thoroughly documented, data to be normalized, and each cost element to be supported by auditable and traceable data sources. Following these best practices could allow the Bureau to enhance its use of the response rate estimate in planning for the decennial census and better inform stakeholders about the reliability of the estimate. It is unclear why officials could only explain how one component of the national response rate estimate was applied to LCO- type estimates in the life cycle cost model. However, according to the Bureau, it is currently documenting how these estimates are used to calculate the costs for operations and how the estimates have changed over time. Having support for the LCO-type estimates would better inform the Bureau's planning efforts for operations that directly rely on response, such as determining workload and hiring enumerators for nonresponse follow-up. The Bureau Plans to Incorporate Additional Methods to Increase Mail Response in 2010: Two of Nine Methods Tested to Increase Response Rate Are Planned to Be Included in the 2010 Census: We determined from reviewing Bureau testing documents that through various national and field tests and experiments, the Bureau tested nine methods to increase self-response to the 2010 Census, as shown in table 1. The Bureau currently plans to implement two of these methods- -the replacement questionnaire and two-column bilingual form (see fig. 5).[Footnote 14] Three other methods for increasing mail response and four methods for increasing response through an electronic data collection system were tested but are not planned for implementation in 2010. An additional method the Bureau plans to include in the 2010 Census to increase response, the Integrated Communications Campaign, has not been tested, although the Bureau conducted a similar campaign in 2000 and plans to test campaign messages with audiences in early 2009. Additional details regarding the testing of these methods can be found in appendix II. Table 1: Self-Response Methods Tested in 2010 Testing Cycle and Effect on Response Rate in Tests: Methods planned for implementation in 2010: 1. Replacement questionnaire; Year(s) tested: 2003, 2006; Effect on response rate in tests (percentages): 8.8 to 10.3. Methods planned for implementation in 2010: 2. Two-column bilingual form; Year(s) tested: 2005, 2007; Effect on response rate in tests (percentages): 1.4 to 2.2. Methods not planned for implementation in 2010: Methods not planned for implementation in 2010; Year(s) tested: [Empty]; Effect on response rate in tests (percentages): [Empty]. Methods not planned for implementation in 2010: 3. Telephone reminder call; Year(s) tested: 2003; Effect on response rate in tests (percentages): 0. Methods not planned for implementation in 2010: 4. Due date on initial questionnaire mailing package; Year(s) tested: 2003, 2006; Effect on response rate in tests (percentages): 0 to 2.0[A]. Methods not planned for implementation in 2010: 5. Messaging to distinguish the replacement from the initial questionnaire; Year(s) tested: 2005; Effect on response rate in tests (percentages): -1.2. Electronic response options: 6. Interactive voice response; Year(s) tested: 2000, 2003; Effect on response rate in tests (percentages): -4.9 to 0[B]. Electronic response options: 7. Letter encouraging internet response instead of sending replacement questionnaire; Year(s) tested: 2005; Effect on response rate in tests (percentages): -3.7. Electronic response options: 8. Computer-assisted telephone interviewing; Year(s) tested: 2000; Effect on response rate in tests (percentages): 2.1. Electronic response options: 9. Internet; Year(s) tested: 2000, 2003; Effect on response rate in tests (percentages): 0 to 2.5. Source: GAO analysis of U.S. Census Bureau information. Notes: We included only those years in which the tests focused primarily on determining the method's effect on response, rather than on data quality, operational feasibility, or other aspects of survey design. A zero effect indicates that the method tested had no statistically significant effect on response rate compared to the control group. Test response rates may not be fully comparable to response rates or return rates calculated for decennial censuses, in part because ineligible housing units, such as those that are vacant or undeliverable as addressed, are treated differently in the denominators. [A] The 2 percentage point increase in response rate was obtained when the due date was tested in conjunction with a compressed mailing schedule, which shortened the time period between the mailing of the questionnaire and Census Day. [B] The Bureau notes that the 2000 test result showing no statistically significant effect on response is difficult to interpret because a portion of the sample in the interactive voice response panel either received the census form late or did not receive it at all. [End of table] Figure 5: Excerpt from Two-Column Bilingual Form from 2008 Dress Rehearsal: This figure is a copy of an excerpt from two-column bilingual form from 2008 rehearsal. [See PDF for image] Source: U.S. Census Bureau. [End of figure] For the 2010 Census, the Bureau does not plan to include three methods it tested to increase mail response, as shown in figures 6, 7, and 8. Two of these methods--a telephone reminder call and messaging to distinguish the initial and replacement questionnaires--were not found to significantly improve response and therefore will not be implemented as part of the 2010 Census. The third method--including a due date on the initial mailing package--was found to generate faster responses when tested in 2003, and increase overall response by 2 percentage points when tested in conjunction with a compressed mailing schedule in 2006.[Footnote 15] However, the Bureau also believes that the use of a due date alone could cause a lower response rate because some people may not send back the census form after the due date. According to Bureau officials, they are not including this method in the decennial census design because they would like to further test it--both with and without a compressed mailing schedule--in a census environment. It will be important for the Bureau to optimize this testing opportunity by designing the test to determine the extent to which the faster and higher response is due to a compressed schedule versus a due date, as well as exploring other test treatments the Bureau has recommended in the past, such as including the due date on multiple mailing pieces. Bureau officials said that additional testing on the use of a due date will be conducted as part of the 2010 Census Program for Evaluations and Experiments, which the National Academy of Sciences recommended in its December 2007 interim report.[Footnote 16] The Bureau did not provide further details on plans for the Census Program for Evaluations and Experiments, but officials have said that they will consider costs and staffing needs in deciding what to evaluate. The academy's final report is due in September 2009.[Footnote 17] Figure 6: Text for Telephone Reminder Call from 2003 Census Test: Hello. A few days ago the U.S. Census Bureau sent a request to your address, asking for your participation in the 2003 National Census Test. If you‘ve already responded to this request, please accept our sincere thanks. If you haven‘t, please take a few minutes to complete your census form and return it by mail. Your response will help us develop the best procedures for counting the U.S. population. Thank you. [See PDF for image] Source: U.S. Census Bureau. [End of figure] Figure 7: Message on Flap to Distinguish Replacement Questionnaire from Initial Questionnaire from 2005 Census Test: This figure is a copy of a message on flap to distinguish replacement questionnaire from initial questionnaire from 2005 census test. [See PDF for image] Source: U.S. Census Bureau. [End of figure] Figure 8: Due Date on Envelope from 2006 Test: [See PDF for image] Source: U.S. Census Bureau. [End of figure] In addition to testing methods to increase mail response, the Bureau also tested four methods to increase self-response through electronic data collection systems: (1) providing for response via an interactive voice response system;( 2) sending a letter to encourage either responding via Internet or returning the initial questionnaire, instead of sending a replacement questionnaire; (3) using computer-assisted telephone interviewing; and (4) providing capability for responding via the Internet. The Bureau found that the first two methods did not increase response in tests and does not plan to include them in the 2010 Census. Computer-assisted telephone interviewing allows respondents to use the telephone to connect with operators who record interview responses electronically. The Bureau found that this method increased the overall response rate in a Census 2000 experiment. However, in a March 2004 report, the Bureau also stated that the method would likely be too costly in terms of hardware, software, and staffing resources, compared to the increase in response it might generate. Computer-assisted telephone interviewing was not tested after the 2000 Census. Although the Internet option was found to increase overall response during the Census 2000 experiment, it did not increase overall response when tested again in 2003, when a replacement mailing was also tested. According to a July 2006 Bureau decision memo titled, "Rationale for the Decision to Eliminate the Internet Option from the DRIS Contract," the Bureau decided not to include the Internet option in the design of the 2010 Census largely because it had underestimated the costs of the contract that included developing the Internet response option. In responding to a draft of this report, the Bureau stated that they made the decision because test results showed that offering the Internet option did not increase overall response and would not offer any cost savings. According to the 2006 memo, the Bureau also determined that the operational risks and costs to develop this option outweighed the potential benefits. In terms of benefits, the Bureau found improvements in data completeness and data timeliness from Internet responses in tests conducted in 2000/2001 for the American Community Survey and in 2003 and 2005 for the National Census Tests. The Bureau noted that these benefits could translate into reduced costs since less follow-up is required to improve data accuracy, and earlier receipt of responses could result in fewer replacement questionnaires that need to be mailed and fewer households that need to be enumerated during nonresponse follow-up. However, with only 6 to 7 percent of the test population using the Internet option, the Bureau concluded that no cost savings could be realized from reducing the number or size of data capture centers (facilities that process returned questionnaires) planned for 2010. Finally, the Bureau stated that the inability to fully test the Internet option and growing concerns about Internet security made it unfeasible for the Bureau to implement the Internet as a response option for 2010. Despite its July 2006 decision, the Bureau recently announced that it is considering including the Internet option as part of the 2010 Census design; however, the Bureau has not developed plans for further testing this option. Testing and Evaluation Plans for the 2010 Communications Campaign Are Not Yet Fully Developed: Through its contractor and subcontractors, the Bureau has taken a number of steps to inform the planning of the communications campaign but has not yet fully developed the campaign's testing and evaluation plans.[Footnote 18] From late 2006 through early 2008, focus groups were conducted with various ethnic groups and hard-to-count populations, such as unattached singles, to identify potential barriers and motivators to participation, to better understand methods of communication that work for different groups, and to develop the campaign's overall creative expression. The Bureau also developed an audience segmentation model using Census 2000 response data, updated through 2006 using American Community Survey data, to provide a more detailed understanding of the characteristics, such as home ownership and unemployment level, of those more or less likely to respond, as well as where they live, in order to better target communications to encourage census participation. In addition, in 2008, a national phone and in-person survey is being conducted to further explore barriers and motivators to response, particularly in hard-to-count populations, which would inform campaign message development. According to agency officials, beginning in early 2009, campaign messages are to be tested by showing storyboards to audiences that will use electronic devices to vote on the messages. These messages will be tested in 14 languages and for other populations, such as Spanish speakers in Puerto Rico. According to the draft plan, testing of events, partnership toolkits, promotional items, and public relations ideas will also be conducted. However, the Bureau has not yet developed detailed plans for this testing because, according to one official, the Bureau intends to further develop testing plans as future fiscal year funding amounts become available. An example of a Census 2000 communications campaign poster is shown in figure 9. Figure 9: Example of a Census 2000 Communications Campaign Poster: This figure is a copy of a census 20000 communications campaign poster. [See PDF for image] Source: U.S. Census Bureau. [End of figure] In addition, although the Bureau expects to award a contract by the end of fiscal year 2008 for an independent evaluation measuring the campaign's performance against its goals of increasing mail response, improving accuracy, and reducing the differential undercount, and improving cooperation with enumerators, it has not yet done so. In the past, the Bureau has said that although evaluations have shown that the Census 2000 communications campaign increased awareness for that census, it was difficult to link increased awareness to changes in respondent behavior. Bureau officials said that they have attempted to analyze Census 2000 data to identify factors that influence behavior, but their research results were inconclusive. Going forward, it will be important for the Bureau to determine its plans for evaluating the 2010 communications campaign so that it does not miss opportunities to collect data in the census environment to inform future campaigns. Bureau Lacks Procedures for Selecting for Testing Methods to Increase Response Rate: For 2010, the Bureau developed a strategy to identify various methods to test for increasing response. Specifically, the Bureau established test objectives and research questions, such as identifying the optimal mix of response options for the public to respond to the 2010 Census and determining security and confidentiality issues surrounding technology. The Bureau also developed the 2010 Census Integrated Program Plan and 2010 Census Operations and Systems Plan to better document its planning. However, Bureau officials did not document the methods that they considered but decided not to test in the 2010 testing cycle or the rationale behind those decisions. Further, for methods that the Bureau decided to test, officials could not provide support for how they selected or prioritized these methods. Although officials said that they considered cost, complexity of the test, and compatibility of experiments in their decisions, they did not specify how they defined or weighed these factors to select and prioritize the nine methods they chose to test. To ensure thorough and comprehensive planning of the decennial census, our past work on lessons learned from Census 2000 highlighted the importance of documentation to support research, testing, and evaluation, and a comprehensive and prioritized plan of goals, objectives, and projects.[Footnote 19] While the Bureau has developed the 2010 Census Integrated Program Plan and 2010 Census Operations and Systems Plan to better document its planning, these and other planning documents we reviewed did not provide support for how the Bureau selected and prioritized methods to test. It is unclear why the Bureau lacks procedures for documenting decisions concerning how it selected for testing methods to increase response. Documenting decisions about methods that were not selected for testing would help the Bureau more effectively build capacity and institutional knowledge about changes in methods to consider in the future. According to Bureau officials, they consider the experience of other survey organizations when identifying methods for increasing response rate. For example, they said that they attend research conferences to learn about experiences of other organizations that conduct national surveys to identify potential methods to increase response rate. Both Bureau officials and some survey experts noted that methods used to increase response in other surveys may only be indirectly applicable to the decennial census. For example, for the Economic Census, officials said that the Bureau sends nonresponding businesses multiple reminder letters; the final letter to large businesses informs them that the Department of Justice will be pursuing nonrespondents.[Footnote 20] Bureau officials said that these methods are less feasible for the decennial census to implement because of the shorter time frame for obtaining responses and concerns about being respondent-friendly to households. In addition, one survey expert noted that methods applicable to small-scale surveys, such as personalizing a cover letter, may be less feasible to implement for the decennial census. Further, survey experts we interviewed generally said that they were unaware of additional methods from censuses undertaken by other countries or private sector surveys that the Bureau could consider to increase mail response. Some experts noted differences between how the United States and other countries conduct their censuses, which may make it difficult to directly transfer other countries' practices to the U.S. census. For example, some European censuses use a population register to collect names, unlike the United States, which builds its survey frame from phone and address lists. In addition, past research has provided evidence that government-sponsored self-administered surveys, such as the decennial census, tend to achieve higher response rates than nongovernmental surveys. Nonetheless, testing modifications to methods that the Bureau has previously considered or tested in earlier studies may yield additional opportunities for increasing response. For example, the Bureau could test various telephone reminder messages stating that response is mandatory by law or providing instructions for obtaining a new census form. Although Bureau officials said that they have previously used the American Community Survey to inform the census design, analyzing respondent behavior to the American Community Survey, because of its similar mailing strategy to the decennial census, could help the Bureau regularly refine its survey methodology for increasing census response. Although the survey forms are different, many concepts, such as targeting the second mailing, modifying the appearance of the mailing, and varying telephone and Internet messages to prompt nonrespondents, could be tested with reasonable inference to the census. Conclusions: Nonresponse follow-up is the largest field operation, and the Bureau estimates that it will cost more the $2 billion. To control the cost of nonresponse follow-up, it will be important for the Bureau to devise a strategy for getting people to return their census forms. A reliable response rate estimate is a critical element necessary for determining the resources needed to carry out nonresponse follow-up. The Bureau did not have support for one of the components of the 2010 response rate estimate--a general decline in responsiveness--and the Bureau does not have procedures for reviewing the estimate after testing to determine whether it should be revised. Establishing procedures for developing the response rate estimate, including documenting data sources and decisions, would enable the Bureau and others to better assess the estimate's reliability. Also, establishing procedures for when, such as after tests or other triggering events, and how to reevaluate the estimate would help the Bureau ensure that it is providing the most current response rate estimate for planning nonresponse follow-up and other activities, such as questionnaire printing. The Bureau's strategy of estimating the response rate conservatively may be prudent given past difficulties with conducting nonresponse follow-up after overestimating the response rate in 1990. Nonetheless, establishing and following procedures for developing, documenting, and reevaluating the estimate are important steps for understanding differences between the estimate and the actual response rate for 2010 and for evaluating the components and underlying assumptions when developing the estimate for the next census. Successful enumeration depends on early research, testing, and evaluation of census methods to increase response. Establishing procedures for selecting and prioritizing the testing of methods--such as the Internet or reminder telephone call--including documenting methods considered but not tested, would help the Bureau demonstrate that it has chosen an optimal research strategy for the decennial census, more effectively build capacity and institutional knowledge about changes in methods to consider in the future, and enable it to more efficiently begin testing for the next census. Recommendations for Executive Action: To enhance credibility of the response rate for determining cost and planning for future census activities, to inform assumptions underlying the 2020 response rate estimate, and to improve the planning and transparency of the Bureau's research and testing, we are recommending that the Secretary of Commerce direct the Bureau to take the following three actions: * Establish procedures for developing the 2020 response rate estimate, including documenting the data sources supporting the estimate's components and decisions that are made in establishing the components and analyzing 2010 data to assess the reasonableness of assumptions used in applying the national estimate's components to the LCO-type estimates. * Establish procedures for reevaluating and updating the 2020 estimate, including identifying events or changes in related operations that should trigger a review and documenting the results of such reviews. * Establish procedures for selecting methods for increasing response rate that will be the subject of research and testing, including requirements for documenting how the Bureau defines and weighs factors used to select methods and documentation on methods considered but not tested. Agency Comments and Our Evaluation: The Secretary of Commerce provided written comments on a draft of this report on September 22, 2008. The comments are reprinted in appendix III. Commerce generally agreed with our conclusions and recommendations and stated that the Bureau is committed to developing and implementing a documented, systematic methodology for establishing response rate estimates for future censuses and reevaluating the estimates throughout the decade. Because the recommendations in our draft report focused on costs and planning for the 2010 Census, we revised our recommendations to reflect actions to be taken to support future census planning, including analyzing 2010 data to assess the reasonableness of assumptions used in applying the national estimate's components to the LCO-type estimates. Commerce also provided technical corrections, which we incorporated as appropriate. In its comments, Commerce disagreed with our statement that the Internet response option increased response in the Census 2000 experiment. Commerce cited a summary statement from a Bureau report that concluded that the use of multiple response modes tested in 2000 does not increase response. However, Bureau analyses from October 25, 2002, and March 2004 on the Census 2000 experiment stated that the overall response rate increased when the Internet was offered as an alternative response mode. We therefore made no changes in the report. Commerce strongly disagreed with our statement about the reason why the Bureau decided not to include the Internet option in the design of the 2010 Census. In the report, we state that underestimating contract costs was the primary reason the Bureau eliminated the Internet option, which we attribute to a July 19, 2006, memo documenting the Bureau's rationale for eliminating the Internet response option from the Decennial Response Integration System (DRIS) contract. This memo states that the decision "was due largely to the fact that the Census Bureau underestimated the FY 2006-2008 contractor costs proposed to develop DRIS." We therefore left this statement unchanged. However, we added the Bureau's explanation provided in its agency comment letter that the decision was based on test results, which showed that offering this option did not increase overall response and would not offer any cost savings. As agreed with your offices, unless you publicly announce the contents of this report earlier, we plan no further distribution until 30 days after its issue date. At that time, we will send copies of this report to the Secretary of Commerce, the Department of Commerce's Inspector General, the Director of the U.S. Census Bureau, and interested congressional committees. We will make copies available to others upon request. This report will also be available at no charge on GAO's Web site at [hyperlink, http://www.gao.gov]. If you or your staff have any questions concerning this report, please contact Mathew J. Scirč at (202) 512-6806 or sciremj@gao.gov or Ronald S. Fecso at (202) 512-2700 or fecsor@gao.gov. Contact points for our Offices of Congressional Relations and Public Affairs may be found on the last page of this report. GAO staff who made major contributions to this report are listed in appendix IV. Signed by: Mathew J. Scirč: Director, Strategic Issues: Signed by: Ronald S. Fecso: Chief Statistician: [End of section] Appendix I: Objectives, Scope, and Methodology: The objectives of this report were to (1) assess how the U.S. Census Bureau (Bureau) develops, supports, and updates the response rate estimate, and the extent to which the Bureau uses the response rate estimate to inform its 2010 planning efforts; (2) describe the methods the Bureau considered for increasing response rates in 2010 and what it did to test these methods; and (3) assess how the Bureau identifies and selects for testing methods to increase response rates, including considering other surveys' methods for increasing response. The scope of our review was limited to the census mailout/mailback universe, which covered more than 80 percent of households in Census 2000. The majority of households in this category have standard city-style addresses, allowing them to receive their census questionnaires in the mail in addition to being expected to return their questionnaires by mail. We excluded from our review operations and methods aimed at enumerating those not included in the mailout/mailback universe, such as the Be Counted initiative and Enumeration at Transitory Locations; those initiatives related to, but not primarily focused on, increasing self-response, such as improving the quality of the Master Address File and providing assistance through Questionnaire Assistance Centers; methods tested prior to 2000 and already implemented in previous censuses; and those primarily intended to improve operational efficiency, such as postal tracking. To determine how the Bureau develops and uses the response rate estimate, we reviewed documentation to support the components of the response rate estimate and research literature on efforts to estimate survey response rate. We interviewed Bureau officials in the Decennial Management Division about the process used to develop and update the estimate, the assumptions on which the estimate is based, and how the estimate is used. We also interviewed experts on survey methodology from Statistics Canada, the U.S. Department of Labor's Bureau of Labor Statistics, and various academic institutions, as well as former Census Bureau officials and survey methodologists to obtain their views on the strengths and weaknesses of the Bureau's process for developing and updating the estimate. We compared responses to a common set of questions we asked the experts in order to identify themes. We also reviewed Bureau strategic planning documents--in particular, the 2010 Census life cycle cost model with supporting documentation, as well as the 2010 Census Integrated Program Plan and the 2010 Census Operations and Systems Plan--to understand the Bureau's use of the estimate, and evaluated the Bureau's practices for using the response rate estimate for generating the life cycle cost estimate against best practices criteria in our Cost Assessment Guide and other relevant GAO products. To describe the methods the Bureau considered for increasing response rates in 2010 and how the Bureau tested these methods, we reviewed Bureau analyses and evaluations of tests on methods to increase mail response and various Bureau planning documents, such as the 2010 Census Operations and Systems Plan. In addition, we reviewed documentation on the Bureau's communications campaign, such as evaluations of the 2000 Partnership and Marketing Program and the 2010 Census Integrated Communications Campaign Plan. We interviewed Bureau officials in the Decennial Management Division to obtain additional context about the methods they considered and tests they conducted. To assess how the Bureau identifies and selects methods to test for increasing response rates, we reviewed various Bureau planning documents, such as research and development planning group action plans, for factors that the Bureau considers in selecting methods to test. We interviewed Bureau officials in the Decennial Management Division and other divisions about the process for identifying and selecting methods for increasing response. We also reviewed our past work on lessons learned from Census 2000 on the importance of documentation and planning in order to evaluate the Bureau's process for selecting and prioritizing methods to test. We interviewed experts on survey methodology from Statistics Canada, the U.S. Department of Labor's Bureau of Labor Statistics, and various academic institutions, as well as former Census Bureau officials and researchers, about additional methods for increasing response that the Bureau could consider and to obtain their perspectives on the Bureau's process for identifying, testing, and implementing methods. We compared responses to a common set of questions we asked the experts in order to identify recurring themes. We also reviewed research literature on methodologies to increase response to mail surveys to identify additional methods that the Bureau could consider. We interviewed the following experts in survey methodology: * Anil Arora, Director General, Census Program Branch, Statistics Canada: * Paul Biemer, Distinguished Fellow in statistics, Research Triangle Institute: * Don A. Dillman, Regents Professor, Thomas S. Foley Distinguished Professor of Government and Public Policy, and Deputy Director for Research and Development in the Social and Economic Sciences Research Center, Washington State University, and former Senior Survey Methodologist, U.S. Census Bureau: * Elizabeth Martin, former Senior Survey Methodologist, U.S. Census Bureau: * Colm O'Muircheartaigh, Vice President for Statistics and Methodology, National Opinion Research Center and Professor, Irving B. Harris Graduate School of Public Policy Studies, University of Chicago: * Kenneth Prewitt, Carnegie Professor of Public Affairs, School of International and Public Affairs, Columbia University and former Director, U.S. Census Bureau: * Clyde Tucker, Senior Survey Methodologist, Bureau of Labor Statistics: We conducted our review from July 2007 through September 2008 in accordance with generally accepted government auditing standards. Those standards require that we plan and perform the audit to obtain sufficient, appropriate evidence to provide a reasonable basis for our findings and conclusions based on our audit objectives. We believe that the evidence obtained provides a reasonable basis for our findings and conclusions based on our audit objectives. [End of section] Appendix II: Testing of Methods to Increase Response to 2010 Census: Replacement mailing. The Bureau plans to include the replacement mailing in the 2010 Census design. The replacement mailing involves sending a new questionnaire to households after households have received the initial census questionnaire in order to increase the likelihood that households will respond. The Bureau tested a targeted replacement mailing--where households that have not returned their initial questionnaires by a cutoff date receive the replacement form-- in the early 1990s and found that this method resulted in a 10 to 12 percent increase in the mail response rate. A blanket replacement mailing--where all households received a replacement questionnaire, including those that had already responded--was planned for Census 2000, but the Bureau dropped it from the design because of operational concerns that became apparent in the 1998 dress rehearsal. A targeted replacement mailing was incorporated into the 2010 design, based on the test results from the 1990s and the Bureau's plans to move to a short- form-only census. In 2003, the Bureau tested the targeted replacement mailing and was able to confirm the results of the tests from the early 1990s. Subsequent testing of the replacement mailing focused on ensuring that the Bureau could implement it successfully, including processing and removing the returns from the nonresponse follow-up workload using the handheld computers.[Footnote 21] The decision to eliminate the handheld computers from nonresponse follow-up and make nonresponse follow-up a paper-based operation diminished the Bureau's ability to remove late mail returns (including replacement questionnaires that are returned) from the nonresponse follow-up workload. The Bureau then examined options to get a better effect from the replacement mailing and decided to try to get the replacement mailings out more quickly. In order to do this, the Bureau decided that it would send a blanket replacement mailing to census tracts identified as low response areas, based on Census 2000 response data and updated demographic data from the American Community Survey. The replacement mailing packets for these tracts can be printed and labeled in advance--approximately 25 million total. Households not in these areas will receive a replacement questionnaire on a targeted basis--that is, only if they are in census tracts identified as medium response areas and if their initial questionnaires have not been received by the cutoff date. The Bureau estimates that it can print and label 15 million of these targeted replacement mailing packages in 4 days. Under this schedule, the Bureau will finish mailing out the replacement questionnaires on April 8, whereas under the original plan the replacement questionnaire mail out would not be completed until April 19. As a result, more households should be able to return the replacement questionnaires in time to be removed from the nonresponse follow-up workload. Two-column bilingual form. The Bureau plans to use a two-column bilingual form in certain locations in 2010. The two-column bilingual form provides two response columns, one in English and one in Spanish. Each column contains the same questions and response options, and respondents are instructed to choose the language that is most comfortable for them. The Bureau tested the bilingual form in 2005 to determine whether it would (1) increase overall response to the census and (2) lower item nonresponse (when a household returns the form but does not respond to a particular question) when compared to the standard English form. In the 2005 test, the two-column bilingual form panel demonstrated a significant positive effect on mail response rates of 2.2 percentage points overall, and a 3.2 percentage point increase in areas with high concentrations of Hispanic and non-White populations, but did not achieve lower rates of item nonresponse. The Bureau conducted another test of the two-column bilingual form in 2007, which produced response rate results similar to the 2005 results. In 2010, the Bureau plans to mail the two-column bilingual form to households in communities with heavy concentrations of Spanish speakers and areas with low English proficiency. Telephone reminder call. The Bureau does not plan to use a telephone reminder call in 2010. The multiple contact strategy that the Bureau used in 2000 included a reminder postcard mailed 1 week after the initial questionnaire packets were mailed. As part of the 2003 National Census Test, the Bureau tested the effect of an automated telephone reminder call in place of a reminder postcard on increasing the mail response rate. Out of a sample of 20,000 households, the Bureau was able to obtain phone numbers for 6,208 (31 percent), and these households received telephone calls to remind them to return their questionnaires if they had not already done so. The initial results indicated a significantly higher cooperation rate for the telephone reminder call panel when compared to the control panel, which received reminder postcards. The Bureau later conducted a supplementary analysis to determine if the higher cooperation rate was related to an underlying higher propensity to cooperate among households with listed telephone numbers. This analysis compared cooperation rates of two groups of households: those for which telephone numbers were available and those for which telephone numbers were not available. Among households for which a telephone number was available, the Bureau observed no significant difference in cooperation rates for housing units that received a reminder postcard compared to those that received a reminder telephone call. The Bureau does not plan to use a telephone reminder call in place of a reminder postcard in the 2010 Census. Due date on initial questionnaire mailing package. The Bureau will not place a due date on the initial questionnaire mailing package as part of the 2010 Census design. The value of placing a due date on the mailing package with the initial questionnaire is that it might evoke a sense of urgency or importance in respondents, leading to increased response to the census. In 2003, when the Bureau tested this method, it found that the inclusion of a due date did not significantly affect the overall response rate. However, it did result in a significantly higher response rate at an earlier point in the test, which can decrease the size of the replacement questionnaire mailing. The Bureau tested a due date again in 2006, this time combined with a "compressed schedule," in which questionnaires were mailed 14 days (as opposed to 21 days, as in the control group) before Census Day. A due date for mailing the questionnaire back was given in the advance letter, on the outgoing envelope and cover letter for the questionnaire mailing, and on the reminder postcard. The final report of the test concluded that the inclusion of the due date and compressed schedule resulted in a significantly higher mail response rate, by 2.0 percentage points. In addition, several measures of data completeness and coverage showed significant improvement. Bureau officials decided not to include the use of a due date in the 2010 Census, noting that they would like to conduct further testing to better understand the effects of a due date versus a compressed schedule. The Bureau is considering whether to test both the use of a due date and a compressed schedule in its 2010 Census Program for Evaluations and Experiments. Messaging to distinguish the replacement from the initial questionnaire. The Bureau does not plan to use a message to distinguish the replacement questionnaire from the initial questionnaire in 2010. The inclusion of a message on the replacement questionnaire informing respondents that they did not need to return the replacement questionnaire if they had already provided a response from the previous mailing was thought to be a way to reduce the number of multiple returns, and possibly improve overall response. The Bureau tested this method in 2005, and while the results showed a significant decline in the submission of multiple returns, they also showed a significant decrease in the response rate of 1.2 percentage points. Including a message to distinguish the replacement from the initial questionnaire was not recommended for, and will not be included in, the 2010 Census. Methods to Increase Self-Response Using Electronic Data Collection Systems: Interactive voice response. The Bureau does not plan to use interactive voice response in 2010. Interactive voice response allows respondents to use the telephone to respond verbally to digitized voice files that contain census questions and instructions. Speech recognition software is used to determine and record responses. Interactive voice response was tested in 2000, with households given the choice of providing their census data via interactive voice response or a paper questionnaire.[Footnote 22] The response rate for the interactive voice response panel was not statistically different from that of the control (mail only) group. However, the results are difficult to interpret because a portion of the sample in the interactive voice response panel either received the census form late or did not receive it at all. In 2003, the Bureau tested two different strategies for getting respondents to provide their census data using interactive voice response. One strategy, known as a push strategy, did not include a paper questionnaire in the initial mailing but rather pushed respondents to reply using interactive voice response. The other strategy, known as a choice strategy, included an initial paper questionnaire with the interactive voice response information on it, allowing respondents to choose their response mode. All nonrespondents received a paper replacement questionnaire. Households that had a choice of responding via paper or interactive voice response had overall response rates similar to households that only received paper, with about 7 percent of respondents given the choice using interactive voice response. Households in the push strategy had a significantly lower response rate--4.9 percentage points lower than households that only received a paper questionnaire. The Bureau does not plan to give respondents the option of providing their census data using interactive voice response in 2010. Letter encouraging Internet response or returning initial questionnaire instead of sending replacement questionnaire. The Bureau will not send a letter encouraging Internet response or returning the initial questionnaire instead of sending a replacement questionnaire in 2010. Instead of mailing a replacement paper questionnaire to nonrespondents, a letter could be sent to households encouraging them either to complete the paper questionnaire that they previously received or to use the Internet to submit their responses. The Bureau tested this method in 2005, sending a replacement mailing that contained a letter with an Internet address and an identification number needed to access the Internet questionnaire in place of a paper replacement questionnaire. Compared to sending out a paper replacement questionnaire, this method resulted in significantly fewer responses overall, decreasing the response rate by 3.7 percentage points. The Bureau will not include a letter encouraging the use of the Internet instead of a replacement questionnaire in 2010, in part because the Internet option was dropped from the 2010 Census design. Computer-assisted telephone interviewing. The Bureau will not use computer-assisted telephone interviewing for increasing response in 2010. Computer-assisted telephone interviewing allows respondents to use the telephone to connect with operators, who conduct interviews and record responses electronically. In a 2000 experiment, households were given the option of returning a paper form or providing their census data via computer-assisted telephone interviewing.[Footnote 23] When compared to households that were only given paper questionnaires, computer-assisted telephone interviewing brought about a 2.06 percentage point improvement in the overall response rate and also had a low item nonresponse rate. However, it entailed substantial costs for hardware, software, and programmer and interviewer time. The Bureau has not tested computer-assisted telephone interviewing since 2000 and does not plan to use this option in 2010. Internet. The Internet response option allows respondents to use an Internet-based questionnaire--with screens designed to resemble the paper questionnaire--to respond to the census. Respondents answer multiple-choice questions by clicking the appropriate buttons and checkboxes and text-entry questions by typing their answers into response fields. The Bureau gave respondents the option to respond by Internet in a 2000 experiment. Some households in the test were given the choice of providing their census data via Internet or a paper questionnaire.[Footnote 24] The experiment found that the Internet response option resulted in a 2.46 percentage point increase in response. In 2003, the Bureau again tested allowing respondents the option of providing their answers via the Internet by sending a paper questionnaire along with instructions for responding via the Internet in the initial mailing.[Footnote 25] Households that had a choice to respond by paper or the Internet had a similar overall response rate to households that were provided only paper, with about 10 percent of respondents choosing to respond by Internet. The Bureau decided not to include the Internet in the 2010 Census, despite including it in the scope of the contract awarded in 2005 for the Decennial Response Integration System. The Bureau noted that this decision was based on a number of factors, including the Bureau's underestimation of the contractor costs for the first 3 years of the contract, as well as test results that indicated that the Internet would not increase the response rate and concerns about the security of respondents' data prior to the Bureau receiving it. Communications campaign. Census 2000 included a greatly expanded outreach and promotion campaign--including, for the first time, paid advertising--in an attempt to increase public awareness of and promote positive attitudes about the census. This program, called the Partnership and Marketing Program, was considered a success after the Bureau reversed the trend of declining mail response rates, and the Bureau made plans to continue the program for 2010. Bureau officials stated that the 2010 campaign consolidates all census communications under a single communications contract. This Integrated Communications Campaign aims to motivate the entire populations of the 50 states, the District of Columbia, Puerto Rico, and other U.S. territories to participate in the census through partnerships (with community groups, businesses, colleges, faith-based organizations, and other targeted groups); public relations; events; Census in Schools; and paid advertisements in broadcast, print, and online media. Bureau officials noted that the advertising campaign would not be included in the dress rehearsal because the Bureau's experience including the advertising campaign in the 1998 Dress Rehearsal did not provide the feedback needed to revise the creative aspects of the campaign. To refine the communications campaign, the Bureau conducted an audience segmentation analysis to identify how to best reach people with the paid advertising campaign. In addition, the Bureau conducted focus groups in 2006 and 2007 to provide information on what motivates individuals to respond. The communications campaign is scheduled to run from mid-2008 through June 2010. In 2008 and 2009, most of the activities are focused on preparing and mobilizing partnerships. Further development of specific messages for various audiences and communication channels will take place from November 2008 through April 2009. Starting in mid-2009, the partnership program will begin outreach to certain hard-to-count populations. Activities and events targeting all audiences will begin in January 2010 and peak in March 2010. [End of section] Appendix III: Comments from the Department of Commerce: The Secretary Of Commerce: Washington, D.C. 20230: September 19, 2008: Mr. Mathew J. Scire: Director: Strategic Issues: United States Government Accountability Office: Washington, DC 20548: Dear Mr. Scire: I enclose the U.S. Department of Commerce's comments in response to recommendations contained in the United States Government Accountability Office's Draft Report, 2010 Census: Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate (GAO-08-1012). Sincerely, Signed by: Carlos M. Gutierrez: Enclosure: Comments in Response to the United States Government Accountability Office Draft Report Entitled: 2010 Census: Census Bureau Needs Procedures for Estimating the Response Rate and Selecting for Testing Methods to Increase Response Rate: September 30, 2008: General Comments: The Census Bureau appreciates the opportunity to review the GAO's draft report on this subject. We are in agreement with GAO's central fording”that the 2010 Census response rate estimate is not fully supported by a documented, systematic methodology, and that it has not been reevaluated since it was first established in 2001. While this is unfortunate, we believe that at this late date we cannot perform additional tests or substantive research that would allow us to establish a revised estimate. While it is possible that another review of available information might lead us to revise the projected response rate, such a revision would involve interpretation; it would not be based on significant test data. Moreover, we feel strongly that despite declining response rates for censuses and surveys throughout the decade, the decennial census environment is unique. For example, all people, households, and governments are involved, there is extensive outreach, partnership, and promotion, and it is a mandatory activity required by the U.S. Constitution. Accordingly, we remain persuaded that the response rate of 64% arrived at through examination of Census 2000 data is viable for planning purposes. We recognize that the actual response rate in 2010 may be different than we project. From our perspective, the major risk comes from overestimating the rate. If the actual rate is significantly lower than we project (as happened in 1990), we will be faced with some major challenges that will be difficult to overcome in the middle of census operations. We have very little leeway in our schedule at that point; the dates for producing apportionment and redistricting data are fixed by law. Thus, we would need to utilize contingency funding”or if necessary, request additional funding from Congress” in order to quickly hire and train more field staff so as to stay on schedule. The Census Bureau is committed to developing and implementing a documented, systematic methodology for establishing response rate estimates for future decennial censuses, beginning with an analysis of 2010 data in preparation for the 2020 Census. We also commit to reevaluating our estimates throughout the decade. Our specific comments on the report follow. Detailed Comments: On page 3, toward the middle of the page, the report states that we have not reevaluated the response rate change to 64% (which we made after the decision to drop use of handheld computers for nonresponse follow-up). Last month we briefed auditors on some of the changes made to the response rate estimate, and our preliminary revisions to the use of replacement mailings. At that time, we also shared with them the Census 2000 data and our analysis of response patterns that led to our decision to lower the estimated rate to 64%. At the top of page 5, the report states that our test results of Internet response "did not always increase the response rate." In our tests in Census 2000 and during this decade, the option never increased the overall response rate, so we suggest the sentence be reworded to accurately reflect our test results. At the bottom of page 5, the GAO recommends "that the Secretary of Commerce direct the Bureau to establish and implement procedures for documenting the process for developing the response rate estimate, as well as establishing when and how to reevaluate the response rate estimate." As we state in our summary comments, the Census Bureau is committed to doing this in preparation for the 2020 Census. On page 6, the report states: "Households that fail to mail back the census questionnaires are included in nonresponse follow-up workload, where enumerators follow up with telephone calls and door-to-door visits, or solicit census data from knowledgeable people, such as neighbors." This sentence should be adjusted to accurately capture the sequence of events that occurs during the census. First, one or more door-to-door visits is made. If no one is home, or if the respondent is busy, contact information is left, after which it may be possible to conduct the enumeration interview over the phone. As a last resort, when no data has been collected and no time is left to make additional visits, census data may be solicited from knowledgeable people, such as neighbors. On pages 7-8, footnote 6 states: "The Bureau also measures the percentage of surveys completed and returned for occupied housing units with deliverable addresses, called the return rate. Although the return rate is considered the more precise measure of public cooperation with the census, it is less useful than response rate for measuring nonresponse follow-up workload " We believe this footnote should be clarified to note that the mail return rate cannot be calculated until after nonresponse follow-up (NRFU) has been completed. The mailout universe includes all the addresses that will turn out to be vacant, and some that will be deleted as nonexistent or non-residential, during the NRFU operation. At the bottom of page 12, the discussion about replacement mailing effects states that our estimate of the effect was based on research during the 1990s. Replacement questionnaires were also tested in the 2003 National Census Survey, the 2005 National Census Test, the 2006 Census Test, and in the 2008 Census Dress Rehearsal. Our estimates are based on these more recent results. Also, we believe we get a larger impact from mailing replacement questionnaires in a test than in a census because during a census, we get a higher response rate in the first place. That is, a higher proportion of households will reply to the initial mailing during the actual census (as a result of paid advertising and partnership efforts), so this reduces the potential number of households that need to respond to the second mailing. Some of the testing that has occurred this decade deals with the timing. of printing and sending replacement questionnaires compared to the enumeration activities of NRFU. We were largely successful in removing very late mail returns from the NRFU workload in an automated environment. Due to the return to a paper-based NRFU operation, our revised plans for 2010 involve printing and distributing the replacement questionnaires earlier in hopes of gaining some of the same benefit. However, given the late decision to move to a paper-based NRFU operation, these plans have not been tested. In the first paragraph on page 16, the last portion needs to be revised to clarify that the operational plans to implement our revised replacement mailing approach are still being finalized. In that regard, the additional analysis that we are conducting does not pertain to fine- tuning the estimated effects; it pertains to the specifics of how to operationalize the approach by Type of Enumeration Area. Finally, because we have never tested these revisions under true decennial census conditions, we have no basis for precisely estimating how much response rate improvement they will yield. At the bottom of page 22, the report states that the Internet response option increased overall response during the Census 2000 experiment. We disagree with such a broad conclusion. During Census 2000, we conducted an experiment to examine both the use of incentives and the use of other response modes (Internet, interactive voice response, and an operator telephone interview). While there were variations in response rates for these different modes depending on whether incentives also were offered, the overall conclusion was: "The use of alternative response modes does not increase overall response rates to the census. Rather, it shifts households who would respond via the paper census to the other mode." (See Synthesis of Results from the Response Mode and Incentive Experiment for Census 2000 (RMIE), March 2003.) On page 23, we strongly disagree with the report's summary of why we decided not to pursue an Internet response option for the 2010 Census. On multiple occasions, the Census Bureau has stated that the decision was made because our tests had shown that offering this option did not increase overall response rate, and thus would not offer any cost savings, much less offset the actual costs that would be incurred to offer the option. While it is true we also have expressed”and continue to have”concerns about Internet security, that was not the primary reason we decided not to pursue this for the 2010 Census. On page 35, with regard to the discussion of the bilingual form, we wish to clarify that the results for item non-response for the 2005 National Census Test were thought to be attributed to a form design problem. The form was tested again in a survey in 2007 with changes to the form design, which corrected to some extent the problems with item non- response. Also, it should be noted that the determination of when to use the bilingual form is based not just on whether an area has a heavy concentration of Spanish speakers, but also on whether an area has low English proficiency. On page 36, the second paragraph regarding telephone reminder calls should say that the initial results indicated a significantly higher cooperation rate. It is also worth mentioning in this paragraph that there are costs to obtaining the phone numbers and making the calls, that there was a low percentage of phone numbers found in 2003, and that there is increasing use of individual cell phones over household landlines. In deciding whether to pursue this option, all of these considerations have to be weighed. [End of section] Appendix IV: GAO Contacts and Staff Acknowledgments: GAO Contacts: Mathew J. Scirč, (202) 512-6806 or sciremj@gao.gov: Ronald S. Fecso, (202) 512-2700 or fecsor@gao.gov: Acknowledgments: In addition to contacts named above, Lisa Pearson, Assistant Director; David Bobruff; Don Brown; and Elizabeth Fan made key contributions to this report. Tom Beall, Susan Etzel, Andrea Levine, Donna Miller, Ellen Rominger, and Elizabeth Wood provided significant technical support. [End of section] Footnotes: [1] The $13.7 billion to $14.5 billion estimate is expressed by the Bureau in nominal year dollars. [2] In this report, we use the term response rate to refer to the national mail response rate the Bureau uses to estimate nonresponse follow-up workload, unless otherwise noted. [3] The Bureau uses the U.S. Postal Service to deliver questionnaires to housing units with city-style addresses (house number and street name). However, in areas where housing units do not receive mail at a city- style address, the Bureau delivers questionnaires through an update/ leave process, in which enumerators deliver a census questionnaire to each housing unit. The household is asked to complete and return the questionnaire by mail. [4] According to the Bureau, these rates reflect mail response received as of the cutoff dates for determining nonresponse follow-up workload, and because these rates were computed differently, caution should be used in comparing the rates. [5] The Census 2000 response rate includes responses received by Internet, which other years do not include. [6] The Bureau also measures the percentage of surveys completed and returned for occupied housing units with deliverable addresses, called the return rate, which is calculated after nonresponse follow-up has been completed. Although the return rate is considered the more precise measure of public cooperation with the census, it is less useful than response rate for measuring nonresponse follow-up workload. [7] The 65 percent rate used as the baseline for the 2010 estimate is the rate that was achieved for the mailout/mailback universe on April 18, 2000, when the nonresponse follow-up universe was identified. [8] The Bureau later described the ability to remove all late mail returns from the nonresponse follow-up workload--both initial questionnaires and replacement questionnaires--as included in the replacement mailing effect. [9] GAO, Cost Assessment Guide: Best Practices for Estimating and Managing Program Costs: Exposure Draft, GAO-07-1134SP (Washington, D.C.: July 2007). [10] The response rates in the American Community Survey analysis are comparable to the Bureau's definition of mail return rates for the decennial census in that vacant and nonexistent housing units are excluded from the denominator in the calculation. The Bureau's mail response rate, as defined for the decennial census, includes vacant and nonexistent housing units in the denominator. [11] The Bureau has used some of these data for identifying hard-to- count populations as part of the planning for the 2010 communications campaign, but these data are not incorporated into the mail response rate estimate. [12] In a blanket replacement mailing, all households in a certain area would receive a replacement questionnaire, regardless of whether they had already responded. In a targeted replacement mailing, only households that have not returned their initial questionnaire by a cutoff date would receive a replacement questionnaire. The original plan for the 2010 Census included using a targeted replacement mailing for nonresponding households in mailout/mailback areas. The revised plan uses a blanket replacement mailing for some areas, targeted replacement mailing for other areas, and no replacement mailing for the remaining areas. [13] In 2000, Type A LCOs were located in inner-city and urban areas. Type B offices were located in urban and metropolitan areas. Type C offices were located in suburban areas, small and medium-size cities, towns, and some rural areas. Type D offices were located in more rural areas. [14] The replacement questionnaire and two-column bilingual English/ Spanish census form are planned to be implemented for the first time in the 2010 Census. [15] Under the compressed mailing schedule, questionnaires were mailed 14 days as opposed to 21 days before Census Day. [16] National Academy of Sciences, Experimentation and Evaluation Plans for the 2010 Census: Interim Report (Washington, D.C.: Dec. 7, 2007). [17] The National Academy of Sciences has convened a panel of experts to review the Bureau's program of research, evaluation, and experimentation for the 2008 Dress Rehearsal and the 2010 Census. The panel will consider priorities for evaluation in the 2010 Census. The panel will conduct its work over a 3-year period, from September 2006 to September 2009. [18] In September 2007, the Bureau awarded a contract to DraftFCB, Inc., a communications agency, to create, produce, and implement an integrated marketing and communications campaign in support of the 2010 Census. The contract, with an estimated value of around $200 million to $300 million, was structured as a series of task orders for specific pieces of the campaign, to be issued over 4 years. [19] GAO, 2000 Census: Lessons Learned for Planning a More Cost- Effective 2010 Census, GAO-03-40 (Washington, D.C.: Oct. 31, 2002). [20] Conducted every 5 years, the Economic Census surveys U.S. businesses and provides official measures of output for industries and geographic areas and key source data for the gross domestic product and other indicators of economic performance. [21] In addition to using the handheld computers to enumerate households during nonresponse follow-up, the Bureau planned to use them for managing enumerators' workloads during nonresponse follow-up. Late mail returns, including replacement questionnaires, would be removed automatically from the nonresponse follow-up workload as soon as they were checked in, so long as the households had not yet been visited by an enumerator. In this way, the response rate estimate would include responses received even after the beginning of the nonresponse follow- up operation. [22] Another interactive voice response test panel was offered an incentive--a telephone calling card good for 30 minutes of domestic calls--for using interactive voice response instead of the paper questionnaire. The incentive did not increase the overall response rate compared to that of the mail-only control panel. [23] Another computer-assisted telephone interviewing test panel was offered an incentive--a telephone calling card good for 30 minutes of domestic calls--for using computer-assisted telephone interviewing instead of the paper questionnaire. The overall response rate for this incentive panel (71.81 percent) was similar to the overall response rate for the paper-only control panel (71.44 percent), though more respondents from the incentive panel chose to use computer-assisted telephone interviewing than the panel where no incentive was offered. [24] Another Internet test panel was offered an incentive--a telephone calling card good for 30 minutes of domestic calls--for using the Internet instead of the paper questionnaire. The overall response rate for this incentive panel (71.50 percent) was similar to the overall response rate for the paper-only control panel (71.44 percent), though more respondents from the incentive panel chose to use the Internet than the panel where no incentive was offered. [25] Another panel received a paper questionnaire along with instructions for responding via the Internet or interactive voice response. This panel had a similar response rate as the paper-only panel, with about 7 percent of respondents using the Internet. A third panel received instructions for responding via the Internet or interactive voice response but no initial paper questionnaire. This panel had a significantly lower response rate--by 5.7 percentage points--than households provided only the paper questionnaire. GAO's Mission: The Government Accountability Office, the audit, evaluation and investigative arm of Congress, exists to support Congress in meeting its constitutional responsibilities and to help improve the performance and accountability of the federal government for the American people. GAO examines the use of public funds; evaluates federal programs and policies; and provides analyses, recommendations, and other assistance to help Congress make informed oversight, policy, and funding decisions. GAO's commitment to good government is reflected in its core values of accountability, integrity, and reliability. Obtaining Copies of GAO Reports and Testimony: The fastest and easiest way to obtain copies of GAO documents at no cost is through GAO's Web site [hyperlink, http://www.gao.gov]. Each weekday, GAO posts newly released reports, testimony, and correspondence on its Web site. To have GAO e-mail you a list of newly posted products every afternoon, go to [hyperlink, http://www.gao.gov] and select "E-mail Updates." Order by Mail or Phone: The first copy of each printed report is free. Additional copies are $2 each. A check or money order should be made out to the Superintendent of Documents. GAO also accepts VISA and Mastercard. Orders for 100 or more copies mailed to a single address are discounted 25 percent. Orders should be sent to: U.S. Government Accountability Office: 441 G Street NW, Room LM: Washington, D.C. 20548: To order by Phone: Voice: (202) 512-6000: TDD: (202) 512-2537: Fax: (202) 512-6061: To Report Fraud, Waste, and Abuse in Federal Programs: Contact: Web site: [hyperlink, http://www.gao.gov/fraudnet/fraudnet.htm]: E-mail: fraudnet@gao.gov: Automated answering system: (800) 424-5454 or (202) 512-7470: Congressional Relations: Ralph Dawn, Managing Director, dawnr@gao.gov: (202) 512-4400: U.S. Government Accountability Office: 441 G Street NW, Room 7125: Washington, D.C. 20548: Public Affairs: Chuck Young, Managing Director, youngc1@gao.gov: (202) 512-4800: U.S. Government Accountability Office: 441 G Street NW, Room 7149: Washington, D.C. 20548:

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.