Child And Family Services Reviews

States and HHS Face Challenges in Assessing and Improving State Performance Gao ID: GAO-04-781T May 13, 2004

In 2001, the Department of Health and Human Services' (HHS) Administration for Children and Families (ACF) implemented the Child and Family Services Reviews (CFSR) to increase states' accountability. The CFSR uses states' data profiles and statewide assessments, as well as interviews and an on-site case review, to measure state performance on 14 outcomes and systemic factors, including child well-being and the provision of caseworker training. The CFSR also requires progress on a program improvement plan (PIP); otherwise ACF may apply financial penalties. This testimony is based on our April 2004 report and addresses (1) ACF's and the states' experiences preparing for and conducting the statewide assessments and on-site reviews; (2) ACF's and the states' experiences developing, funding, and implementing items in PIPs; and (3) any additional efforts that ACF has taken beyond the CFSR to improve state performance. For the April 2004 report, we surveyed all 50 states, the District of Columbia, and Puerto Rico regarding their experiences throughout the CFSR process, visited 5 states to obtain first-hand information, and conducted a content analysis of all 31 available PIPs as of January 1, 2004. We also interviewed HHS officials--including those in all 10 regional offices--and key child welfare experts.

ACF and many state officials perceive the CFSR as a valuable process and a substantial undertaking, but some data enhancements could improve its reliability. ACF staff in 8 of the 10 regions considered the CFSR a helpful tool to improve outcomes for children. Further, 26 of 36 states responding to a relevant question in our survey commented that they generally or completely agreed with the results of the final CFSR report, even though none of the 41 states with final CFSR reports released through 2003 has achieved substantial conformity on all 14 outcomes and systemic factors. Additionally, both ACF and the states have dedicated substantial financial and staff resources to the process. Nevertheless, several state officials and child welfare experts we interviewed questioned the accuracy of the data used in the review process. While ACF officials contend that stakeholder interviews and case reviews complement the data profiles, many state officials and experts reported that additional data from the statewide assessment could bolster the evaluation of state performance. Program improvement planning is under way, but uncertainties have affected the development, funding, and implementation of state PIPs. Officials from 3 of the 5 states we visited said ACF's PIP-related instructions were unclear, and at least 9 states reported in our survey that challenges to implementing their plans include insufficient funding, staff, and time. While ACF has provided some guidance, ACF and state officials remain uncertain about PIP monitoring efforts and how ACF will apply financial penalties if states fail to achieve their stated PIP objectives. Since 2001, ACF's focus has been almost exclusively on the CFSRs and regional staff report limitations in providing assistance to states in helping them to meet key federal goals. While staff from half of ACF's regions told us they would like to provide more targeted assistance to states, and state officials in all 5 of the states we visited said that ACF's existing technical assistance efforts could be improved, ACF officials acknowledged that regional staff might still be adjusting to the new way ACF oversees child welfare programs. In the April 2004 report, we recommended that the Secretary of HHS ensure that ACF uses the best available data to measure state performance. We also recommended that the Secretary clarify PIP guidance and provide guidance to regional officials on how to better integrate their many oversight responsibilities. In commenting on a draft of the April 2004 report, HHS acknowledged that the CFSR is a new process that continues to evolve, and noted several steps it has taken to address the data quality concerns we raised in that report.



GAO-04-781T, Child And Family Services Reviews: States and HHS Face Challenges in Assessing and Improving State Performance This is the accessible text file for GAO report number GAO-04-781T entitled 'Child And Family Services Reviews: States and HHS Face Challenges in Assessing and Improving State Performance' which was released on May 13, 2004. This text file was formatted by the U.S. General Accounting Office (GAO) to be accessible to users with visual impairments, as part of a longer term project to improve GAO products' accessibility. Every attempt has been made to maintain the structural and data integrity of the original printed product. Accessibility features, such as text descriptions of tables, consecutively numbered footnotes placed at the end of the file, and the text of agency comment letters, are provided but may not exactly duplicate the presentation or format of the printed version. The portable document format (PDF) file is an exact electronic replica of the printed version. We welcome your feedback. Please E-mail your comments regarding the contents or accessibility features of this document to Webmaster@gao.gov. This is a work of the U.S. government and is not subject to copyright protection in the United States. It may be reproduced and distributed in its entirety without further permission from GAO. Because this work may contain copyrighted images or other material, permission from the copyright holder may be necessary if you wish to reproduce this material separately. Testimony: Before the Subcommittee on Human Resources, Committee on Ways and Means, House of Representatives: United States General Accounting Office: GAO: For Release on Delivery Expected at 10:00 a.m. EDT: Thursday, May 13, 2004: Child and Family Services Reviews: States and HHS Face Challenges in Assessing and Improving State Performance: Statement of Cornelia M. Ashby, Director, Education, Workforce, and Income Security Issues: GAO-04-781T: GAO Highlights: Highlights of GAO-04-781T, a testimony before the Subcommittee on Human Resources, Committee on Ways and Means, House of Representative Why GAO Did This Study: In 2001, the Department of Health and Human Services‘ (HHS) Administration for Children and Families (ACF) implemented the Child and Family Services Reviews (CFSR) to increase states‘ accountability. The CFSR uses states‘ data profiles and statewide assessments, as well as interviews and an on-site case review, to measure state performance on 14 outcomes and systemic factors, including child well-being and the provision of caseworker training. The CFSR also requires progress on a program improvement plan (PIP); otherwise ACF may apply financial penalties. This testimony is based on our April 2004 report and addresses (1) ACF‘s and the states‘ experiences preparing for and conducting the statewide assessments and on-site reviews; (2) ACF‘s and the states‘ experiences developing, funding, and implementing items in PIPs; and (3) any additional efforts that ACF has taken beyond the CFSR to improve state performance. For the April 2004 report, we surveyed all 50 states, the District of Columbia, and Puerto Rico regarding their experiences throughout the CFSR process, visited 5 states to obtain first-hand information, and conducted a content analysis of all 31 available PIPs as of January 1, 2004. We also interviewed HHS officials”including those in all 10 regional offices”and key child welfare experts. What GAO Found: ACF and many state officials perceive the CFSR as a valuable process and a substantial undertaking, but some data enhancements could improve its reliability. ACF staff in 8 of the 10 regions considered the CFSR a helpful tool to improve outcomes for children. Further, 26 of 36 states responding to a relevant question in our survey commented that they generally or completely agreed with the results of the final CFSR report, even though none of the 41 states with final CFSR reports released through 2003 has achieved substantial conformity on all 14 outcomes and systemic factors. Additionally, both ACF and the states have dedicated substantial financial and staff resources to the process. Nevertheless, several state officials and child welfare experts we interviewed questioned the accuracy of the data used in the review process. While ACF officials contend that stakeholder interviews and case reviews complement the data profiles, many state officials and experts reported that additional data from the statewide assessment could bolster the evaluation of state performance. Program improvement planning is under way, but uncertainties have affected the development, funding, and implementation of state PIPs. Officials from 3 of the 5 states we visited said ACF‘s PIP-related instructions were unclear, and at least 9 states reported in our survey that challenges to implementing their plans include insufficient funding, staff, and time. While ACF has provided some guidance, ACF and state officials remain uncertain about PIP monitoring efforts and how ACF will apply financial penalties if states fail to achieve their stated PIP objectives. Since 2001, ACF‘s focus has been almost exclusively on the CFSRs and regional staff report limitations in providing assistance to states in helping them to meet key federal goals. While staff from half of ACF‘s regions told us they would like to provide more targeted assistance to states, and state officials in all 5 of the states we visited said that ACF‘s existing technical assistance efforts could be improved, ACF officials acknowledged that regional staff might still be adjusting to the new way ACF oversees child welfare programs. In the April 2004 report, we recommended that the Secretary of HHS ensure that ACF uses the best available data to measure state performance. We also recommended that the Secretary clarify PIP guidance and provide guidance to regional officials on how to better integrate their many oversight responsibilities. In commenting on a draft of the April 2004 report, HHS acknowledged that the CFSR is a new process that continues to evolve, and noted several steps it has taken to address the data quality concerns we raised in that report. www.gao.gov/cgi-bin/getrpt?GAO-04-781T. To view the full product, including the scope and methodology, click on the link above. For more information, contact Corneila Ashby at (202) 512-8403 or ashbyc@gao.gov. [End of section] Mr. Chairman and Members of the Subcommittee: Thank you for inviting me here today to discuss states' efforts to comply with federal Child and Family Services Reviews (CFSR). As you are aware, in 2001, the Department of Health and Human Services' (HHS) Administration for Children and Families' (ACF) began implementing the CFSRs to hold states accountable for improving child welfare outcomes. Unlike prior federal reviews--which determined states' adherence to certain process measures--ACF designed the CFSR as an outcome-oriented approach to assess children's safety; their timely placement in permanent homes; and their mental, physical, and educational well- being; and it developed certain standards against which to measure states' success in these areas.[Footnote 1] ACF also designed the reviews to assess states' performance across a range of systemic factors, such as caseworker training and foster parent licensing. The CFSR has multiple phases, consisting of a statewide assessment; an on- site review, which culminates in the release of a final report; and the development and implementation of a program improvement plan (PIP) when states are found to be deficient. Pursuant to CFSR regulations, ACF can withhold federal funds if states do not show adequate progress implementing their PIPs. My testimony today will focus on three key issues: (1) ACF's and the states' experiences preparing for and conducting the statewide assessments and on-site reviews; (2) ACF's and the states' experiences developing, funding, and implementing items in their PIPs; and (3) additional efforts, if any, that ACF has taken beyond the CFSR to help ensure that all states meet federal goals of safety, permanency, and well-being for children. My comments are based on the findings from our April 2004 report.[Footnote 2] Those findings were based on a survey of all 50 states, the District of Columbia, and Puerto Rico regarding their experiences during each phase of the CFSR process;[Footnote 3] post-survey follow up phone calls with key states;[Footnote 4] and site visits to California, Florida, New York, Oklahoma, and Wyoming to obtain first-hand information on states' experiences. We selected these states for diversity in their location, size, program administration, performance on the CFSR, and the timing of their review. We also examined all 31 approved PIPs available as of January 1, 2004, and conducted interviews with ACF's senior officials, regional staff from all 10 regions, ACF contractors, staff from all 10 national resource centers,[Footnote 5] and key child welfare experts. We conducted our work between May 2003 and February 2004 in accordance with generally accepted government auditing standards. In summary, ACF and many state officials perceive the CFSR as a valuable process and a substantial undertaking, but some data enhancements could improve its reliability. ACF staff in 8 of the 10 regions considered the CFSR a helpful tool to improve outcomes for children, and 26 of 36 states responding to a relevant question in our survey commented that they generally or completely agreed with the results of the final CFSR report, even though none of the 41 states with final CFSR reports released through 2003 has achieved substantial conformity on all CFSR outcomes and systemic factors. Additionally, both ACF and the states have dedicated substantial financial and staff resources to the process. Nevertheless, several state officials and child welfare experts we interviewed questioned the accuracy of the data used in the review process and noted that additional data from the statewide assessment could bolster the evaluation of state performance. While states' PIP planning is under way, uncertainties have affected the development, funding, and implementation of these plans. Officials from 3 of the 5 states we visited said ACF's PIP-related instructions were unclear, and at least 9 of the 25 states reporting on PIP implementation in our survey stated that insufficient funding, staff, and time, as well as high caseloads, were among the greatest challenges. While ACF has provided some guidance, ACF and state officials remain uncertain about PIP monitoring efforts and how ACF will apply financial penalties if states fail to achieve their stated PIP objectives. Further, since 2001, ACF's focus has been almost exclusively on the CFSRs and regional staff report limitations in providing assistance to states in helping them to meet key federal goals. To improve its oversight, we recommended in our April 2004 report that the Secretary of HHS ensure that ACF use the best available data to measure state performance, clarify PIP guidance, and help regional offices better integrate their oversight responsibilities. Background: ACF's Children's Bureau administers and oversees federal funding to states for child welfare services under Titles IV-B and IV-E of the Social Security Act, and states and counties provide these child welfare services, either directly or indirectly through contracts with private agencies.[Footnote 6] Among other activities, ACF staff are responsible for developing appropriate policies and procedures for states to follow to obtain and use federal child welfare funds, reviewing states' planning documents required by Title IV-B, conducting states' data system reviews, assessing states' use of Title IV-E funds, and providing technical assistance to states through all phases of the CFSR process. In addition, ACF staff coordinate the work of the 10 resource centers to provide additional support and assistance to the states. Spurred by the passage of the 1997 Adoption and Safe Families Act (ASFA), ACF launched the CFSR in 2001 to improve its existing monitoring efforts, which had once been criticized for focusing exclusively on states' compliance with regulations rather than on their performance over a full range of child welfare services. The CFSR process combines a statewide self-assessment, an on-site case file review that is coupled with stakeholder interviews,[Footnote 7] and the development and implementation of a 2-year PIP with performance benchmarks to measure process in improving noted deficiencies. In assessing performance through the CFSR, ACF relies, in part, on its own data systems, known as NCANDS and AFCARS, which were designed prior to CFSR implementation to capture, report, and analyze the child welfare information collected by the states.[Footnote 8] Today, these systems provide the national data necessary for ACF to calculate national standards for key performance items against which all states are measured and to determine, in part, whether or not states are in substantial conformity on CFSR outcomes and systemic factors.[Footnote 9] Once ACF approves the PIP, states are required to submit quarterly progress reports. Pursuant to CFSR regulations, federal child welfare funds can be withheld if states do not show adequate PIP progress, but these penalties are suspended during the 2-year PIP implementation term.[Footnote 10] In preparation for the next round of CFSRs, ACF officials have formed a Consultation Work Group of ACF staff, child welfare administrators, data experts, and researchers who will propose recommendations on the CFSR measures and processes. The group's resulting proposals for change, if any, are not yet available. The CFSR Is a Valuable Yet Substantial Undertaking, but Data Enhancements Could Improve Its Reliability: ACF and many state officials perceive the CFSR as a valuable process-- highlighting many areas needing improvement--and a substantial undertaking, but some state officials and child welfare experts told us that data enhancements could improve its reliability. ACF staff in 8 of the 10 regions considered the CFSR a helpful tool to improve outcomes for children. Further, 26 of the 36 states responding to a relevant question in our survey commented that they generally or completely agreed with the results of the final CFSR report, even though none of the 41 states with final CFSR reports released through 2003 has achieved substantial conformity on all 14 outcomes and systemic factors. In addition, both ACF and the states have dedicated substantial financial and staff resources to the process. However, several state officials and child welfare experts we interviewed questioned the accuracy of the data used to compile state profiles and establish the national standards. While ACF officials in the central office contend that stakeholder interviews and case reviews compliment the data profiles, many state officials and experts reported that additional data from the statewide assessment could bolster the evaluation of state performance. The CFSR Is a Valuable Process for ACF and the States: ACF and state officials support the objectives of the review, especially in focusing on children's outcomes and strengthening relationships with stakeholders, and told us they perceive the process as valuable. For example, ACF officials from 8 regional offices noted that the CFSRs were more intensive and more comprehensive than the other types of reviews they had conducted in the past, creating a valuable tool for regional officials to monitor states' performance. In addition, state officials from every state we visited told us that the CFSR process helped to improve collaboration with community stakeholders. Furthermore, state staff from 4 of the 5 states we visited told us the CFSR led to increased public and legislative attention to critical issues in child welfare. For example, caseworkers in Wyoming told us that without the CFSR they doubted whether their state agency's administration would have focused on needed reforms. They added that the agency used the CFSR findings to request legislative support for the hiring of additional caseworkers. Along with the value associated with improved stakeholder relations, the ACF officials we talked to and many state officials reported that the process has been helpful in highlighting the outcomes and systemic factors, as well as other key performance items that need improvement. According to our survey, 26 of the 36 states that commented on the findings of the final CFSR report indicated that they generally or completely agreed with the findings, even though performance across the states was low in certain key outcomes and performance items. For example, not one of the 41 states with final reports released through 2003 was found to be in substantial conformity with either the outcome measure that assesses the permanency and stability of children's living situations or with the outcome measure that assesses whether states had enhanced families' capacity to provide for their children's needs. Moreover, across all 14 outcomes and systemic factors, state performance ranged from achieving substantial conformity on as few as 2 outcomes and systemic factors to as many as 9.[Footnote 11] As figure 1 illustrates, the majority of states were determined to be in substantial conformity with half or fewer of the 14 outcomes and systemic factors assessed. Figure 1: State Performance on the 14 CFSR Outcomes and Systemic Factors: [See PDF for image] [End of figure] States' performance on the outcomes related to safety, permanency, and well-being--as well as the systemic factors--is determined by their performance on an array of items, such as establishing permanency goals, ensuring worker visits with parents and children, and providing accessible services to families. The CFSR showed that many states need improvement in the same areas. For example, across all 41 states reviewed through 2003, the 10 items most frequently rated as needing improvement included assessing the needs and services of children, parents, and foster parents (40 states); assessing the mental health of children (37 states); and establishing the most appropriate permanency goal for the child (36 states). ACF and the States Report That Reviews Have Been a Substantial Undertaking: Given the value that ACF and the states have assigned to the CFSR process, both have spent substantial financial resources and staff time to prepare for and implement the reviews. In fiscal years 2001-03, when most reviews were scheduled, ACF budgeted an additional $300,000 annually for CFSR-related travel. In fiscal year 2004, when fewer reviews were scheduled, ACF budgeted about $225,000. To further enhance its capacity to conduct the reviews, and to obtain additional logistical and technical assistance, ACF spent approximately $6.6 million annually to hire contractors. Specifically, ACF has let three contracts to assist with CFSR-related activities, including training reviewers to conduct the on-site reviews, tracking final reports and PIP documents, and, as of 2002, writing the CFSR final reports. Additionally, ACF hired 22 new staff to build central and regional office capacity and dedicated 4 full-time staff and 2 state government staff temporarily on assignment with ACF to assist with the CFSR process. To build a core group of staff with CFSR expertise, ACF created the National Review Team, composed of central and regional office staff with additional training in and experience with the review process. In addition, to provide more technical assistance to the states, ACF reordered the priorities of the national resource centers to focus their efforts primarily on helping states with the review process. Like ACF, states also spent financial resources on the review. While some states did not track CFSR expenses--such as staff salaries, training, or administrative costs--of the 25 states that reported such information in our survey, the median expense to date was $60,550, although states reported spending as little as $1,092 and as much as $1,000,000 on the CFSR process.[Footnote 12] Although ACF officials told us that states can use Title IV-E funds to pay for some of their CFSR expenses, only one state official addressed the use of these funds in our survey, commenting that it was not until after the on-site review occurred that the state learned these funds could have been used to offset states' expenses. States also reported that they dedicated staff time to prepare for the statewide assessment and to conduct the on-site review, which sometimes had a negative impact on some staffs' regular duties. According to our survey, 45 states reported dedicating up to 200 full-time staff equivalents (FTE), with an average of 47 FTEs, to the statewide assessment process.[Footnote 13] Similarly, 42 states responded that they dedicated between 3 and 130 FTEs, with an average of 45 FTEs, to the on-site review process. For some caseworkers, dedicating time to the CFSR meant that they were unable or limited in their ability to manage their typical workload. For example, Wyoming caseworkers whose case files were selected for the on-site review told us that they needed to be available to answer reviewers' questions all day every day during the on-site review, which they said prevented them from conducting necessary child abuse investigations or home visits. Child welfare-related stakeholders--such as judges, lawyers, and foster parents--also contributed time to the CFSR. States and Child Welfare Experts Report That Several Data Improvements Could Enhance CFSR Reliability: State officials in the 5 states we visited, as well as child welfare experts, reported on several data improvements that could enhance the reliability of CFSR findings. In particular, they highlighted inaccuracies with the AFCARS and NCANDS data that are used for establishing the national standards and creating the statewide data profiles, which are then used to determine if states are in substantial conformity. These concerns echoed the findings of a prior GAO study on the reliability of these data sources, which found that states are concerned that the national standards used in the CFSR are based on unreliable information and should not be used as a basis for comparison and potential financial penalty.[Footnote 14] Furthermore, many states needed to resubmit their statewide data after finding errors in the data profiles ACF would have used to measure compliance with the national standards.[Footnote 15] According to our national survey, of the 37 states that reported on resubmitting data for the statewide data profile, 23 needed to resubmit their statewide data at least once, with one state needing to resubmit as many as five times to accurately reflect revised data. Four states reported in our survey that they did not resubmit their data profiles because they did not know they had this option or they did not have enough time to resubmit before the review. In addition to expressing these data concerns, child welfare experts as well as officials in all of the states we visited commented that existing practices that benefit children might conflict with actions needed to attain the national standards. For example, officials in New York said that they recently implemented an initiative to facilitate adoptions. Because these efforts focus on the backlog of children who have been in foster care for several years, New York officials predict that their performance on the national standard for adoption will be lower since many of the children in the initiative have already been in care for more than 2 years. Experts and officials from multiple states also commented that they believe the on-site review case sample of 50 cases is too small to provide an accurate picture of statewide performance, although ACF officials stated that the case sampling is supplemented with additional information.[Footnote 16] For example, Oklahoma officials we visited commented that they felt the case sample size was too small, especially since they annually assess more than 800 of their own cases--using a procedure that models the federal CFSR--and obtain higher performance results than the state received on its CFSR. Furthermore, because not every case in the states' sample is applicable to each item measured in the on-site review, we found that sometimes as few as 1 or 2 cases were being used to evaluate states' performance on an item. For example, Wyoming had only 2 on-site review cases applicable for the item measuring the length of time to achieve a permanency goal of adoption, but for 1 of these cases, reviewers determined that appropriate and timely efforts had not been taken to achieve finalized adoptions within 24 months, resulting in the item being assigned a rating of area needing improvement.[Footnote 17] While ACF officials acknowledged the insufficiency of the sample size,[Footnote 18] they contend that the case sampling is augmented by stakeholder interviews for all items and applicable statewide data for the five CFSR items with corresponding national standards, therefore providing sufficient evidence for determining states' conformity. All of the states we visited experienced discrepant findings between the aggregate data from the statewide assessment and the information obtained from the on-site review. We also found that in these 5 states, ACF had assigned an overall rating of area needing improvement for 10 of the 11 instances in which discrepancies occurred. ACF officials acknowledged the challenge of resolving data discrepancies, noting that such complications can delay the release of the final report and increase or decrease the number of items that states must address in their PIPs. While states have the opportunity to resolve discrepancies by submitting additional information explaining the discrepancy or by requesting an additional case review, only 1 state to date has decided to pursue the additional case review.[Footnote 19] Further, several state officials and experts also told us that additional data from the statewide assessments--or other data sources compiled by the states-- could bolster the evaluation of states' performance, but they found this information to be missing or insufficiently used in the final reports. For example, child welfare experts and state officials from California and New York--who are using alternative data sources to AFCARS and NCANDS, such as longitudinal data that track children's placements over time--told us that the inclusion of this more detailed information would provide a more accurate picture of states' performance nationwide. An HHS official told us that alternative data are used only to assess state performance in situations in which a state does not have NCANDS data, since states are not mandated to have these systems. Given their concerns with the data used in the review process, state officials in 4 of the 5 states believed that the threshold for achieving substantial conformity was difficult to achieve. While an ACF official told us that different thresholds for the national standards had been considered, ACF policy makers ultimately concluded that a threshold at the 75th percentile of the nationwide data would be used. ACF officials recognize that they have set a high standard. However, they believe it is attainable and supportive of their overall approach to move states to the standard through continuous improvement. Program Improvement Planning Under Way, but Uncertainties Challenge Plan Development, Implementation, and Monitoring: Forty-one states are engaged in program improvement planning, but many uncertainties, such as those related to federal guidance and monitoring and the availability of state resources, have affected the development, implementation, and funding of the PIPs. State PIPs include strategies such as revising or developing policies, training caseworkers, and engaging stakeholders, and ACF has issued regulations and guidance to help states develop and implement their plans. Nevertheless, states reported uncertainty about how to develop their PIPs and commented on the challenges they faced during implementation. For example, officials from 2 of the states we visited told us that ACF had rejected their PIPs before final approval, even though these officials said that the plans were based on examples of approved PIPs that regional officials had provided. Further, at least 9 of the 25 states responding to a question in our survey on PIP implementation indicated that insufficient time, funding, and staff, as well as high caseloads, were the greatest challenges they faced. As states progress in PIP implementation, some ACF officials expressed a need for more guidance on how to monitor state accomplishments, and both ACF and state officials were uncertain about how the estimated financial penalties would be applied if states fail to achieve the goals described in their plans. State Plans Include a Variety of Strategies to Address Identified Weaknesses: State plans include a variety of strategies to address weaknesses identified in the CFSR review process. However, because most states had not completed PIP implementation by the time of our analysis, the extent to which states have improved outcomes for children has not been determined.[Footnote 20] While state PIPs varied in their detail, design, and scope, according to our analysis of 31 available PIPs, these state plans have focused to some extent on revising or developing policies; reviewing and reporting on agency performance; improving information systems; and engaging stakeholders such as courts, advocates, foster parents, private providers, or sister agencies in the public sector.[Footnote 21] Table 1 shows the number of states that included each of the six categories and subcategories of strategies we developed for the purposes of this study. Table 1: Number of States Including Each of the PIP Strategy Categories Used in This Study: PIP strategy category: Policies and procedures; Description (number of states that included the strategy in their PIP): * Review, modify, or develop/implement any policy, procedure or case practice standard (31); * Enhance foster home/parent licensing standards (7); * Develop child and family assessment tools, such as protocols for risk/ safety determinations (28); * Identify and adopt any promising practices (19). PIP strategy category: Data collection and analysis; Description (number of states that included the strategy in their PIP): * Review and report on agency performance through self-assessments or internal audits/review (31); * Apply federal CFSR or similar process for internal statewide case reviews (16); * Improve information and data collection systems (31). PIP strategy category: Staff supports; Description (number of states that included the strategy in their PIP): * Train and develop caseworkers (through dissemination and training on policy or through revisions to overall curriculum) (30); * Assess and monitor staff responsibilities, skills, or performance (24); * Recruit additional staff/retain staff (14); * Lower caseloads (11); * Increase caseworker pay (1). PIP strategy category: Foster parent supports/services and resources for children and families; Description (number of states that included the strategy in their PIP): * Train and develop foster families'/ providers' skills and capacities (27); * Recruit and retain foster families (22); * Increase involvement of foster or birth families in case (18); * Expand service array for children and families (includes developing or enhancing transportation systems to transport siblings and parents for visits, creating one-stop centers for assistance, modifying visitation services, and providing any additional support services) (27); * Engage stakeholders such as courts, advocates, foster homes, private providers, or sister agencies in public sector, e.g., mental health (can include consultation, training, or formal partnering to improve services or placements) (31); * Create or improve monitoring of contracts with private providers to enhance service delivery (includes development of performance based or outcome-based contracts or other evaluations of provider performance) (25). PIP strategy category: State legislative supports; Description (number of states that included the strategy in their PIP): * State request for legislative action to support any of the above strategies (20). PIP strategy category: Federal technical assistance; Description (number of states that included the strategy in their PIP): * State request technical assistance from ACF or any resource center to support any of the above strategies (27). Source: GAO analysis. [End of table] Our analysis also showed that many states approached PIP development by building on state initiatives in place prior to the on-site review. Of the 42 surveyed states reporting in our survey on this topic, 30 said that their state identified strategies for the PIP by examining ongoing state initiatives. For example, local officials in New York City and state officials in California told us that state reform efforts--borne in part from legal settlements--have become the foundation for the PIP. State officials in California informed us that reform efforts initiated prior to the CFSR, such as implementing a new system for receiving and investigating reports of abuse and neglect and developing more early intervention programs, became integral elements in the PIP. Insufficient Guidance Hampered State Planning Efforts, but ACF Has Taken Steps to Clarify Expectations and Improve Technical Assistance: ACF has provided states with regulations and guidance to facilitate PIP development, but some states believe the requirements have been unclear. For example, several states commented in our survey that multiple aspects of the PIP approval process were unclear, such as how much detail and specificity the agency expects the plan to include; what type of feedback states could expect to receive; when states could expect to receive such feedback; and whether a specific format was required. Officials in the states we visited echoed survey respondents' concerns with officials from 3 of the 5 states informing us that ACF had given states different instructions regarding acceptable PIP format and content. For example, California and Florida officials told us that their program improvement plans had been rejected prior to final approval, even though they were based on examples of approved plans that regional officials had provided. In addition, California officials told us that they did not originally know how much detail the regional office expected in the PIP and believed that the level of detail the regional office staff ultimately required was too high. Specifically, officials in California said that the version of their plan that the region accepted included 2,932 action steps--a number these officials believe is too high given their state's limited resources and the 2- year time frame to implement the PIP. ACF officials have undertaken several steps to clarify their expectations for states and to improve technical assistance. For example, in 2002, 2 years after ACF released the CFSR regulations and a procedures manual, ACF offered states additional guidance and provided a matrix format to help state officials prepare their plans. ACF officials told us the agency sends a team of staff from ACF and resource centers to the state to provide intensive on-site technical assistance, when it determines that a state is slow in developing its PIP. Further, ACF has sent resource center staff to states to provide training almost immediately after the completion of the on-site review to encourage state officials to begin PIP development before the final report is released. Our survey results indicate that increasing numbers of states are developing their PIPs early in the CFSR process, which may reflect ACF's emphasis on PIP development. According to our analysis, of the 18 states reviewed in 2001, only 2 started developing their PIPs before or during the statewide assessment phase. Among states reviewed in 2003, this share increased to 5 of 9. Evidence suggests that lengthy time frames for PIP approval have not necessarily delayed PIP implementation, and ACF has made efforts to reduce the time the agency takes to approve states' PIPs. For example, officials in 3 of the 5 states we visited told us they began implementing new action steps before ACF officially approved their plans because many of the actions in their PIPs were already under way. In addition, according to our survey, of the 28 states reporting on this topic, 24 reported that they had started implementing their PIP before ACF approved it. Further, our analysis shows that the length of time between the PIP due date, which statute sets at 90 days after the release of the final CFSR report, and final ACF PIP approval has ranged considerably--from 45 to 349 business days. For almost half of the plans, ACF's approval occurred 91 to 179 business days after the PIP was due. Our analysis indicated that ACF has recently reduced the time lapse by 46 business days. This shorter time lapse for PIP approval may be due, in part, to the ACF's emphasis on PIP development. According to one official, ACF has directed states to concentrate on submitting a plan that can be quickly approved. Another ACF official added that because of ACF's assistance with PIP development, states are now submitting higher-quality PIPs that require fewer revisions. State and Federal Uncertainties Cloud PIP Implementation and Monitoring: Program improvement planning has been ongoing, but uncertainties have made it difficult for states to implement their plans and ACF to monitor state performance. Such uncertainties include not knowing whether state resources are adequate to implement the plans and how best to monitor state reforms. In answering a survey question about PIP implementation challenges, a number of states identified insufficient funding, staff, and time--as well as high caseloads--as their greatest obstacles. Figure 2 depicts these results. Figure 2: Most Common Challenges Affecting States' PIP Implementation: [See PDF for image] Note: This is based on responses from 25 states. The results reported in the figure are a sum of the states reporting that the issue was a challenge to PIP implementation to a very great extent, great extent, moderate extent, or some/little extent. States not included answered no extent, no basis to judge, or not applicable. [End of figure] One official from Pennsylvania commented that because of the state's budget shortfall, no additional funds were available for the state to implement its improvement plan, so most counties must improve outcomes with little or no additional resources. A Massachusetts official reported that fiscal problems in his state likely would lead the state to lay off attorneys and caseworkers and to cut funding for family support programs. While state officials acknowledged that they do not have specific estimates of PIP implementation expenses because they have not tracked this information in their state financial systems, many states indicated that to cope with financial difficulties, they had to be creative and use resources more efficiently to fund PIP strategies. Of the 26 states responding to a question in our survey on PIP financing, 12 said that they were financing the PIP strategies by redistributing current funding, and 7 said that they were using no-cost methods. In an example of the latter, Oklahoma officials reported pursuing in-kind donations from a greeting card company so that they could send thank-you notes to foster parents, believing this could increase foster parent retention and engagement. Aside from funding challenges, states also reported that PIP implementation has been affected by staff workloads, but these comments were mixed. In Wyoming, for example, caseworkers told us that their high caseloads would prevent them from implementing many of the positive action steps included in their improvement plan. In contrast, Oklahoma caseworkers told us that the improvement plan priorities in their state--such as finding permanent homes for children--have helped them become more motivated, more organized, and more effective with time management. ACF officials expressed uncertainty about how best to monitor states' progress and apply estimated financial penalties when progress was slow or absent, and 3 of the 5 states we visited reported frustration with the limited guidance ACF had provided on the PIPs quarterly reporting process. For example, 4 regional offices told us that they did not have enough guidance on or experience with evaluating state quarterly reports. Some regional offices told us they require states to submit evidence of each PIP action step's completion, such as training curricula or revised policies, but one ACF official acknowledged that this is not yet standard procedure, although the agency is considering efforts to make the quarterly report submission procedures more uniform. Moreover, ACF staff from 1 region told us that because PIP monitoring varies by region, they were concerned about enforcing penalties. Shortly before California's quarterly report was due, state officials told us they still did not know how much detail to provide; how to demonstrate whether they had completed certain activities; or what would happen if they did not reach the level of improvement specified in the plan. Based on data from the states that have been reviewed to date, the estimated financial penalties range from a total of $91,492 for North Dakota to $18,244,430 for California, but the impact of these potential penalties remains unclear. While ACF staff from most regional offices told us that potential financial penalties are not the driving force behind state reform efforts, some contend that the estimated penalties affect how aggressively states pursue reform in their PIPs. For example, regional office staff noted that 1 state's separate strategic plan included more aggressive action steps than those in its PIP because the state did not want to be liable for penalties if it did not meet its benchmarks for improvement. State officials also had mixed responses as to how the financial penalties would affect PIP implementation. An official in Wyoming said that incurring the penalties was equivalent to shutting down social service operations in 1 local office for a month, while other officials in the same state thought it would cost more to implement PIP strategies than it would to incur financial penalties if benchmarks were unmet. Nevertheless, these officials also said that while penalties are a consideration, they have used the CFSR as an opportunity to provide better services. One official in another state agreed that it would cost more to implement the PIP than to face financial penalties, but this official was emphatic in the state's commitment to program improvement. ACF's Focus Rests Almost Exclusively on Implementing the CFSR: To implement the CFSRs, ACF has focused its activities almost entirely on the CFSR review process, and regional staff report limitations in providing assistance to states in helping them to meet key federal goals. ACF officials told us the CFSR has become the agency's primary mechanism for monitoring states and facilitating program improvement, but they acknowledged that regional office staff might not have realized the full utility of the CFSR as a tool to integrate all existing training and technical assistance efforts. Further, according to ACF officials, meetings to discuss a new system of training and technical assistance are ongoing, though recommendations were not available at the time of publication of our April 2004 report. Levels of resource center funding, the scope and objectives of the resource centers' work, and the contractors who operate the resource centers are all subject to change before the current cooperative agreements expire at the close of fiscal year 2004. ACF officials told us that the learning opportunities in the Children's Bureau are intentionally targeted at the CFSR, but staff in 3 regions told us that this training should cover a wider range of subjects-- including topics outside of the CFSR process--so that regional officials could better meet states' needs. All 18 of the courses that ACF has provided to its staff since 2001 have focused on such topics as writing final CFSR reports and using data for program improvement, and while ACF officials in the central office said that the course selection reflects both the agency's prioritization of the CFSR process and staff needs, our interviews with regional staff suggest that some of them wish to obtain additional non-CFSR training. In addition, although ACF organizes biennial conferences for state and federal child welfare officials, staff from 5 regions told us that they wanted more substantive interaction with their ACF colleagues, such as networking at conferences, to increase their overall child welfare expertise. Further, staff from 6 of the 10 regions told us that their participation in conferences is limited because of funding constraints. ACF staff in all 10 regions provide ongoing assistance or ad hoc counseling to states, either through phone, e-mail, or on-site support, but staff from 6 regions told us they would like to conduct site visits with states more regularly to improve their relationships with state officials and provide more targeted assistance. Further, staff in 4 regions felt their travel funds were constrained and explained that they try to stretch their travel dollars by addressing states' non-CFSR needs, such as court improvements, during CFSR-related visits. While an ACF senior official from the central office confirmed that CFSR-related travel constituted 60 percent of its 2002 child welfare-monitoring budget, this official added that CFSR spending represents an infusion of funding rather than a reprioritization of existing dollars, and stated that regional administrators have discretion over how the funds are allocated within their regions. In addition, the same official stated that he knew of no instance in which a region requested more money for travel than it received. Concerns from state officials in all 5 of the states we visited echoed those of regional office staff and confirmed the need for improvements to the overall training and technical assistance structure. For example, state officials in New York and Wyoming commented that ACF staff from their respective regional offices did not have sufficient time to spend with them on CFSR matters because regional staff were simultaneously occupied conducting reviews in other states. However, our survey results revealed that states reviewed in 2003 had much higher levels of satisfaction with regional office assistance than those states reviewed in 2001, which suggests improvements to regional office training and technical assistance as the process evolved. Concluding Observations: ACF and the states have devoted considerable resources to the CFSR process, but to date, no state has passed the threshold for substantial conformity on all CFSR measures, and concerns remain regarding the validity of some data sources and the limited use of all available information to determine substantial conformity. The majority of states surveyed agreed that CFSR results are similar to their own evaluation of areas needing improvement. However, without using more reliable data--and in some cases, additional data from state self-assessments-- to determine substantial conformity, ACF may be over-or under- estimating the extent to which states are actually meeting the needs of the children and families in their care. These over-or under-estimates can, in turn, affect the scope and content of the PIPs that states must develop in response. In addition, the PIP development, approval, and monitoring processes remain unclear to some, potentially reducing states' credibility with their stakeholders and straining the federal/state partnership. Similarly, regional officials are unclear as to how they can accomplish their various training and technical assistance responsibilities, including the CFSR. Without clear guidance on how to systematically prepare and monitor PIP-related documents, and how regional officials can integrate their many oversight responsibilities, ACF has left state officials unsure of how their progress over time will be judged and potentially complicated its own monitoring efforts. To ensure that ACF uses the best available data in measuring state performance, we recommended in our April 2004 report that the Secretary of HHS expand the use of additional data states may provide in their statewide assessments and consider alternative data sources when available, such as longitudinal data that track children's placements over time, before making final CFSR determinations. In addition, to ensure that ACF regional offices and states fully understand the PIP development, approval, and monitoring processes, and that regional offices fully understand ACF's prioritization of the CFSR as the primary mechanism for child welfare oversight, we recommended that the Secretary of HHS issue clarifying guidance on the PIP process and evaluate states' and regional offices' adherence to this instruction and provide guidance to regional offices explaining how to better integrate the many training and technical assistance activities for which they are responsible, such as participation in state planning meetings and the provision of counsel to states on various topics, with their new CFSR responsibilities. In response to the first recommendation, HHS acknowledged that the CFSR is a new process that continues to evolve, and also noted several steps it has taken to address the data quality concerns we raise in our report. We believe that our findings from the April 2004 report, as well as a previous report on child welfare data and states' information systems, fully address HHS's initial actions, as well as the substantial resources the agency has already dedicated to the review process. However, to improve its oversight of state performance, our recommendation was meant to encourage HHS to take additional actions to improve its use of data in conducting these reviews. In response to the second recommendation, HHS said that it has continued to provide technical assistance and training to states and regional offices, when appropriate. HHS noted that it is committed to continually assessing and addressing training and technical assistance needs. In this context, our recommendation was intended to encourage HHS to enhance existing training efforts and focus both on state and on regional officials' understanding of how to incorporate the CFSR process into their overall improvement and oversight efforts. Mr. Chairman, this concludes my prepared statement. I would be pleased to respond to any questions that you or other members of the subcommittee may have. [End of section] Appendix I: GAO Contacts Acknowledgments: GAO Contacts: For further contacts regarding this testimony, please call Cornelia M. Ashby at (202) 512-8403. Individuals making key contributions to this testimony include Diana Pietrowiak and Joy Gambino. [End of section] Related GAO Products: D.C. Family Court: Operations and Case Management Have Improved, but Critical Issues Remain. GAO-04-685T. Washington, D.C.: April 23, 2004. Child and Family Services Reviews: Better Use of Data and Improved Guidance Could Enhance HHS's Oversight of State Performance. GAO-04- 333 Washington, D.C.: April 20, 2004. Child Welfare: Improved Federal Oversight Could Assist States in Overcoming Key Challenges. GAO-04-418T. Washington, D.C.: January 28, 2004. D.C. Family Court: Progress Has Been Made in Implementing Its Transition. GAO-04-234. Washington, D.C.: January 6, 2004. Child Welfare: States Face Challenges in Developing Information Systems and Reporting Reliable Child Welfare Data. GAO-04-267T. Washington, D.C.: November 19, 2003. Child Welfare: Enhanced Federal Oversight of Title IV-B Could Provide States Additional Information to Improve Services. GAO-03-956. Washington, D.C.: September 12, 2003. Child Welfare: Most States Are Developing Statewide Information Systems, but the Reliability of Child Welfare Data Could be Improved. GAO-03-809. Washington, D.C.: July 31, 2003. D.C. Child and Family Services: Key Issues Affecting the Management of Its Foster Care Cases. GAO-03-758T. Washington, D.C.: May 16, 2003. Child Welfare and Juvenile Justice: Federal Agencies Could Play a Stronger Role in Helping States Reduce the Number of Children Placed Solely to Obtain Mental Health Services. GAO-03-397. Washington, D.C.: April 21, 2003. Foster Care: States Focusing on Finding Permanent Homes for Children, but Long-Standing Barriers Remain. GAO-03-626T. Washington, D.C.: April 8, 2003. Child Welfare: HHS Could Play a Greater Role in Helping Child Welfare Agencies Recruit and Retain Staff. GAO-03-357. Washington, D.C.: March 31, 2003. Foster Care: Recent Legislation Helps States Focus on Finding Permanent Homes for Children, but Long-Standing Barriers Remain. GAO-02-585. Washington, D.C.: June 28, 2002. District of Columbia Child Welfare: Long-Term Challenges to Ensuring Children's Well-Being. GAO-01-191. Washington, D.C.: December 29, 2000. Child Welfare: New Financing and Service Strategies Hold Promise, but Effects Unknown. GAO/T-HEHS-00-158. Washington, D.C.: July 20, 2000. Foster Care: States' Early Experiences Implementing the Adoption and Safe Families Act. GAO/HEHS-00-1. Washington, D.C.: December 22, 1999. Foster Care: HHS Could Better Facilitate the Interjurisdictional Adoption Process. GAO/HEHS-00-12. Washington, D.C.: November 19, 1999. Foster Care: Effectiveness of Independent Living Services Unknown. GAO/ HEHS-00-13. Washington, D.C.: November 10, 1999. Foster Care: Kinship Care Quality and Permanency Issues. GAO/HEHS-99- 32. Washington, D.C.: May 6, 1999. Juvenile Courts: Reforms Aim to Better Serve Maltreated Children. GAO/ HEHS-99-13. Washington, D.C.: January 11, 1999. Child Welfare: Early Experiences Implementing a Managed Care Approach. GAO/HEHS-99-8. Washington, D.C.: October 21, 1998. Foster Care: Agencies Face Challenges Securing Stable Homes for Children of Substance Abusers. GAO/HEHS-98-182. Washington, D.C.: September 30, 1998. FOOTNOTES [1] The CFSR measures state performance on 45 performance items, which correspond to 7 outcomes and 7 systemic factors. The outcomes relate to children's safety, permanency, and well-being, and the systemic factors address state agency management and responsiveness to the community. Six national standards, as reported in the Adoption and Foster Care Analysis and Reporting System (AFCARS) and the National Child Abuse and Neglect Data System (NCANDS), apply to 5 of the 45 items. Three of these standards are based on the 75th percentile of all states' performance--adoption; stability of foster care placements; and length of time to achieve reunification, guardianship, or permanent placement with relatives--because a higher incidence is desirable. However, the remaining three standards--recurrence of maltreatment, incidence of child abuse/neglect in foster care, and foster care re-entries--are based on the 25th percentile of state performance because lower incidence is a desired outcome for these measures. [2] U.S. General Accounting Office, Child and Family Services Reviews: Better Use of Data and Improved Guidance Could Enhance HHS's Oversight of State Performance (GAO-04- 333, April 20, 2004). [3] We achieved a 98 percent response rate from this survey; Puerto Rico was the only non-respondent. [4] The 10 states participating in our phone follow-up surveys were Arkansas, Iowa, Kansas, Mississippi, North Dakota, New Jersey, Pennsylvania, Rhode Island, Utah, and West Virginia. [5] ACF has established cooperative agreements with 10 national resource centers to help states implement federal legislation intended to ensure the safety, permanency, and well-being of children and families. ACF sets the resource centers' areas of focus, and although each center has a different area of expertise, such as organizational improvement or information technology, all of them conduct needs assessments, sponsor national conference calls with states, collaborate with other resource centers and agencies, and provide on-site training and technical assistance to states. [6] Title IV-B of the Social Security Act, consisting of two subparts, is the primary source of federal funding for services to help families address problems that lead to child abuse and neglect and to prevent the unnecessary separation of children from their families. Funding under Title IV-E of the Social Security Act is used primarily to pay for the room and board of children in foster care. [7] The term stakeholder refers to two groups: (1) agency stakeholders, such as judges or advocates, whose responsibilities are closely related to the work of the child welfare agency and who can comment on the agency's overall performance on outcomes and systemic factors, and (2) case-specific stakeholders, such as parents, caseworkers, children, or others who are interviewed to provide first-hand information that supplements reviewers' assessment of paper or electronic case files. [8] States began voluntarily reporting to NCANDS in 1990, and in 1995 started reporting to AFCARS on the demographic characteristics of adoptive and foster children and their parents, as well as foster children's type of placement and permanency goals. We recently issued a report on states' child welfare information systems and the reliability of child welfare data. U.S. General Accounting Office, Child Welfare: Most States Are Developing Statewide Information Systems, but the Reliability of Child Welfare Data Could Be Improved, GAO-03-809 (Washington, D.C.: July 31, 2003). [9] States achieve substantial conformity on outcomes and systemic factors when at least 90 percent of applicable cases are substantially achieved; stakeholder interviews confirm that state plan and other program requirements are in place and functioning as described in the applicable regulations or statute; and performance on items with national standards, where applicable, meets the applicable threshold. [10] The formula for calculating penalties is based in part on each state's allocation of federal child welfare funds from Titles IV-B and IV-E and the number of outcomes and systemic factors for which substantial conformity has not been achieved. [11] California and Puerto Rico were determined to be in substantial conformity on 2 outcomes and systemic factors, while North Dakota achieved substantial conformity on 9. [12] These values are state-reported and reflect officials' estimates of costs associated with all CFSR-related activities except those incurred during PIP implementation. In reporting on their expenses, states were instructed to include the value of training, travel, infrastructure, technology, food, administrative supplies, and any other expenses associated with the CFSR process. States were also asked to provide supporting documentation for this particular question, but most states were unable to provide documentation. Many states reported that they did not track CFSR-related expenses. The 25 states that did provide estimates were in different phases of the CFSR. [13] The number of FTEs participating in each phase of the CFSR is state-reported. While states were not given specific instructions for how to calculate FTEs, they were asked to report only on the phases of the CFSR that they had started or completed. Therefore, states' responses varied depending on the phase of the CFSR process they were in and the methods they used to calculate FTEs. [14] GAO-03-809. [15] ACF provides states with their statewide data about 6 months prior to the on-site review, during which time states are allowed to make corrections to the data and resubmit the updated data so it can be used when determining state conformity with CFSR measures. [16] According to our calculations--which assumed that the attribute of interest occurred in about 50 percent of the cases--a sample size of 50 would produce percentage estimates with a 95 percent margin of error of approximately plus or minus 14 percentage points. This level of variability is a limitation when attempting to interpret estimates based on this sample size. [17] Because 1 of the 2 cases applicable to the adoption measure was assigned a rating of area needing improvement, 50 percent of the cases for this item were assigned a rating of area needing improvement. As a result, the item was given an overall rating of area needing improvement since both cases would have needed to be assigned a rating of strength for this item to meet the 85 percent threshold necessary to assign an overall rating of strength. [18] An ACF statistician also confirmed that the CFSR sample is too small to generalize to the states' populations and that the three sites, from which cases are selected, also are not representative. [19] Virginia requested an additional case review to resolve a discrepancy between the statewide data and on-site review findings for the item measuring the state's performance on foster care re-entries. According to an ACF regional official, the state met the national standard for this item but the case review findings showed the state did not meet the threshold for this measure. At the time of publication of our April 2004 report, ACF and the state were still finalizing plans to conduct the additional case review, and until the review is completed, the state cannot receive its final report. [20] As we reported in our April 2004 report, only Delaware and North Carolina had completed the 2-year term of their PIPs, and ACF was still analyzing the states' progress and had not determined if there has been overall improvement or if ACF will apply financial penalties. [21] Although 41 states were developing or implementing PIPs when our April 2004 report was published, we reviewed the 31 available PIPs that ACF had approved as of January 1, 2004.

The Justia Government Accountability Office site republishes public reports retrieved from the U.S. GAO These reports should not be considered official, and do not necessarily reflect the views of Justia.