PITTSBURGH (AP) 鈥 For the two weeks that the Hackneys鈥 baby girl lay in a Pittsburgh hospital bed weak from dehydration, her parents rarely left her side, sometimes sleeping on the fold-out sofa in the room.
They stayed with their daughter around the clock when she was moved to a rehab center to regain her strength. Finally, the 8-month-old stopped batting away her bottles and started putting on weight again.
鈥淪he was doing well and we started to ask when can she go home,鈥 Lauren Hackney said. 鈥淎nd then from that moment on, at the time, they completely stonewalled us and never said anything.鈥
The couple was stunned when child welfare officials showed up, told them they were negligent and took their daughter away.
鈥淭hey had custody papers and they took her right there and then,鈥 Lauren Hackney recalled. 鈥淎nd we started crying.鈥
More than a year later, their daughter, now 2, remains in foster care. The Hackneys, who have developmental disabilities, are struggling to understand how taking their daughter to the hospital when she refused to eat could be seen as so neglectful that she鈥檇 need to be taken from her home.
They wonder if an artificial intelligence tool that the Allegheny County Department of Human Services uses to predict which children could be at risk of harm singled them out because of their disabilities.
The U.S. Justice Department is asking the same question. The agency is investigating the county鈥檚 child welfare system to determine whether its use of the influential algorithm discriminates against people with disabilities or other protected groups, The Associated Press has learned. Later this month, federal civil rights attorneys will interview the Hackneys and Andrew Hackney鈥檚 mother, Cynde Hackney-Fierro, the grandmother said.
Lauren Hackney has attention-deficit hyperactivity disorder that affects her memory, and her husband, Andrew, has a comprehension disorder and nerve damage from a stroke suffered in his 20s. Their baby girl was just 7 months old when she began refusing to drink her bottles. Facing a nationwide shortage of formula, they traveled from Pennsylvania to West Virginia looking for some and were forced to change brands. The baby didn鈥檛 seem to like it.
Her pediatrician first reassured them that babies sometimes can be fickle with feeding and offered ideas to help her get back her appetite, they said.
When she grew lethargic days later, they said, the same doctor told them to take her to the emergency room. The Hackneys believe medical staff alerted child protective services after they showed up with a baby who was dehydrated and malnourished.
That鈥檚 when they believe their information was fed into the Allegheny Family Screening Tool, which county officials say is standard procedure for neglect allegations. Soon, a social worker appeared to question them, and their daughter was sent to foster care.
Over the past six years, Allegheny County has served as a real-world laboratory for testing AI-driven child welfare tools that crunch reams of data about local families to try to predict which children are likely to face danger in their homes. Today, child welfare agencies in at least 26 states and Washington, D.C., have considered using algorithmic tools, and jurisdictions in at least 11 have deployed them, according to the American Civil Liberties Union.
The Hackneys鈥 story 鈥 based on interviews, internal emails and legal documents 鈥 illustrates the opacity surrounding these algorithms. Even as they fight to regain custody of their daughter, they can鈥檛 question the 鈥渞isk score鈥 Allegheny County鈥檚 tool may have assigned to her case because officials won鈥檛 disclose it to them. And neither the county nor the people who built the tool have ever explained which variables may have been used to measure the Hackneys鈥 abilities as parents.
鈥淚t鈥檚 like you have an issue with someone who has a disability,鈥 Andrew Hackney said in an interview from their apartment in suburban Pittsburgh. 鈥淚n that case 鈥 you probably end up going after everyone who has kids and has a disability.鈥
As part of a yearlong investigation, the AP obtained the fields of data underpinning several algorithms deployed by child welfare agencies, including some marked 鈥淐ONFIDENTIAL,鈥 offering rare insight into the mechanics driving these emerging technologies. Among the factors they have used to calculate a family鈥檚 risk, whether outright or by proxy: race, poverty rates, disability status and family size. They include whether a mother smoked before she was pregnant and whether a family had previous child abuse or neglect complaints.
What they measure matters. found that when Allegheny's algorithm flagged people who accessed county services for mental health and other behavioral health programs, that could add up to three points to a child鈥檚 risk score, a significant increase on a scale of 20.
Allegheny County spokesman Mark Bertolet declined to address the Hackney case and did not answer detailed questions about the status of the federal probe or critiques of the data powering the tool, including by the ACLU.
鈥淎s a matter of policy, we do not comment on lawsuits or legal matters,鈥 Bertolet said in an email.
Justice Department spokeswoman Aryele Bradford declined to comment.
NOT MAGIC
Child welfare algorithms plug vast amounts of public data about local families into complex statistical models to calculate what they call a risk score. The number that鈥檚 generated is then used to advise social workers as they decide which families should be investigated, or which families need additional attention 鈥 a weighty decision that can sometimes mean life or death.
A number of local leaders have tapped into AI technology while under pressure to make systemic changes, such as in Oregon during a foster care crisis and in Los Angeles County after a series of high-profile child deaths in one of the nation鈥檚 largest county child welfare systems.
LA County鈥檚 Department of Children and Family Services Director Brandon Nichols says algorithms can help identify high-risk families and improve outcomes in a deeply strained system. Yet he could not explain how the screening tool his agency uses works.
鈥淲e鈥檙e sort of the social work side of the house, not the IT side of the house,鈥 Nichols said in an interview. 鈥淗ow the algorithm functions, in some ways is, I don鈥檛 want to say is magic to us, but it鈥檚 beyond our expertise and experience.鈥
Nichols and officials at two other child welfare agencies referred detailed questions about their AI tools to the outside developers who created them.
In Larimer County, Colorado, one official acknowledged she didn鈥檛 know what variables were used to assess local families.
鈥淭he variables and weights used by the Larimer Decision Aide Tool are part of the code developed by Auckland and thus we do not have this level of detail,鈥 Jill Maasch, a Larimer County Human Services spokeswoman, said in an email, referring to the developers.
In Pennsylvania, California and Colorado, county officials have opened up their data systems to the two academic developers who select data points to build their algorithms. Rhema Vaithianathan, a professor of health economics at New Zealand鈥檚 Auckland University of Technology, and Emily Putnam-Hornstein, a professor at the University of North Carolina at Chapel Hill鈥檚 School of Social Work, said in an email that their work is transparent and that they make their computer models public.
鈥淚n each jurisdiction in which a model has been fully implemented we have released a description of fields that were used to build the tool, along with information as to the methods used,鈥 they said by email.
A on the Allegheny County website includes pages of coded variables and statistical calculations.
Vaithianathan and Putnam-Hornstein鈥檚 work has been hailed in reports published by UNICEF and the Biden administration alike for devising computer models that promise to lighten caseworkers鈥 loads by drawing from a set of simple factors. They have described using such tools as a moral imperative, insisting that child welfare officials should draw from all data at their disposal to make sure children aren鈥檛 maltreated.
Through tracking their work across the country, however, the AP found their tools can set families up for separation by rating their risk based on personal characteristics they cannot change or control, such as race or disability, rather than just their actions as parents.
In Allegheny County, a sprawling county of 1.2 million near the Ohio border, the algorithm has accessed an array of external data, including jail, juvenile probation, Medicaid, welfare, health and birth records, all held in a vast countywide 鈥渄ata warehouse.鈥 The tool uses that information to predict the risk that a child will be placed in foster care two years after a family is first investigated.
County officials have told the AP they鈥檙e proud of their cutting-edge approach, and even expanded their work to build another algorithm focused on newborns. They have said they monitor their risk scoring tool closely and update it over time, including removing variables such as welfare benefits and birth records.
Vaithianathan and Putnam-Hornstein declined the AP鈥檚 repeated interview requests to discuss how they choose the specific data that powers their models. But in a 2017 report, they detailed the methods used to build the first version of Allegheny鈥檚 tool, including a footnote that described a statistical cutoff as 鈥渞ather arbitrary but based on trial and error.鈥
鈥淭his footnote refers to our exploration of more than 800 features from Allegheny鈥檚 data warehouse more than five years ago,鈥 the developers said by email.
That approach is borne out in their design choices, which differ from county to county.
In the same 2017 report, the developers acknowledged that using race data didn鈥檛 substantively improve the model鈥檚 accuracy, but they continued to study it in Douglas County, Colorado, though they ultimately opted against including it in that model. To address community concerns that a tool could harden racial bias in Los Angeles County, the developers excluded people鈥檚 criminal history, ZIP code and geographic indicators, but have continued to use those data points in the Pittsburgh area.
When asked about the inconsistencies, the developers pointed to their published methodology documents.
鈥淲e detail various metrics used to assess accuracy 鈥 while also detailing 鈥榚xternal validations,鈥欌 the developers said via email.
When Oregon鈥檚 Department of Human Services built an algorithm inspired by Allegheny鈥檚, it factored in a child鈥檚 race as it predicted a family鈥檚 risk, and also applied a 鈥渇airness correction鈥 to mitigate racial bias. Last June, the due to equity concerns after an AP investigation in April revealed potential racial bias in such tools.
Justice Department last fall when federal civil rights attorneys started inquiring about additional discrimination concerns in Allegheny鈥檚 tool, three sources told the AP. They spoke on the condition of anonymity, saying the Justice Department asked them not to discuss the confidential conversations. Two said they also feared professional retaliation.
IQ TESTS, PARENTING CLASS
With no answers on when they could get their daughter home, the Hackneys鈥 lawyer in October filed a federal civil rights complaint on their behalf that questioned how the screening tool was used in their case.
Over time, Allegheny鈥檚 tool has tracked if members of the family have diagnoses for schizophrenia or mood disorders. It鈥檚 also measured if parents or other children in the household have disabilities, by noting whether any family members received Supplemental Security Income, a federal benefit for people with disabilities. The county said that it factors in SSI payments in part because children with disabilities are more likely to be abused or neglected.
The county also said disabilities-aligned data can be 鈥減redictive of the outcomes鈥 and it 鈥渟hould come as no surprise that parents with disabilities 鈥 may also have a need for additional supports and services.鈥 In an emailed statement, the county added that elsewhere in the country, social workers also draw on data about mental health and other conditions that may affect a parent鈥檚 ability to safely care for a child.
The Hackneys have been ordered to take parenting classes and say they have been taxed by all of the child welfare system鈥檚 demands, including IQ tests and downtown court hearings.
People with disabilities are overrepresented in the child welfare system, yet there鈥檚 no evidence that they harm their children at higher rates, said Traci LaLiberte, a University of Minnesota expert on child welfare and disabilities.
Including data points related to disabilities in an algorithm is problematic because it perpetuates historic biases in the system and it focuses on people鈥檚 physiological traits rather than behavior that social workers are brought in to address, LaLiberte said.
The Los Angeles tool weighs if any children in the family have ever gotten special education services, have had prior developmental or mental health referrals or used drugs to treat mental health.
鈥淭his is not unique to caseworkers who use this tool; it is common for caseworkers to consider these factors when determining possible supports and services,鈥 the developers said by email.
Before algorithms were in use, the child welfare system had long distrusted parents with disabilities. Into the 1970s, they were regularly sterilized and institutionalized, LaLiberte said. A landmark federal report in 2012 noted parents with psychiatric or intellectual disabilities lost custody of their children as much as 80 percent of the time.
Across the U.S., it鈥檚 extremely rare for any child welfare agencies to require disabilities training for social workers, LaLiberte鈥檚 research has found. The result: Parents with disabilities are often judged by a system that doesn鈥檛 understand how to assess their capacity as caregivers, she said.
The Hackneys experienced this firsthand. When a social worker asked Andrew Hackney how often he fed the baby, he answered literally: two times a day. The worker seemed appalled, he said, and scolded him, saying babies must eat more frequently. He struggled to explain that the girl鈥檚 mother, grandmother and aunt also took turns feeding her each day.
FOREVER FLAGGED
Officials in Allegheny County have said that building AI into their processes helps them 鈥渕ake decisions based on as much information as possible,鈥 and noted that the algorithm merely harnesses data social workers can already access.
That can include decades-old records. The Pittsburgh-area tool has tracked whether parents were ever on public benefits or had a history with the criminal justice system 鈥 even if they were minors at the time, or if it never resulted in charges or convictions.
The AP found those design choices can stack the deck against people who grew up in poverty, hardening historical inequities that persist in the data, or against people with records in the juvenile or criminal justice systems, long after society has granted redemption. And critics say that algorithms can create a self-fulfilling prophecy by influencing which families are targeted in the first place.
Allegheny County provided researchers at the ACLU and the nonprofit Human Rights Data Analysis Group three months of data from 2021, when the tool was live, allowing the team to audit how the algorithm had actually affected families鈥 scores.
鈥淭hese predictors have the effect of casting permanent suspicion and offer no means of recourse for families marked by these indicators,鈥 the researchers found. 鈥淭hey are forever seen as riskier to their children.鈥
As child welfare algorithms become more common, parents who have experienced social workers鈥 scrutiny fear the models won鈥檛 let them escape their pasts, no matter how old or irrelevant their previous scrapes with the system may have been.
Charity Chandler-Cole, who serves on the Los Angeles County Commission for Children and Families, is one of them. She landed in foster care as a teen after being arrested for shoplifting underwear for her younger sister. Then as an adult, she said, social workers once showed up at her apartment after someone spuriously reported that a grand piano was thrown at her nephew who was living at her home 鈥 even though they didn鈥檛 own such an instrument.
The local algorithm could tag her for her prior experiences in foster care and juvenile probation, as well as the unfounded child abuse allegation, Chandler-Cole says. She wonders if AI could also properly assess that she was quickly cleared of any maltreatment concerns, or that her nonviolent offense as a teen was legally expunged.
鈥淎 lot of these reports lack common sense,鈥 said Chandler-Cole, now the mother of four and CEO of an organization that works with the court system to help children in foster care. 鈥淵ou are automatically putting us in these spaces to be judged with these labels. It just perpetuates additional harm.鈥
Chandler-Cole鈥檚 fellow commissioner Wendy Garen, by contrast, argues 鈥渕ore is better鈥 and that by drawing on all available data, risk scoring tools can help make the agency鈥檚 work more thorough and effective.
GLOBAL INFLUENCE
Even as their models have come under scrutiny for their accuracy and fairness, the developers have started new projects with child welfare agencies in Northampton County, Pennsylvania, and Arapahoe County, Colorado. The states of California and Pennsylvania, as well as New Zealand and Chile, have also asked them to do preliminary work.
And as word of their methods has spread in recent years, Vaithianathan has given lectures highlighting screening tools in Colombia and Australia. She also recently advised researchers in Denmark and officials in the United Arab Emirates on how to use technology to target child services.
鈥淩hema is one of the world leaders and her research can help to shape the debate in Denmark,鈥 a Danish researcher said on LinkedIn last year, regarding Vaithianathan鈥檚 advisory role related to a local child welfare tool that was being piloted.
Last year, the U.S. Department of Health and Human Services funded a national study, co-authored by Vaithianathan and Putnam-Hornstein, that concluded that their overall approach in Allegheny could be a model for other places.
HHS鈥 Administration for Children and Families spokeswoman Debra Johnson declined to say whether the Justice Department鈥檚 probe would influence her agency鈥檚 future support for an AI-driven approach to child welfare.
Especially as budgets tighten, cash-strapped agencies are desperate to find more efficient ways for social workers to focus on children who truly need protection. At a 2021 panel, Putnam-Hornstein acknowledged that 鈥渢he overall screen-in rate remained totally flat鈥 in Allegheny since their tool had been implemented.
Meanwhile, foster care and the separation of families can have lifelong developmental consequences for the child.
A 2012 HHS study found 95% of babies who are reported to child welfare agencies go through more than one caregiver and household change during their time in foster care, instability that researchers noted can itself be a form of trauma.
The Hackneys鈥 daughter already has been placed in two foster homes and has now spent more than half of her short life away from her parents as they try to convince social workers they are worthy.
Meanwhile, they say they're running out of money in the fight for their daughter. With barely enough left for food from Andrew Hackney鈥檚 wages at a local grocery store, he had to shut off his monthly cell phone service. They鈥檙e struggling to pay for the legal fees and gas money needed to attend appointments required of them.
In February, their daughter was diagnosed with a disorder that can disrupt her sense of taste, according to Andrew Hackney鈥檚 lawyer, Robin Frank, who added that the girl has continued to struggle to eat, even in foster care.
All they have for now are twice-weekly visits that last a few hours before she鈥檚 taken away again. Lauren Hackney鈥檚 voice breaks as she worries her daughter may be adopted and soon forget her own family. They say they yearn to do what many parents take for granted 鈥 put their child to sleep at night in her own bed.
鈥淚 really want to get my kid back. I miss her, and especially holding her. And of course, I miss that little giggly laugh,鈥 Andrew Hackney said, as his daughter sprang toward him with excitement during a recent visit. 鈥淚t hurts a lot. You have no idea how bad.鈥
___
Burke reported from San Francisco. Associated Press video journalist Jessie Wardarski and photojournalist Maye-E Wong in Pittsburgh contributed to this report.
___
Follow Sally Ho and Garance Burke on Twitter at @_sallyho and @garanceburke.
___
Contact AP’s global investigative team at Investigative@ap.org or