Beyond Predictive Algorithms in Child Welfare (2403.05573v1)
Abstract: Caseworkers in the child welfare (CW) sector use predictive decision-making algorithms built on risk assessment (RA) data to guide and support CW decisions. Researchers have highlighted that RAs can contain biased signals which flatten CW case complexities and that the algorithms may benefit from incorporating contextually rich case narratives, i.e. - casenotes written by caseworkers. To investigate this hypothesized improvement, we quantitatively deconstructed two commonly used RAs from a United States CW agency. We trained classifier models to compare the predictive validity of RAs with and without casenote narratives and applied computational text analysis on casenotes to highlight topics uncovered in the casenotes. Our study finds that common risk metrics used to assess families and build CWS predictive risk models (PRMs) are unable to predict discharge outcomes for children who are not reunified with their birth parent(s). We also find that although casenotes cannot predict discharge outcomes, they contain contextual case signals. Given the lack of predictive validity of RA scores and casenotes, we propose moving beyond quantitative risk assessments for public sector algorithms and towards using contextual sources of information such as narratives to study public sociotechnical systems.
- Roles for computing in social change. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, pp. 252–260, 2020.
- Algorithmic profiling of job seekers in Austria: how austerity politics are made effective. Front. Big Data 3: 5. doi: 10.3389/fdata, 2020.
- Street-level algorithms and ai in bureaucratic decision-making: A caseworker perspective. Proceedings of the ACM on Human-Computer Interaction, 5(CSCW1):1–23, 2021.
- Machine Bias, May 2016.
- Narrative paths and negotiation of power in birth stories. Proceedings of the ACM on Human-Computer Interaction, 3(CSCW):1–27, 2019.
- Human-Centered Data Science: An Introduction. MIT Press, 2022.
- Augintel. About us: Social impact meets A.I., 2020.
- Augintel. Allegheny county DHS case study: Unlocking the data in case notes with natural language processing, may 2022.
- S. Barocas and A. D. Selbst. Big data’s disparate impact. Calif. L. Rev., 104:671, 2016.
- Comparing grounded theory and topic modeling: Extreme divergence or unlikely convergence? Journal of the Association for Information Science and Technology, 68(6):1397–1410, 2017.
- AAPI onine development handbook the adult-adolescent parenting inventory (AAPI-2). Family Development Resources, Inc, 2010.
- Family members’ perspectives of child protection services, a metasynthesis of the literature. Children and Youth Services Review, p. 106094, 2021.
- D. M. Blei. Probabilistic topic models. Commun. ACM, 55(4):77–84, apr 2012. doi: 10 . 1145/2133806 . 2133826
- Latent dirichlet allocation. J. Mach. Learn. Res., 3:993–1022, Mar. 2003.
- E. A. Bosk. What counts? Quantification, worker judgment, and divergence in child welfare decision making. Human Service Organizations: Management, Leadership & Governance, 42(2):205–224, 2018.
- V. Braun and V. Clarke. Using thematic analysis in psychology. Qualitative Research in Psychology, 3(2):77–101, 2006. doi: 10 . 1191/1478088706qp063oa
- L. Breiman. Random forests. Machine Learning, 45(1):5–32, Oct 2001. doi: 10 . 1023/A:1010933404324
- Toward algorithmic accountability in public services: A qualitative study of affected community perspectives on algorithmic decision-making in child welfare services. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems, p. 41. ACM, 2019.
- J. Burrell and M. Fourcade. The society of algorithms. Annual Review of Sociology, 47(1):213–237, 2021. doi: 10 . 1146/annurev-soc-090820-020800
- Chapter 8 - precision medicine in digital pathology via image analysis and machine learning. In S. Cohen, ed., Artificial Intelligence and Deep Learning in Pathology, pp. 149–173. Elsevier, 2021. doi: 10 . 1016/B978-0-323-67538-3 . 00008-7
- M. J. Camasso and R. Jagannathan. Decision making in child protective services: A risky business? Risk analysis, 33(9):1636–1649, 2013.
- Capacity Building Center for States. Child Protective Services: A guide for caseworkers, 2018.
- Who is the ”human” in human-centered machine learning: The case of predicting mental health from social media. Proc. ACM Hum.-Comput. Interact., 3(CSCW), nov 2019. doi: 10 . 1145/3359249
- SMOTE: Synthetic minority over-sampling technique. Journal of Artificial Intelligence Research, 16:321–357, jun 2002. doi: 10 . 1613/jair . 953
- How child welfare workers reduce racial disparities in algorithmic decisions. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, CHI ’22. Association for Computing Machinery, New York, NY, USA, 2022. doi: 10 . 1145/3491102 . 3501831
- Child Welfare Information Gateway. Family Preservation Services.
- Child Welfare Information Gateway. What is child abuse and neglect? Recognizing the signs and symptoms, 2019.
- Child Welfare Information Gateway. Concurrent planning for timely permanency for children, 2021.
- Child Welfare Information Gateway. The use of safety and risk assessments in child protection cases, 2022.
- A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Conference on Fairness, Accountability and Transparency, pp. 134–148, 2018.
- A. Christin. Algorithms in practice: Comparing web journalism and criminal justice. Big Data & Society, 4(2):2053951717718855, 2017. doi: 10 . 1177/2053951717718855
- Reconciling data-driven crime analysis with human-centered algorithms. Cities, 124:103604, 2022. doi: 10 . 1016/j . cities . 2022 . 103604
- O. H. R. Commission. Under suspicion: Concerns about child welfare, 2017.
- The role of administrative data in the big data revolution in social science research. Social science research, 59:1–12, 2016.
- Measuring the potential for child maltreatment: The reliability and validity of the Adult Adolescent Parenting Inventory—2. Child Abuse & Neglect, 30(1):39–53, 2006. doi: 10 . 1016/j . chiabu . 2005 . 08 . 011
- V. A. Copeland. “It’s the Only System We’ve Got”: Exploring emergency response decision-making in child welfare. Columbia Journal of Race and Law, 11(3):43–74, 2021.
- C. Cortes and V. Vapnik. Support-vector networks. Machine Learning, 20(3):273–297, Sep 1995. doi: 10 . 1007/BF00994018
- A validity perspective on evaluating the justified use of data-driven decision-making algorithms. In 2023 IEEE Conference on Secure and Trustworthy Machine Learning (SaTML), pp. 690–704. IEEE, 2023.
- Toward cultural bias evaluation datasets: The case of Bengali gender, religious, and national identity. In S. Dev, V. Prabhakaran, D. Adelani, D. Hovy, and L. Benotti, eds., Proceedings of the First Workshop on Cross-Cultural Considerations in NLP (C3NLP), pp. 68–83. Association for Computational Linguistics, Dubrovnik, Croatia, May 2023. doi: 10 . 18653/v1/2023 . c3nlp-1 . 8
- A. J. Dettlaff and R. Boyd. Racial disproportionality and disparities in the child welfare system: Why do they exist, and what can be done to address them? The ANNALS of the American Academy of Political and Social Science, 692(1):253–274, 2020. doi: 10 . 1177/0002716220980329
- Disentangling substantiation: The influence of race, income, and risk on the substantiation decision in child welfare. Children and Youth Services Review, 33(9):1630–1637, 2011.
- J. Dressel and H. Farid. The accuracy, fairness, and limits of predicting recidivism. Science Advances, 4(1):eaao5580, 2018. doi: 10 . 1126/sciadv . aao5580
- Edx. Adult-Adolescent Parenting Inventory-2 (AAPI-2), 2015.
- V. Eubanks. Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin’s Press, 2018.
- Family Development Resources. Inventory scoring system for assessing parenting practices.
- Examining risks of racial biases in NLP tools for child protective services. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’23, p. 1479–1492. Association for Computing Machinery, New York, NY, USA, 2023. doi: 10 . 1145/3593013 . 3594094
- Publication bias in the social sciences: Unlocking the file drawer. Science, 345(6203):1502–1505, 2014. doi: 10 . 1126/science . 1255484
- A review on ensembles for the class imbalance problem: Bagging-, boosting-, and hybrid-based approaches. IEEE Transactions on Systems, Man, and Cybernetics, Part C (Applications and Reviews), 42(4):463–484, 2012. doi: 10 . 1109/TSMCC . 2011 . 2161285
- J. M. Geiger and L. Schelbe. Foster care placement. In The Handbook on Child Welfare Practice, pp. 219–248. Springer, 2021.
- The devil is in the details: Interrogating values embedded in the allegheny family screening tool. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’23, p. 1292–1310. Association for Computing Machinery, New York, NY, USA, 2023. doi: 10 . 1145/3593013 . 3594081
- P. Gillingham. Decision-making about the adoption of information technology in social welfare agencies: Some key considerations. European Journal of Social Work, 21(4):521–529, 2018.
- J. D. Goldhaber-Fiebert and L. Prince. Impact evaluation of a predictive risk modeling tool for Allegheny county’s child welfare office. Technical report, 2019.
- The hidden inconsistencies introduced by predictive algorithms in judicial decision making. arXiv preprint arXiv:2012.00289, 2020.
- S. Ho and G. Burke. An algorithm that screens for child neglect raises concerns., 2022.
- Work of the unemployed: An inquiry into individuals’ experience of data usage in public services and possibilities for their agency. In Designing Interactive Systems Conference 2021, DIS ’21, p. 438–448. Association for Computing Machinery, New York, NY, USA, 2021. doi: 10 . 1145/3461778 . 3462003
- Shifting concepts of value: Designing algorithmic decision-support systems for public services. In Proceedings of the 11th Nordic Conference on Human-Computer Interaction: Shaping Experiences, Shaping Society, pp. 1–12, 2020.
- Hornby Zeller Associates INC. Allegheny county predictive risk modeling tool implementation: Process evaluation. Technical report, 2018.
- The meaning and measurement of bias: Lessons from natural language processing. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, FAT* ’20, p. 706. Association for Computing Machinery, New York, NY, USA, 2020. doi: 10 . 1145/3351095 . 3375671
- N. Japkowicz and S. Stephen. The class imbalance problem: A systematic study. Intelligent Data Analysis, pp. 429–449, 2002.
- Improving human-AI partnerships in child welfare: Understanding worker practices, challenges, and desires for algorithmic decision support. In Proceedings of the 2022 CHI Conference on Human Factors in Computing Systems, CHI ’22. Association for Computing Machinery, New York, NY, USA, 2022. doi: 10 . 1145/3491102 . 3517439
- “Why do I care what’s similar?” Probing challenges in AI-assisted child welfare decision-making through worker-AI interface design concepts. In Designing Interactive Systems Conference, DIS ’22, p. 454–470. Association for Computing Machinery, New York, NY, USA, 2022. doi: 10 . 1145/3532106 . 3533556
- E. Keddell. The ethics of predictive risk modelling in the Aotearoa/New Zealand child welfare context: Child abuse prevention or neo-liberal tool? Critical Social Policy, 35(1):69–88, 2015. doi: 10 . 1177/0261018314543224
- R. S. Kirk. Psychometric properties of the trauma and post-trauma well-being assessment domains of the North Carolina Family Assessment Scale for General and Reunification Services (NCFAS G+R). Journal of Public Child Welfare, 9(5):444–462, 2015. doi: 10 . 1080/15548732 . 2015 . 1090364
- R. S. Kirk and K. Reed-Ashcraft. NCFAS North Carolina Family Assessment Scale Research Report. National Family Preservation Network, pp. 1–22, 2004.
- Retention of child welfare caseworkers: The role of case severity and workplace resources. Children and Youth Services Review, 126:106039, 2021.
- The misuse of AUC: What high impact risk assessment gets wrong. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’23, p. 1570–1583. Association for Computing Machinery, New York, NY, USA, 2023. doi: 10 . 1145/3593013 . 3594100
- North Carolina Family Assessment Scale: Measurement Properties for Youth Mental Health Services. Research on Social Work Practice, 20(2):202–211, 2010. doi: 10 . 1177/1049731509334180
- Y. Lei. 3 - Individual intelligent method-based fault diagnosis. In Y. Lei, ed., Intelligent Fault Diagnosis and Remaining Useful Life Prediction of Rotating Machinery, pp. 67–174. Butterworth-Heinemann, 2017. doi: 10 . 1016/B978-0-12-811534-3 . 00003-2
- Imbalanced-learn: A Python toolbox to tackle the curse of imbalanced datasets in machine learning. Journal of Machine Learning Research, 18(17):1–5, 2017.
- AUC: A misleading measure of the performance of predictive distribution models. Global Ecology and Biogeography, 17(2):145–151, 2008. doi: 10 . 1111/j . 1466-8238 . 2007 . 00358 . x
- P. Marks. Algorithmic hiring needs a human face. Commun. ACM, 65(3):17–19, feb 2022. doi: 10 . 1145/3510552
- A human-centered review of algorithms in decision-making in higher education. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, CHI ’23. Association for Computing Machinery, New York, NY, USA, 2023. doi: 10 . 1145/3544548 . 3580658
- Assessing risk: A comparison of tools for child welfare practice with Indigenous families, may 2017.
- E. S. Y. Moon and S. Guha. A human-centered review of algorithms in homelessness research. 2024. doi: 10 . 1145/3613904 . 3642392
- Towards a non-ideal methodological framework for responsible ml, 2024.
- Machine learning and grounded theory method: Convergence, divergence, and combination. In Proceedings of the 19th International Conference on Supporting Group Work, pp. 3–8. ACM, 2016.
- National Family Preservation Network. NCFAS North Carolina Family Assessment Scale Scale & Definitions (v. 2.0). 2009.
- How we do things with words: Analyzing text as social and cultural data. Frontiers in Artificial Intelligence, 3:62, 2020.
- L. H. Nguyen and S. Holmes. Ten quick tips for effective dimensionality reduction. JPLoS Comput Biol., 15(6), june 2019. doi: 10 . 1371/journal . pcbi . 1006907
- W. S. Noble. What is a support vector machine? Nature Biotechnology, 24(12):1565–1567, Dec 2006. doi: 10 . 1038/nbt1206-1565
- A Collaborative Way of Knowing: Bridging Computational Communication Research and Grounded Theory Ethnography. Journal of Communication, 70(3):447–472, 2020.
- The fallacy of AI functionality. In 2022 ACM Conference on Fairness, Accountability, and Transparency, FAccT ’22, p. 959–972. Association for Computing Machinery, New York, NY, USA, Jun 2022. doi: 10 . 1145/3531146 . 3533158
- Datafied child welfare services: Unpacking politics, economics and power. Policy Studies, 41(5):507–526, 2020.
- Confidence and professional judgment in assessing children’s risk of abuse. Research on Social Work Practice, 20(6):621–628, 2010. doi: 10 . 1177/1049731510368050
- D. E. Roberts. Racial harm: Dorothy Roberts explains how racism works in the child welfare system. Colorlines, 5(3):19, Fall 2002.
- D. E. Roberts. Torn Apart How the Child Welfare System Destroys Black Families–and How Abolition Can Build a Safer World. Basic Books, New York, NY, USA, 2022.
- Modeling assumptions clash with the real world: Transparency, equity, and community challenges for student assignment algorithms. arXiv preprint arXiv:2101.10367, 2021.
- S. Robertson and N. Salehi. What if I don’t like any of the choices? The limits of preference elicitation for participatory algorithm design. arXiv preprint arXiv:2007.06718, 2020.
- Family surveillance by algorithm: The rapidly spreading tools few have heard of, 2021.
- Child welfare system: Interaction of policy, practice and algorithms. In Companion Proceedings of the 2020 ACM International Conference on Supporting Group Work, pp. 119–122, 2020.
- A framework of high-stakes algorithmic decision-making for the public sector developed through a case study of child-welfare. Proceedings of the ACM on Human-Computer Interaction, 5(CSCW2), 2021.
- A human-centered review of algorithms used within the us child welfare system. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems, pp. 1–15, 2020.
- D. Saxena and S. Guha. Algorithmic harms in child welfare: Uncertainties in practice, organization, and street-level decision-making. ACM J. Responsib. Comput., sep 2023. doi: 10 . 1145/3616473
- Rethinking ”Risk” in algorithmic systems through a computational narrative analysis of casenotes in child-welfare. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems, pp. 1–19, 2023.
- Unpacking invisible work practices, constraints, and latent power relationships in child welfare through casenote analysis. CHI ’22. Association for Computing Machinery, New York, NY, USA, 2022. doi: 10 . 1145/3491102 . 3517742
- How to train a (bad) algorithmic caseworker: A quantitative deconstruction of risk assessments in child welfare. In CHI Conference on Human Factors in Computing Systems Extended Abstracts, pp. 1–7, 2022.
- C. S. Schwalbe. Strengthening the integration of actuarial risk assessment with clinical judgment in an evidence based practice framework. Children and Youth Services Review, 30(12):1458–1464, 2008.
- Auditing risk prediction of long-term unemployment. Proc. ACM Hum.-Comput. Interact., 6(GROUP), jan 2022. doi: 10 . 1145/3492827
- Fairness and abstraction in sociotechnical systems. In Proceedings of the Conference on Fairness, Accountability, and Transparency, FAT* ’19, p. 59–68. Association for Computing Machinery, New York, NY, USA, 2019. doi: 10 . 1145/3287560 . 3287598
- A. Shlonsky and D. Wagner. The next step: Integrating actuarial risk assessment and clinical judgment into an evidence-based practice framework in cps case management. Children and youth services review, 27(4):409–427, 2005.
- B. Shneiderman. Human-Centered AI. Oxford University Press, 01 2022. doi: 10 . 1093/oso/9780192845290 . 001 . 0001
- N. L. Sidell. Social Work Documentation. NASW Press, Washington, DC, USA, 2015.
- A strengths-based approach to supervised visitation in child welfare. Child Care in Practice, 20(1):98–119, 2014.
- Imagining new futures beyond predictive systems in child welfare: A qualitative study with impacted stakeholders. In FAccT ’22: 2022 ACM Conference on Fairness, Accountability, and Transparency, Seoul, Republic of Korea, June 21 - 24, 2022, pp. 1162–1177. ACM, 2022. doi: 10 . 1145/3531146 . 3533177
- Z. Strassburger. Medical decision making for youth in the foster care system. J. Marshall L. Rev., 49(4):1103–1154, 2016.
- D. G. N. Y. Times. San Francisco had an ambitious plan to tackle school segregation. It made it worse., April 2019.
- Using a machine learning tool to support high-stakes decisions in child protection. AI Magazine, 42(1):53–60, 2021.
- Hospital Injury Encounters of Children Identified by a Predictive Risk Model for Screening Child Maltreatment Referrals: Evidence From the Allegheny Family Screening Tool. JAMA pediatrics, 174(11):e202770, Nov 2020. doi: 10 . 1001/jamapediatrics . 2020 . 2770
- Developing predictive models to support child maltreatment hotline screening decisions: Allegheny county methodology and implementation. Center for Social data Analytics, 2017.
- M. Veale and I. Brass. Administration by algorithm? Public management meets public sector machine learning. (3375391), 2019.
- Fairness and accountability design needs for algorithmic support in high-stakes public sector decision-making. In Proceedings of the 2018 chi conference on human factors in computing systems, pp. 1–14, 2018.
- M. Wieringa. What to account for when accounting for algorithms: A systematic literature review on algorithmic accountability. In Proceedings of the 2020 Conference on Fairness, Accountability, and Transparency, FAT* ’20, p. 1–18. Association for Computing Machinery, New York, NY, USA, 2020. doi: 10 . 1145/3351095 . 3372833
- B. Williamson. Digital education governance: Data visualization, predictive analytics, and ‘real-time’policy instruments. Journal of Education Policy, 31(2):123–141, 2016.
- RepliCHI - CHI Should Be Replicating and Validating Results More: Discuss. In CHI ’11 Extended Abstracts on Human Factors in Computing Systems, CHI EA ’11, p. 463–466. Association for Computing Machinery, New York, NY, USA, 2011. doi: 10 . 1145/1979742 . 1979491