The Situate AI Guidebook: Co-Designing a Toolkit to Support Multi-Stakeholder Early-stage Deliberations Around Public Sector AI Proposals (2402.18774v2)
Abstract: Public sector agencies are rapidly deploying AI systems to augment or automate critical decisions in real-world contexts like child welfare, criminal justice, and public health. A growing body of work documents how these AI systems often fail to improve services in practice. These failures can often be traced to decisions made during the early stages of AI ideation and design, such as problem formulation. However, today, we lack systematic processes to support effective, early-stage decision-making about whether and under what conditions to move forward with a proposed AI project. To understand how to scaffold such processes in real-world settings, we worked with public sector agency leaders, AI developers, frontline workers, and community advocates across four public sector agencies and three community advocacy groups in the United States. Through an iterative co-design process, we created the Situate AI Guidebook: a structured process centered around a set of deliberation questions to scaffold conversations around (1) goals and intended use or a proposed AI system, (2) societal and legal considerations, (3) data and modeling constraints, and (4) organizational governance factors. We discuss how the guidebook's design is informed by participants' challenges, needs, and desires for improved deliberation processes. We further elaborate on implications for designing responsible AI toolkits in collaboration with public sector agency stakeholders and opportunities for future work to expand upon the guidebook. This design approach can be more broadly adopted to support the co-creation of responsible AI toolkits that scaffold key decision-making processes surrounding the use of AI in the public sector and beyond.
- 2022. Community jury - Azure Application Architecture Guide. https://learn.microsoft.com/en-us/azure/architecture/guide/responsible-innovation/community-jury/
- 2023. Algorithmic Impact Assessment tool — Government of Canada. https://www.canada.ca/en/government/system/digital-government/digital-government-innovations/responsible-use-ai/algorithmic-impact-assessment.html
- 2023. A guiding framework to vetting public sector technology vendors. https://www.fordfoundation.org/work/learning/research-reports/a-guiding-framework-to-vetting-public-sector-technology-vendors/
- J Khadijah Abdurahman. 2021. Calculating the Souls of Black Folk: Predictive Analytics in the New York City Administration for Children’s Services. In Colum. J. Race & L. Forum, Vol. 11. HeinOnline, 75.
- Accenture. 2018. Public services in the era of artificial intelligence. https://www.accenture.com/us-en/services/public-service/artificial-intelligence
- Artificial intelligence and fraud detection. Innovative Technology at the Interface of Finance and Operations: Volume I (2022), 223–247.
- Development of a predictive model to identify inpatients at risk of re-admission within 30 days of discharge (PARR-30). BMJ open 2, 4 (2012), e001667.
- Virginia Braun and Victoria Clarke. 2019. Reflecting on reflexive thematic analysis. Qualitative research in sport, exercise and health 11, 4 (2019), 589–597.
- Toward algorithmic accountability in public services: A qualitative study of affected community perspectives on algorithmic decision-making in child welfare services. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–12.
- How Child Welfare Workers Reduce Racial Disparities in Algorithmic Decisions. In CHI Conference on Human Factors in Computing Systems. 1–22.
- A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. In Conference on Fairness, Accountability and Transparency. PMLR, 134–148.
- A case study of algorithm-assisted decision making in child maltreatment hotline screening decisions. Proceedings of Machine Learning Research 81 (2018), 1–15. http://proceedings.mlr.press/v81/chouldechova18a.html
- A Validity Perspective on Evaluating the Justified Use of Data-driven Decision-making Algorithms. arXiv preprint arXiv:2206.14983 (2022).
- Tim Dare and Eileen Gambrill. 2017. Ethical analysis: Predictive risk models at call screening for Allegheny County. Alleghany County Analytics (2017).
- “Many miles to go…”: a systematic review of the implementation of patient decision support interventions into routine clinical practice. BMC medical informatics and decision making 13 (2013), 1–10.
- Government by algorithm: Artificial intelligence in federal administrative agencies. NYU School of Law, Public Law Research Paper 20-54 (2020).
- Virginia Eubanks. 2018. Automating inequality: How high-tech tools profile, police, and punish the poor. St. Martin’s Press.
- Data Ethics Decision Aid (DEDA): a dialogical framework for ethical inquiry of AI and data projects in the Netherlands. Ethics and Information Technology (2021), 1–17.
- Datasheets for datasets. Commun. ACM 64, 12 (2021), 86–92.
- The Devil is in the Details: Interrogating Values Embedded in the Allegheny Family Screening Tool. In Proceedings of the 2023 ACM Conference on Fairness, Accountability, and Transparency. 1292–1310.
- Barriers and facilitators to implementing shared decision-making in clinical practice: a systematic review of health professionals’ perceptions. Implementation science 1 (2006), 1–12.
- Ben Green. 2020. The false promise of risk assessments: epistemic reform and the limits of fairness. In Proceedings of the 2020 conference on fairness, accountability, and transparency. 594–606.
- Ben Green. 2022. The flaws of policies requiring human oversight of government algorithms. Computer Law & Security Review 45 (2022), 105681.
- Sally Ho and Garance Burke. 2022. An algorithm that screens for child neglect raises concerns. AP News (2022).
- Sally Ho and Garance Burke. 2023. Child welfare algorithm faces Justice Department scrutiny. AP News (2023).
- Intelligent tutors as teachers’ aides: exploring teacher needs for real-time analytics in blended classrooms. In Proceedings of the seventh international learning analytics & knowledge conference. 257–266.
- Shifting concepts of value: Designing algorithmic decision-support systems for public services. NordiCHI (2020), 1–12. https://doi.org/10.1145/3419249.3420149
- NPR June. 2022. Oregon is dropping an artificial intelligence tool used in child welfare system. Link: https://www. npr. org/2022/06/02/1102661376/oregon-drops-artificial-intelligence-child-abuse-cases (2022).
- Studying Up Public Sector AI: How Networks of Power Relations Shape Agency Decisions Around AI Design and Use. (2023).
- Improving Human-AI Partnerships in Child Welfare: Understanding Worker Practices, Challenges, and Desires for Algorithmic Decision Support. In CHI Conference on Human Factors in Computing Systems. 1–18.
- “Why Do I Care What’s Similar?” Probing Challenges in AI-Assisted Child Welfare Decision-Making through Worker-AI Interface Design Concepts. In Designing Interactive Systems Conference. 454–470.
- Human decisions and machine predictions. The quarterly journal of economics 133, 1 (2018), 237–293.
- An action-oriented AI policy toolkit for technology audits by community advocates and activists. In Proceedings of the 2021 ACM conference on fairness, accountability, and transparency. 772–781.
- Supporting reflective public thought with considerit. In Proceedings of the ACM 2012 conference on Computer Supported Cooperative Work. 265–274.
- Understanding Frontline Workers’ and Unhoused Individuals’ Perspectives on AI Used in Homeless Services. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–17.
- Algorithms and Decision-Making in the Public Sector. Annual Review of Law and Social Science 17 (2021), 1–38.
- Co-designing checklists to understand organizational challenges and opportunities around fairness in AI. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1–14.
- Shefeh Prisilia Mbuy and Marco Ortolani. 2022. Algorithmic Impact Assessment for an Ethical Use of AI in SMEs. https://doi.org/10.14236/ewic/HCI2022.34
- Model cards for model reporting. In Proceedings of the conference on fairness, accountability, and transparency. 220–229.
- Assembling accountability: algorithmic impact assessment for the public interest. Available at SSRN 3877437 (2021).
- J Murray. 2021. Public sector vs. private sector: What’s the difference. The Balance Small Business (2021).
- Arvind Narayanan. 2019. How to recognize AI snake oil. Arthur Miller Lecture on Science and Ethics (2019).
- NIST. 2023a. NIST AI RMF Playbook. https://airc.nist.gov/AI_RMF_Knowledge_Base/Playbook
- NIST. 2023b. NIST Risk Management Framework. https://csrc.nist.gov/projects/risk-management/about-rmf
- Ziad Obermeyer and Ezekiel J Emanuel. 2016. Predicting the future—big data, machine learning, and clinical medicine. The New England journal of medicine 375, 13 (2016), 1216.
- Predictive risk modelling in health: options for New Zealand and Australia. Australian Health Review 35, 1 (2011), 45–51.
- Samir Passi and Solon Barocas. 2019. Problem formulation and fairness. In Proceedings of the conference on fairness, accountability, and transparency. 39–48.
- Allied Media Projects. 2018. A People’s Guide To Tech. https://alliedmedia.org/resources/peoples-guide-to-ai
- Data cards: Purposeful and transparent dataset documentation for responsible ai. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency. 1776–1826.
- The fallacy of AI functionality. In 2022 ACM Conference on Fairness, Accountability, and Transparency. 959–972.
- Where responsible AI meets reality: Practitioner perspectives on enablers for shifting organizational practices. Proceedings of the ACM on Human-Computer Interaction 5, CSCW1 (2021), 1–23.
- Algorithmic Impact Assessments: A Practical Framework for Public Agency. AI Now (2018).
- A methodology for creating AI FactSheets. arXiv preprint arXiv:2006.13796 (2020).
- Dorothy Roberts. 2022. Torn apart: how the child welfare system destroys black families–and how abolition can build a safer world. Basic Books.
- Modeling assumptions clash with the real world: Transparency, equity, and community challenges for student assignment algorithms. In Proceedings of the 2021 CHI Conference on Human Factors in Computing Systems. 1–14.
- Family surveillance by algorithm: The rapidly spreading tools few have heard of. American Civil Liberties Union (ACLU) (2021). https://www.aclu.org/sites/default/files/field_document/2021.09.28a_family_surveillance_by_algorithm.pdf
- Family surveillance by algorithm: The rapidly spreading tools few have heard of. American Civil Liberties Union (ACLU)(2021).
- A human-centered review of algorithms used within the US child welfare system. In Proceedings of the 2020 CHI Conference on Human Factors in Computing Systems. 1–15.
- Rethinking” Risk” in Algorithmic Systems Through A Computational Narrative Analysis of Casenotes in Child-Welfare. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–19.
- Katie Shilton. 2013. Values levers: Building ethics into design. Science, Technology, & Human Values 38, 3 (2013), 374–397.
- Imagining new futures beyond predictive systems in child welfare: A qualitative study with impacted stakeholders. In Proceedings of the 2022 ACM Conference on Fairness, Accountability, and Transparency. 1162–1177.
- Intelligent tutoring systems: Why teachers abandoned a technology aimed at automating teaching processes. (2021).
- Children in the public benefit system at risk of maltreatment: Identification via predictive modeling. American journal of preventive medicine 45, 3 (2013), 354–359.
- Fairness and accountability design needs for algorithmic support in high-stakes public sector decision-making. In Proceedings of the 2018 chi conference on human factors in computing systems. 1–14.
- Against Predictive Optimization: On the Legitimacy of Decision-Making Algorithms that Optimize Predictive Accuracy. Available at SSRN (2022).
- Seeing like a toolkit: How toolkits envision the work of AI ethics. Proceedings of the ACM on Human-Computer Interaction 7, CSCW1 (2023), 1–27.
- Unremarkable ai: Fitting intelligent decision support into critical, clinical decision-making processes. In Proceedings of the 2019 CHI Conference on Human Factors in Computing Systems. 1–11.
- Creating Design Resources to Scaffold the Ideation of AI Concepts. In Proceedings of the 2023 ACM Designing Interactive Systems Conference. 2326–2346.
- Investigating How Practitioners Use Human-AI Guidelines: A Case Study on the People+ AI Guidebook. In Proceedings of the 2023 CHI Conference on Human Factors in Computing Systems. 1–13.