2000 character limit reached
Concrete Problems in AI Safety, Revisited (2401.10899v1)
Published 18 Dec 2023 in cs.CY and cs.AI
Abstract: As AI systems proliferate in society, the AI community is increasingly preoccupied with the concept of AI Safety, namely the prevention of failures due to accidents that arise from an unanticipated departure of a system's behavior from designer intent in AI deployment. We demonstrate through an analysis of real world cases of such incidents that although current vocabulary captures a range of the encountered issues of AI deployment, an expanded socio-technical framing will be required for a more complete understanding of how AI systems and implemented safety mechanisms fail and succeed in real life.
- Concrete Problems in AI Safety. arXiv:1606.06565 [cs], June 2016. URL http://arxiv.org/abs/1606.06565. arXiv: 1606.06565.
- Exploring or exploiting? social and ethical implications of autonomous experimentation in ai. In Workshop on Fairness, Accountability, and Transparency in Machine Learning, 2016.
- Abeba Birhane and Jelle van Dijk. Robot rights? let’s talk about human welfare instead. arXiv preprint arXiv:2001.05046, 2020.
- National Transportation Safety Board. Collision between a car operating with automated vehicle control systems and a tractor-semitrailer truck, 2017a. URL https://ntsb.gov/investigations/Pages/HWY18FH010.aspx.
- National Transportation Safety Board. Driver errors, overreliance on automation, lack of safeguards, led to fatal tesla crash, 2017b. URL https://www.ntsb.gov/news/press-releases/pages/pr20170912.aspx.
- National Transportation Safety Board. Collision between vehicle controlled by developmental automated driving system and pedestrian, 2018. URL https://ntsb.gov/investigations/Pages/HWY18FH010.aspx.
- ” you might also like:” privacy risks of collaborative filtering. In 2011 IEEE symposium on security and privacy, pp. 231–246. IEEE, 2011.
- On evaluating adversarial robustness. arXiv preprint arXiv:1902.06705, 2019.
- Show your work: Improved reporting of experimental results. arXiv preprint arXiv:1909.03004, 2019.
- Three Reasons Why: Framing the Challenges of Assuring AI. In Alexander Romanovsky, Elena Troubitsyna, Ilir Gashi, Erwin Schoitsch, and Friedemann Bitsch (eds.), Computer Safety, Reliability, and Security, Lecture Notes in Computer Science, pp. 281–287, Cham, 2019. Springer International Publishing. ISBN 978-3-030-26250-1. doi: 10.1007/978-3-030-26250-1˙22.
- A general safety framework for learning-based control in uncertain robotic systems. IEEE Transactions on Automatic Control, 64(7):2737–2752, 2018.
- The netflix recommender system: Algorithms, business value, and innovation. ACM Transactions on Management Information Systems (TMIS), 6(4):1–19, 2015.
- Ghost Work: How to Stop Silicon Valley from Building a New Global Underclass. Eamon Dolan Books, 2019.
- Ai safety gridworlds. arXiv preprint arXiv:1711.09883, 2017a.
- AI Safety Gridworlds. November 2017b. URL https://arxiv.org/abs/1711.09883.
- Sidneyeve Matrix. The netflix effect: Teens, binge watching, and on-demand digital media trends. Jeunesse: Young People, Texts, Cultures, 6(1):119–138, 2014.
- How to break anonymity of the netflix prize dataset. arXiv preprint cs/0610105, 2006.
- Laurent Orseau and MS Armstrong. Safely interruptible agents. 2016.
- Pots: Protective optimization technologies. arXiv preprint arXiv:1806.02711, 2018.
- SoK: Security and privacy in machine learning. In 2018 IEEE European Symposium on Security and Privacy (EuroS&P), pp. 399–414. IEEE, 2018.
- Humans and automation: Use, misuse, disuse, abuse. Human factors, 39(2):230–253, 1997.
- Safe exploration techniques for reinforcement learning–an overview. In International Workshop on Modelling and Simulation for Autonomous Systems, pp. 357–375. Springer, 2014.
- Saving face: Investigating the ethical concerns of facial recognition auditing. arXiv preprint arXiv:2001.00964, 2020.
- Charlie Schmidt. Md anderson breaks with ibm watson, raising questions about artificial intelligence in oncology. JNCI: Journal of the National Cancer Institute, 109(5), 2017.
- Green ai. arXiv preprint arXiv:1907.10597, 2019.
- Towards verified artificial intelligence. arXiv preprint arXiv:1606.08514, 2016.
- Ryan Singel. Netflix spilled your brokeback mountain secret, lawsuit claims. Threat Level (blog), Wired, 2009.
- Ryan Singel. Netflix cancels recommendation contest after privacy lawsuit. Retrieved March, 29:2018, 2010.
- Eliza Strickland. Ibm watson, heal thyself: How ibm overpromised and underdelivered on ai health care. IEEE Spectrum, 56(4):24–31, 2019.
- Energy and policy considerations for deep learning in nlp. arXiv preprint arXiv:1906.02243, 2019.
- Inc. Tesla. Tesla vehicle safety report. https://www.tesla.com/VehicleSafetyReport, 2019.
- The University of Texas System Administration. Special review of procurement procedures related to the m.d. anderson cancer center oncology expert advisor project, 2016.
- Towards Dynamic Safety Management for Autonomous Systems. In Engineering Safe Autonomy, pp. 193–204, 2019. ISBN 978-1-72936-176-4.
- Inc. Uber. Uber atg safety report. https://www.uber.com/us/en/atg/safety/, 2019.
- Kush R. Varshney. Engineering Safety in Machine Learning. arXiv:1601.04126 [cs, stat], January 2016. URL http://arxiv.org/abs/1601.04126. arXiv: 1601.04126.
- Eric Weiss. ‘inadequate safety culture’ contributed to uber automated test vehicle crash - ntsb calls for federal review process for automated vehicle testing on public roads, 2019. URL https://www.ntsb.gov/news/press-releases/Pages/NR20191119c.aspx.