Explainability through uncertainty: Trustworthy decision-making with neural networks (2403.10168v1)
Abstract: Uncertainty is a key feature of any machine learning model and is particularly important in neural networks, which tend to be overconfident. This overconfidence is worrying under distribution shifts, where the model performance silently degrades as the data distribution diverges from the training data distribution. Uncertainty estimation offers a solution to overconfident models, communicating when the output should (not) be trusted. Although methods for uncertainty estimation have been developed, they have not been explicitly linked to the field of explainable artificial intelligence (XAI). Furthermore, literature in operations research ignores the actionability component of uncertainty estimation and does not consider distribution shifts. This work proposes a general uncertainty framework, with contributions being threefold: (i) uncertainty estimation in ML models is positioned as an XAI technique, giving local and model-specific explanations; (ii) classification with rejection is used to reduce misclassifications by bringing a human expert in the loop for uncertain observations; (iii) the framework is applied to a case study on neural networks in educational data mining subject to distribution shifts. Uncertainty as XAI improves the model's trustworthiness in downstream decision-making tasks, giving rise to more actionable and robust machine learning systems in operations research.
- On calibration of modern neural networks. In International Conference on Machine Learning, pages 1321–1330. PMLR, 2017.
- K Murphy. Probabilistic machine learning: Advanced topics, chapter 20. MIT Press, 2022.
- Chip Huyen. Designing Machine Learning Systems, chapter 8. O’Reilly Media, Inc., 2022.
- Can you trust your model’s uncertainty? evaluating predictive uncertainty under dataset shift. Advances in neural information processing systems, 32, 2019.
- Quod erat demonstrandum?-towards a typology of the concept of explanation for the design of explainable ai. Expert Systems with Applications, 213:118888, 2023. doi:10.1016/j.eswa.2022.118888.
- A survey on uncertainty estimation in deep learning classification systems from a bayesian perspective. ACM Computing Surveys (CSUR), 54(9):1–35, 2021. doi:10.1145/3477140.
- Big data analytics in operations management. Production and Operations Management, 27(10):1868–1883, 2018. doi:10.1111/poms.12838.
- Donghee Shin. The effects of explainability and causability on perception, trust, and acceptance: Implications for explainable ai. International Journal of Human-Computer Studies, 146:102551, 2021. doi:10.1016/j.ijhcs.2020.102551.
- Explainable models of credit losses. European Journal of Operational Research, 301(1):386–394, 2022. doi:10.1016/j.ejor.2021.11.009.
- An explainable ai decision-support-system to automate loan underwriting. Expert Systems with Applications, 144:113100, 2020. doi:10.1016/j.eswa.2019.113100.
- A new hybrid classification algorithm for customer churn prediction based on logistic regression and decision trees. European Journal of Operational Research, 269(2):760–772, 2018. doi:10.1016/j.ejor.2018.02.009.
- Predicting customer demand for remanufactured products: A data-mining approach. European Journal of Operational Research, 281(3):543–558, 2020. doi:10.1016/j.ejor.2019.08.015.
- An analytical framework for supply network risk propagation: A bayesian network approach. European Journal of Operational Research, 243(2):618–627, 2015. doi:10.1016/j.ejor.2014.10.034.
- A data analytics approach to building a clinical decision support system for diabetic retinopathy: Developing and deploying a model ensemble. Decision Support Systems, 101:12–27, 2017. doi:10.1016/j.dss.2017.05.012.
- To imprison or not to imprison: an analytics model for drug courts. Annals of Operations Research, 303:101–124, 2021. doi:10.1007/s10479-021-03984-7.
- Peeking inside the black-box: a survey on explainable artificial intelligence (xai). IEEE access, 6:52138–52160, 2018. doi:10.1109/ACCESS.2018.2870052.
- Deep learning in business analytics and operations research: Models, applications and managerial implications. European Journal of Operational Research, 281(3):628–641, 2020. ISSN 0377-2217. doi:10.1016/j.ejor.2019.09.018. Featured Cluster: Business Analytics: Defining the field and identifying a research agenda.
- Deep learning for credit scoring: Do or don’t? European Journal of Operational Research, 295(1):292–305, 2021. ISSN 0377-2217. doi:10.1016/j.ejor.2021.03.006.
- Autoencoders for strategic decision support. Decision Support Systems, 150:113422, 2021. ISSN 0167-9236. doi:10.1016/j.dss.2020.113422. Interpretable Data Science For Decision Making.
- Using shared sell-through data to forecast wholesaler demand in multi-echelon supply chains. European Journal of Operational Research, 288(2):466–479, 2021. ISSN 0377-2217. doi:10.1016/j.ejor.2020.05.059.
- Forecasting remaining useful life: Interpretable deep learning approach via variational bayesian inferences. Decision Support Systems, 125:113100, 2019. ISSN 0167-9236. doi:10.1016/j.dss.2019.113100.
- A hybrid neural variational cf-nade for collaborative filtering using abstraction and generation. Expert Systems with Applications, 179:115047, 2021. ISSN 0957-4174. doi:10.1016/j.eswa.2021.115047.
- Alireza Ghahtarani. A new portfolio selection problem in bubble condition under uncertainty: Application of z-number theory and fuzzy neural network. Expert Systems with Applications, 177:114944, 2021. ISSN 0957-4174. doi:10.1016/j.eswa.2021.114944.
- Inductive gaussian representation of user-specific information for personalized stress-level prediction. Expert Systems with Applications, 178:114912, 2021. ISSN 0957-4174. doi:10.1016/j.eswa.2021.114912.
- Bayesian neural networks for flight trajectory prediction and safety assessment. Decision Support Systems, 131:113246, 2020. ISSN 0167-9236. doi:10.1016/j.dss.2020.113246.
- Understanding the uncertainty of traffic time prediction impacts on parking lot reservation in logistics centers. Annals of Operations Research, pages 1–23, 2022. doi:10.1007/s10479-022-04734-z.
- Learning uncertainty with artificial neural networks for predictive process monitoring. Applied Soft Computing, 125:109134, 2022. doi:10.1016/j.asoc.2022.109134.
- Exploring bayesian deep learning for urgent instructor intervention need in mooc forums. In International Conference on Intelligent Tutoring Systems, pages 78–90. Springer, 2021. doi:10.1007/978-3-030-80421-3_10.
- Explainable deep learning for efficient and robust pattern recognition: A survey of recent developments. Pattern Recognition, 120:108102, 2021. doi:10.1016/j.patcog.2021.108102.
- Aleatoric and epistemic uncertainty in machine learning: An introduction to concepts and methods. Machine Learning, 110(3):457–506, 2021. doi:10.1007/s10994-021-05946-3.
- Aleatoric and epistemic uncertainty with random forests. In Advances in Intelligent Data Analysis XVIII: 18th International Symposium on Intelligent Data Analysis, IDA 2020, Konstanz, Germany, April 27–29, 2020, Proceedings 18, pages 444–456. Springer, 2020. doi:10.1007/978-3-030-44584-3_35.
- Dropout as a bayesian approximation: Representing model uncertainty in deep learning. In international conference on machine learning, pages 1050–1059. PMLR, 2016.
- Simple and scalable predictive uncertainty estimation using deep ensembles. Advances in neural information processing systems, 30, 2017.
- Aleatory or epistemic? does it matter? Structural safety, 31(2):105–112, 2009. doi:10.1016/j.strusafe.2008.06.020.
- Towards a rigorous science of interpretable machine learning. arXiv preprint arXiv:1702.08608, 2017. doi:10.48550/arXiv.1702.08608.
- Uncertainty-based rejection in machine learning: Implications for model development and interpretability. Electronics, 11(3):396, 2022. doi:10.3390/electronics11030396.
- Performance measures for classification systems with rejection. Pattern Recognition, 63:437–450, 2017. doi:10.1016/j.patcog.2016.10.011.
- Bayesian autoencoders with uncertainty quantification: Towards trustworthy anomaly detection. Expert Systems with Applications, 209:118196, 2022. ISSN 0957-4174. doi:10.1016/j.eswa.2022.118196.
- Andrew Gordon Wilson. The case for bayesian deep learning. arXiv preprint arXiv:2001.10995, 2020. doi:10.48550/arXiv.2001.10995.
- Decomposition of uncertainty in bayesian deep learning for efficient and risk-sensitive learning. In International Conference on Machine Learning, pages 1184–1193. PMLR, 2018.
- Development of a bayesian belief network-based dss for predicting and understanding freshmen student attrition. European Journal of Operational Research, 281(3):575–587, 2020. ISSN 0377-2217. doi:10.1016/j.ejor.2019.03.037. Featured Cluster: Business Analytics: Defining the field and identifying a research agenda.
- Predicting student dropout in subscription-based online learning environments: The beneficial impact of the logit leaf model. Decision Support Systems, 135:113325, 2020. ISSN 0167-9236. doi:10.1016/j.dss.2020.113325.
- Uplift modeling for preventing student dropout in higher education. Decision Support Systems, 134:113320, 2020. ISSN 0167-9236. doi:10.1016/j.dss.2020.113320.
- Predicting student performance using sequence classification with time-based windows. Expert Systems with Applications, 209:118182, 2022. ISSN 0957-4174. doi:10.1016/j.eswa.2022.118182.
- A decision support framework to incorporate textual data for early student dropout prediction in higher education. Decision Support Systems, page 113940, 2023. doi:10.1016/j.dss.2023.113940.
- Mooc dropout prediction: How to measure accuracy? In Proceedings of the fourth (2017) acm conference on learning@ scale, pages 161–164, 2017. doi:10.1145/3051457.3053974.
- Student success prediction in moocs. User Modeling and User-Adapted Interaction, 28(2):127–203, 2018. doi:10.1007/s11257-018-9203-z.
- Learning analytics should not promote one size fits all: The effects of instructional conditions in predicting academic success. The Internet and Higher Education, 28:68–84, 2016. doi:10.1016/j.iheduc.2015.10.002.
- MITx and HarvardX. Hmxpc13_di_v2_5-14-14.csv. In HarvardX-MITx Person-Course Academic Year 2013 De-Identified dataset, version 2.0. Harvard Dataverse, 2014. doi:10.7910/DVN/26147/OCLJIV.
- Gaussian processes for unconstraining demand. European Journal of Operational Research, 275(2):621–634, 2019. ISSN 0377-2217. doi:10.1016/j.ejor.2018.11.065.
- Active learning strategies for interactive elicitation of assignment examples for threshold-based multiple criteria sorting. European Journal of Operational Research, 293(2):658–680, 2021. ISSN 0377-2217. doi:10.1016/j.ejor.2020.12.055.