Model-Free Local Recalibration of Neural Networks (2403.05756v1)
Abstract: Artificial neural networks (ANNs) are highly flexible predictive models. However, reliably quantifying uncertainty for their predictions is a continuing challenge. There has been much recent work on "recalibration" of predictive distributions for ANNs, so that forecast probabilities for events of interest are consistent with certain frequency evaluations of them. Uncalibrated probabilistic forecasts are of limited use for many important decision-making tasks. To address this issue, we propose a localized recalibration of ANN predictive distributions using the dimension-reduced representation of the input provided by the ANN hidden layers. Our novel method draws inspiration from recalibration techniques used in the literature on approximate Bayesian computation and likelihood-free inference methods. Most existing calibration methods for ANNs can be thought of as calibrating either on the input layer, which is difficult when the input is high-dimensional, or the output layer, which may not be sufficiently flexible. Through a simulation study, we demonstrate that our method has good performance compared to alternative approaches, and explore the benefits that can be achieved by localizing the calibration based on different layers of the network. Finally, we apply our proposed method to a diamond price prediction problem, demonstrating the potential of our approach to improve prediction and uncertainty quantification in real-world applications.
- Arya, S., Mount, D. M., Netanyahu, N. S., Silverman, R., and Wu, A. Y. (1998), “An optimal algorithm for approximate nearest neighbor searching fixed dimensions,” Journal of the ACM (JACM), 45, 891–923.
- Blum, J. R., Hanson, D. L., and Koopmans, L. H. (1963), “On the strong law of large numbers for a class of stochastic processes,” Zeitschrift für Wahrscheinlichkeitstheorie und Verwandte Gebiete, 2, 1–11.
- Chen, Q., Zhao, B., Wang, H., Li, M., Liu, C., Li, Z., Yang, M., and Wang, J. (2021), “Spann: Highly-efficient billion-scale approximate nearest neighborhood search,” Advances in Neural Information Processing Systems, 34, 5199–5212.
- Dawid, A. P. (1984), “Present Position and Potential Developments: Some Personal Views: Statistical Theory: The Prequential Approach,” Journal of the Royal Statistical Society. Series A (General), 147, 278–292.
- Dheur, V., and Taieb, S. B. (2023), “A Large-Scale Study of Probabilistic Calibration in Neural Network Regression,” .
- Gal, Y., and Ghahramani, Z. (2016), “Dropout as a bayesian approximation: Representing model uncertainty in deep learning,” in international conference on machine learning, PMLR, pp. 1050–1059.
- Gneiting, T., Balabdaoui, F., and Raftery, A. E. (2007), “Probabilistic Forecasts, Calibration and Sharpness,” Journal of the Royal Statistical Society Series B: Statistical Methodology, 69, 243–268.
- Gneiting, T.— (2007), “Strictly Proper Scoring Rules, Prediction, and Estimation,” Journal of the American Statistical Association, 102, 359–378.
- Guo, C., Pleiss, G., Sun, Y., and Weinberger, K. Q. (2017), “On calibration of modern neural networks,” in International conference on machine learning, PMLR, pp. 1321–1330.
- Kuleshov, V., and Deshpande, S. (2022), “Calibrated and sharp uncertainties in deep learning via density estimation,” in International Conference on Machine Learning, PMLR, pp. 11683–11693.
- Kuleshov, V., Fenner, N., and Ermon, S. (2018), “Accurate uncertainties for deep learning using calibrated regression,” in International conference on machine learning, PMLR, pp. 2796–2804.
- Kull, M., Filho, T. S., and Flach, P. (2017a), “Beta calibration: a well-founded and easily implemented improvement on logistic calibration for binary classifiers,” PMLR, vol. 54 of Proceedings of Machine Learning Research, pp. 623–631.
- Kull, M., Silva Filho, T.— (2017b), “Beyond Sigmoids: How to obtain well-calibrated probabilities from binary classifiers with beta calibration,” Electronic Journal of Statistics, 11, 5052–5080.
- Kumar, A., Liang, P. S., and Ma, T. (2019), “Verified uncertainty calibration,” Advances in Neural Information Processing Systems, 32.
- Menéndez, P., Fan, Y., Garthwaite, P., and Sisson, S. (2014), “Simultaneous adjustment of bias and coverage probabilities for confidence intervals,” Computational Statistics and Data Analysis, 70, 35–50.
- Minderer, M., Djolonga, J., Romijnders, R., Hubis, F., Zhai, X., Houlsby, N., Tran, D., and Lucic, M. (2021), “Revisiting the calibration of modern neural networks,” Advances in Neural Information Processing Systems, 34, 15682–15694.
- Mukhoti, J., Kulharia, V., Sanyal, A., Golodetz, S., Torr, P., and Dokania, P. (2020), “Calibrating deep neural networks using focal loss,” Advances in Neural Information Processing Systems, 33, 15288–15299.
- Platt, J. et al. (1999), “Probabilistic outputs for support vector machines and comparisons to regularized likelihood methods,” Advances in large margin classifiers, 10, 61–74.
- Prangle, D., Blum, M. G., Popovic, G., and Sisson, S. (2014), “Diagnostic tools for approximate Bayesian computation using the coverage property,” Australian & New Zealand Journal of Statistics, 56, 309–329.
- Rodrigues, G., Prangle, D., and Sisson, S. A. (2018), “Recalibration: A post-processing method for approximate Bayesian computation,” Computational Statistics & Data Analysis, 126, 53–66.
- Si, P., Bishop, A., and Kuleshov, V. (2021), “Autoregressive Quantile Flows for Predictive Uncertainty Estimation,” arXiv preprint arXiv:2112.04643.
- Silva Filho, T., Song, H., Perello-Nieto, M., Santos-Rodriguez, R., Kull, M., and Flach, P. (2023), “Classifier calibration: a survey on how to assess and improve predicted class probabilities,” Machine Learning, 1–50.
- Tran, M.-N., Nguyen, N., Nott, D., and Kohn, R. (2020), “Bayesian deep net GLM and GLMM,” Journal of Computational and Graphical Statistics, 29, 97–113.
- Utpala, S., and Rai, P. (2020), “Quantile regularization: towards implicit calibration of regression models,” arXiv preprint arXiv:2002.12860.
- Xiong, M., Deng, A., Koh, P. W., Wu, J., Li, S., Xu, J., and Hooi, B. (2023), “Proximity-Informed Calibration for Deep Neural Networks,” .