Leveraging Self-Consistency for Data-Efficient Amortized Bayesian Inference (2310.04395v4)
Abstract: We propose a method to improve the efficiency and accuracy of amortized Bayesian inference by leveraging universal symmetries in the joint probabilistic model of parameters and data. In a nutshell, we invert Bayes' theorem and estimate the marginal likelihood based on approximate representations of the joint model. Upon perfect approximation, the marginal likelihood is constant across all parameter values by definition. However, errors in approximate inference lead to undesirable variance in the marginal likelihood estimates across different parameter values. We penalize violations of this symmetry with a \textit{self-consistency loss} which significantly improves the quality of approximate inference in low data regimes and can be used to augment the training of popular neural density estimators. We apply our method to a number of synthetic problems and realistic scientific models, discovering notable advantages in the context of both neural posterior and likelihood approximation.
- Robust model training and generalisation with studentising flows, 2020.
- Neural networks enable efficient and accurate simulation-based inference of evolutionary parameters from adaptation dynamics. PLoS Biology, 20(5):e3001633, 2022.
- Approximate Bayesian computation with domain expert in the loop. In International Conference on Machine Learning, pp. 1893–1905. PMLR, 2022.
- Variational inference: A review for statisticians. Journal of the American statistical Association, 112(518):859–877, 2017.
- A guide to constraining effective field theories with machine learning. Physical Review D, 98(5):052004, 2018.
- Madminer: Machine learning-based inference for particle physics. Computing and Software for Big Science, 4:1–25, 2020a.
- Mining gold from implicit models to improve likelihood-free inference. Proceedings of the National Academy of Sciences, 117(10):5242–5249, 2020b.
- Some models are useful, but how do we know which ones? towards a unified bayesian model taxonomy. Statistics Surveys, 17(none), 2023. ISSN 1935-7516. doi: 10.1214/23-ss145. URL http://dx.doi.org/10.1214/23-SS145.
- Stan: A probabilistic programming language. Journal of statistical software, 76(1), 2017.
- The frontier of simulation-based inference. Proceedings of the National Academy of Sciences, 2020.
- Real-time gravitational wave science with neural posterior estimation. Physical Review Letters, 127(24), December 2021. ISSN 1079-7114. doi: 10.1103/physrevlett.127.241103. URL http://dx.doi.org/10.1103/PhysRevLett.127.241103.
- Neural importance sampling for rapid and reliable gravitational-wave inference. Physical Review Letters, 130(17), April 2023. ISSN 1079-7114. doi: 10.1103/physrevlett.130.171403. URL http://dx.doi.org/10.1103/PhysRevLett.130.171403.
- Neural spline flows. Advances in neural information processing systems, 32, 2019.
- On optimality of kernels for approximate bayesian computation using sequential monte carlo, June 2011.
- Deep adaptive design: Amortizing sequential bayesian experimental design. In International Conference on Machine Learning, 2021. URL https://api.semanticscholar.org/CorpusID:232104961.
- Compositional score modeling for simulation-based inference, 2022.
- Bayesian Data Analysis (3rd Edition). Chapman and Hall/CRC, 2013.
- Variational methods for simulation-based inference. In International Conference on Learning Representations, 2022.
- Training deep neural density estimators to identify mechanistic models of neural dynamics. Elife, 2020.
- Automatic posterior transformation for likelihood-free inference. In International Conference on Machine Learning, 2019.
- A Kernel Two-Sample Test. The Journal of Machine Learning Research, 13:723–773, 2012.
- Aleatoric and Epistemic Uncertainty in Machine Learning: An Introduction to Concepts and Methods. Machine Learning, 110(3):457–506, March 2021. ISSN 0885-6125, 1573-0565. doi: 10.1007/s10994-021-05946-3. URL http://arxiv.org/abs/1910.09457.
- Köthe, U. A review of change of variable formulas for generative modeling, 2023.
- Simulation intelligence: Towards a new generation of scientific methods. arXiv preprint, 2021.
- Set transformer: A framework for attention-based permutation-invariant neural networks. In Chaudhuri, K. and Salakhutdinov, R. (eds.), Proceedings of the 36th International Conference on Machine Learning, volume 97 of Proceedings of Machine Learning Research, pp. 3744–3753. PMLR, 2019.
- Flow matching for generative modeling. In The 11th International Conference on Learning Representations, 2023. URL https://openreview.net/forum?id=PqvMRDCJT9t.
- Benchmarking simulation-based inference. arXiv preprint, 2021.
- Simulating ratios of normalizing constants via a simple identity: a theoretical exploration. Statistica Sinica, 1996.
- Spectral normalization for generative adversarial networks. In International Conference on Learning Representations, 2018.
- Dissecting the dynamics of the hes1 genetic oscillator. Journal of Theoretical Biology, 254(4):784–798, oct 2008. doi: 10.1016/j.jtbi.2008.07.013.
- Neal, R. M. MCMC using Hamiltonian dynamics. May 2011. doi: 10.1201/b10905.
- Score matched neural exponential families for likelihood-free inference. J. Mach. Learn. Res., 23:38–1, 2022.
- Sequential neural likelihood: Fast likelihood-free inference with autoregressive flows. 2019.
- BayesFlow: Learning complex stochastic models with invertible neural networks. IEEE transactions on neural networks and learning systems, 2020.
- JANA: Jointly Amortized Neural Approximation of Complex Bayesian Models. In Evans, R. J. and Shpitser, I. (eds.), Proceedings of the 39th Conference on Uncertainty in Artificial Intelligence, volume 216 of Proceedings of Machine Learning Research, pp. 1695–1706. PMLR, 2023.
- Variational inference with normalizing flows. In Proceedings of the 32nd International Conference on International Conference on Machine Learning - Volume 37, ICML’15, pp. 1530–1538. JMLR.org, 2015.
- Graphical test for discrete uniformity and its applications in goodness-of-fit evaluation and multiple sample comparison. Statistics and Computing, 32(2):1–21, 2022.
- Consistency models for scalable and fast simulation-based inference, 2023.
- Sequential neural score estimation: Likelihood-free inference with conditional score based diffusion models, 2022.
- Designing attractive models via automated identification of chaotic and oscillatory dynamical regimes. Nature Communications, 2(1), 2011. doi: 10.1038/ncomms1496.
- Sequential monte carlo without likelihoods. Proceedings of the National Academy of Sciences, 104(6):1760–1765, February 2007. doi: 10.1073/pnas.0607208104. URL https://doi.org/10.1073/pnas.0607208104.
- Improved Techniques for Training Consistency Models, October 2023.
- Consistency models. In International Conference on Machine Learning, 2023.
- Validating Bayesian inference algorithms with simulation-based calibration. arXiv preprint, 2018.
- loo: Efficient leave-one-out cross-validation and WAIC for Bayesian models, 2022.
- Mental speed is high until age 60 as revealed by analysis of over a million participants. Nature Human Behaviour, 6(5):700–708, May 2022. doi: 10.1038/s41562-021-01282-7.
- Watanabe, S. Algebraic geometry and statistical learning theory, volume 25. Cambridge university press, 2009.
- Sequential neural posterior and likelihood approximation. arXiv preprint, 2021.
- Deep sets, 2017.
- Probabilistic damage detection using a new likelihood-free Bayesian inference method. Journal of Civil Structural Health Monitoring, 13(2-3):319–341, 2023.
- Sensitivity-guided iterative parameter identification and data generation with BayesFlow and PELS-VAE for model calibration. Advanced Modeling and Simulation in Engineering Sciences, 10(1):1–28, 2023.
- Marvin Schmitt (15 papers)
- Daniel Habermann (3 papers)
- Ullrich Köthe (52 papers)
- Paul-Christian Bürkner (58 papers)
- Desi R. Ivanova (8 papers)
- Stefan T. Radev (31 papers)