On the Asymptotic Mean Square Error Optimality of Diffusion Models (2403.02957v4)
Abstract: Diffusion models (DMs) as generative priors have recently shown great potential for denoising tasks but lack theoretical understanding with respect to their mean square error (MSE) optimality. This paper proposes a novel denoising strategy inspired by the structure of the MSE-optimal conditional mean estimator (CME). The resulting DM-based denoiser can be conveniently employed using a pre-trained DM, being particularly fast by truncating reverse diffusion steps and not requiring stochastic re-sampling. We present a comprehensive (non-)asymptotic optimality analysis of the proposed diffusion-based denoiser, demonstrating polynomial-time convergence to the CME under mild conditions. Our analysis also derives a novel Lipschitz constant that depends solely on the DM's hyperparameters. Further, we offer a new perspective on DMs, showing that they inherently combine an asymptotically optimal denoiser with a powerful generator, modifiable by switching re-sampling in the reverse process on or off. The theoretical findings are thoroughly validated with experiments based on various benchmark datasets
- MIMO channel estimation using score-based generative models. IEEE Transactions on Wireless Communications, 22(6):3698–3713, 2023.
- The perception-distortion tradeoff. In IEEE/CVF Conference on Computer Vision and Pattern Recognition, pp. 6228–6237, 2018.
- Callahan, J. J. Advanced Calculus: A Geometric View. Springer New York, NY, 2010.
- Probability Theory: Independence, Interchangeability, Martingales. Springer NY, 3rd edition, 1997.
- Come-closer-diffuse-faster: Accelerating conditional diffusion models for inverse problems through stochastic contraction. In 2022 IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 12403–12412, 2022.
- Inversion by direct iteration: An alternative to denoising diffusion for image restoration, 2023. arXiv: 2303.11435.
- Image super-resolution using deep convolutional networks. IEEE Transactions on Pattern Analysis and Machine Intelligence, 38(2):295–307, 2016.
- A general derivative identity for the conditional mean estimator in Gaussian noise and some applications. In IEEE International Symposium on Information Theory (ISIT), pp. 1183–1188, 2020.
- Image denoising: The deep learning revolution and beyond—a survey paper. SIAM Journal on Imaging Sciences, 16(3):1594–1654, 2023.
- DiracDiffusion: Denoising and incremental reconstruction with assured data-consistency, 2023. arXiv: 2303.14353.
- Feller, W. On the theory of stochastic processes, with particular reference to applications. In Proceedings of the [First] Berkeley Symposium on Mathematical Statistics and Probability, volume 1, pp. 403–433. 1949.
- On the mean square error optimal estimator in one-bit quantized systems. IEEE Transactions on Signal Processing, 71:1968–1980, 2023.
- Applied Biomedical Engineering. IntechOpen, Rijeka, Aug 2011.
- Regularisation of neural networks by enforcing Lipschitz continuity. Machine Learning, 110:393 – 416, 2018.
- Iterative α𝛼\alphaitalic_α-(de)blending: A minimalist deterministic diffusion model. In ACM SIGGRAPH Conference Proceedings, 2023.
- Denoising diffusion probabilistic models. In Proceedings of the 34th International Conference on Neural Information Processing Systems, pp. 6840–6851, 2020.
- Wavedm: Wavelet-based diffusion models for image restoration, 2023. arXiv: 2305.13819.
- Denoising diffusion medical models, 2023. arXiv: 2304.09383.
- Robust compressed sensing using generative models. In Advances in Neural Information Processing Systems, volume 33, pp. 713–727, 2020.
- Stochastic solutions for linear inverse problems using the prior implicit in a denoiser. In Advances in Neural Information Processing Systems, volume 34, pp. 13242–13254, 2021.
- Bayesian denoising: From MAP to MMSE using consistent cycle spinning. IEEE Signal Processing Letters, 20(3):249–252, 2013.
- Variational diffusion models. In Advances in Neural Information Processing Systems, volume 34, pp. 21696–21707, 2021.
- An asymptotically MSE-optimal estimator based on Gaussian mixture models. IEEE Transactions on Signal Processing, 70:4109–4123, 2022.
- Diffwave: A versatile diffusion model for audio synthesis. In International Conference on Learning Representations, 2021.
- MNIST handwritten digit database. ATT Labs [Online]. Available: http://yann.lecun.com/exdb/mnist, 2, 2010.
- Loizou, P. C. Speech Enhancement: Theory and Practice. CRC Press, Inc., USA, 2nd edition, 2013.
- Diffusion model based posterior sampling for noisy linear inverse problems, 2023. arXiv: 2211.12343.
- Approximation by finite mixtures of continuous density functions that vanish at infinity. Cogent Mathematics & Statistics,, 7(1):1750861, 2020.
- Improved denoising diffusion probabilistic models. In Proceedings of the 38th International Conference on Machine Learning, volume 139, pp. 8162–8171, 2021.
- MMSE based MAP estimation for image denoising. Optics & Laser Technology, 57:252–264, 2014.
- Deep learning techniques for inverse problems in imaging. IEEE Journal on Selected Areas in Information Theory, 1(1):39–56, 2020.
- Librispeech: An ASR corpus based on public domain audio books. In IEEE International Conference on Acoustics, Speech and Signal Processing (ICASSP), pp. 5206–5210, 2015.
- Pedregosa, F. et al. Scikit-learn: Machine learning in Python. Journal of Machine Learning Research, 12:2825–2830, 2011.
- U-net: Convolutional networks for biomedical image segmentation. In Medical Image Computing and Computer-Assisted Intervention, pp. 234–241. Springer, 2015.
- Image super-resolution via iterative refinement. IEEE Transactions on Pattern Analysis and Machine Intelligence, 45(4):4713–4726, 2023.
- Statistical Signal Processing: Detection, Estimation, and Time Series Analysis. Pearson, 1991.
- Deep unsupervised learning using nonequilibrium thermodynamics. In Proceedings of the 32nd International Conference on Machine Learning, volume 37, pp. 2256–2265, 2015.
- Denoising diffusion implicit models. In International Conference on Learning Representations, 2021a.
- Generative modeling by estimating gradients of the data distribution. In Advances in Neural Information Processing Systems, volume 32, 2019.
- Score-based generative modeling through stochastic differential equations. In International Conference on Learning Representations, 2021b.
- Consistency models, 2023. arXiv: 2303.01469.
- Revisiting denoising diffusion probabilistic models for speech enhancement: Condition collapse, efficiency and refinement. Proceedings of the AAAI Conference on Artificial Intelligence, 37(11):13627–13635, Jun. 2023.
- Attention is all you need. In Advances in Neural Information Processing Systems, volume 30, 2017.
- Speech enhancement with score-based generative models in the complex STFT domain. In Proc. Interspeech, pp. 2928–2932, 2022.
- Deblurring via stochastic refinement. In IEEE/CVF Conference on Computer Vision and Pattern Recognition (CVPR), pp. 16272–16282, 2022.
- Fashion-MNIST: a novel image dataset for benchmarking machine learning algorithms, 2017. arXiv: 1708.07747.
- Diffusion model for generative image denoising, 2023. arXiv: 2302.02398.
- Compressive sensing by learning a Gaussian mixture model from measurements. IEEE Transactions on Image Processing, 24(1):106–119, 2015.
- Yang, Z. et al. Eliminating Lipschitz singularities in diffusion models, 2023. arXiv: 2306.11251.
- Wide residual networks, 2017. arXiv: 1605.07146.
- Loss functions for image restoration with neural networks. IEEE Transactions on Computational Imaging, 3(1):47–57, 2017.
- Denoising diffusion models for plug-and-play image restoration. In 2023 IEEE/CVF Conference on Computer Vision and Pattern Recognition Workshops (CVPRW), pp. 1219–1229, 2023.