Papers
Topics
Authors
Recent
Detailed Answer
Quick Answer
Concise responses based on abstracts only
Detailed Answer
Well-researched responses based on abstracts and relevant paper content.
Custom Instructions Pro
Preferences or requirements that you'd like Emergent Mind to consider when generating responses
Gemini 2.5 Flash
Gemini 2.5 Flash 28 tok/s
Gemini 2.5 Pro 40 tok/s Pro
GPT-5 Medium 16 tok/s Pro
GPT-5 High 13 tok/s Pro
GPT-4o 103 tok/s Pro
Kimi K2 197 tok/s Pro
GPT OSS 120B 471 tok/s Pro
Claude Sonnet 4 38 tok/s Pro
2000 character limit reached

Projection Regret: Reducing Background Bias for Novelty Detection via Diffusion Models (2312.02615v1)

Published 5 Dec 2023 in cs.LG and cs.CV

Abstract: Novelty detection is a fundamental task of machine learning which aims to detect abnormal ($\textit{i.e.}$ out-of-distribution (OOD)) samples. Since diffusion models have recently emerged as the de facto standard generative framework with surprising generation results, novelty detection via diffusion models has also gained much attention. Recent methods have mainly utilized the reconstruction property of in-distribution samples. However, they often suffer from detecting OOD samples that share similar background information to the in-distribution data. Based on our observation that diffusion models can \emph{project} any sample to an in-distribution sample with similar background information, we propose \emph{Projection Regret (PR)}, an efficient novelty detection method that mitigates the bias of non-semantic information. To be specific, PR computes the perceptual distance between the test image and its diffusion-based projection to detect abnormality. Since the perceptual distance often fails to capture semantic changes when the background information is dominant, we cancel out the background bias by comparing it against recursive projections. Extensive experiments demonstrate that PR outperforms the prior art of generative-model-based novelty detection methods by a significant margin.

Definition Search Book Streamline Icon: https://streamlinehq.com
References (70)
  1. A survey of outlier detection methodologies. Artificial intelligence review, 22:85–126, 2004.
  2. Explaining and harnessing adversarial examples. In ICLR, 2015.
  3. Efficient out-of-distribution detection in digital pathology using multi-head convolutional neural networks. In MIDL, pages 465–478, 2020.
  4. Out-of-distribution detection for automotive perception. In 2021 IEEE International Intelligent Transportation Systems Conference (ITSC), pages 2938–2943. IEEE, 2021.
  5. Codit: Conformal out-of-distribution detection in time-series data. arXiv preprint arXiv:2207.11769, 2022.
  6. C. M. Bishop. Novelty detection and neural network validation. In ICANN, 1994.
  7. Input complexity and out-of-distribution detection with likelihood-based generative models. In ICLR, 2020.
  8. Likelihood regret: An out-of-distribution detection score for variational autoencoder. In NeurIPS, 2020.
  9. Adversarially learned anomaly detection. In ICDM, 2018.
  10. Backpropagated gradient representations for anomaly detection. In ECCV, 2020.
  11. High-resolution image synthesis with latent diffusion models. In CVPR, 2022.
  12. Re-imagen:retrieval-augmented text-to-image generator. In ICLR, 2023.
  13. Language models are few-shot learners. arXiv preprint arXiv:2005.14165, 2020.
  14. Adversarial perturbations fool deepfake detectors. In IJCNN, 2020.
  15. Addressing semantic drift in question generation for semi-supervised question answering. In EMNLP, 2019.
  16. Image super-resolution via iterative refinement. arXiv preprint arXiv:2104.07636, 2021.
  17. Multiscale score matching for out-of-distribution detection. In ICLR, 2021.
  18. Unsupervised out-of-distribution detection with diffusion inpainting. arXiv preprint arXiv:2302.10326, 2023.
  19. Implicit generation and generalization in energy-based models. In NeurIPS, 2019.
  20. The unreasonable effectiveness of deep features as a perceptual metric. In CVPR, 2018.
  21. Consistency models. In ICML, 2023.
  22. Improved contrastive divergence training of energy-based models. In ICML, 2021.
  23. Vaebm: A symbiosis between variational autoencoders and energy-based models. In ICLR, 2021.
  24. Guiding energy-based models via contrastive latent variables. In ICLR, 2023.
  25. Nvae: A deep hierarchical variational autoencoder. In NeurIPS, 2020.
  26. Generative modeling by estimating gradients of the data distribution. In NeurIPS, 2019.
  27. Image quality assessment: From error visibility to structural similarity. IEEE Transactions on Image Processing, 2004.
  28. Score-based generative modeling through stochastic differential equations. In ICLR, 2021a.
  29. Elucidating the design space of diffusion-based generative models. In NeurIPS, 2022.
  30. Learning multiple layers of features from tiny images. Master’s thesis, Department of Computer Science, University of Toronto, 2009.
  31. Reading digits in natural images with unsupervised feature learning. In NIPS Workshop on Deep Learning and Unsupervised Feature Learning, 2011.
  32. Unsupervised representation learning by predicting image rotations. In ICLR, 2018.
  33. Do deep generative models know what they don’t know? In ICLR, 2019.
  34. Enhancing the reliability of out-of-distribution image detection in neural networks. In ICLR, 2018.
  35. A simple unified framework for detecting out-of-distribution samples and adversarial attacks. In NeurIPS, 2018.
  36. Lsun: Construction of a large-scale image dataset using deep learning with humans in the loop. arXiv preprint arXiv:1506.03365, 2015.
  37. Imagenet: A large-scale hierarchical image database. In CVPR, 2009.
  38. Describing textures in the wild. In CVPR, 2014.
  39. Csi: Novelty detection via contrastive learning on distributionally shifted instances. In NeurIPS, 2020.
  40. Invariant risk minimization. arXiv preprint arXiv:1907.02893, 2019.
  41. Auto-encoding variational bayes. In NeurIPS, 2014.
  42. Glow: Generative flow with invertible 1x1 convolutions. In NeurIPS, 2018.
  43. Memorizing normality to detect anomaly: memory-augmented deep autoencoder for unsupervised anomaly detection. In ICCV, 2019.
  44. Unsupervised anomaly detection with generative adversarial networks to guide marker discovery. In IPMI, 2017.
  45. Ocgan: One-class novelty detection using gans with constrained latent representations. In CVPR, 2019.
  46. Drocc: Deep robust one-class classification. In ICML, 2020.
  47. Pytorch: An imperative style, high-performance deep learning library. In NeurIPS, 2019.
  48. Pixelcnn++: a pixelcnn implementation with discretized logistic mixture likelihood and other modifications. In ICLR, 2017.
  49. P-kdgan: Progressive knowledge distillation with gans for one-class novelty detection. In IJCAI, 2020.
  50. Using self-supervised learning can improve model robustness and uncertainty. In NeurIPS, 2019.
  51. U-net: Convolutional networks for biomedical image segmentation. In MICCAI, 2015.
  52. Label-efficient semantic segmentation with diffusion models. In ICLR, 2022.
  53. Adaptive prototype learning and allocation for few-shot segmentation. In CVPR, 2021.
  54. Unsupervised representation for semantic segmentation by implicit cycle-attention contrastive learning. In AAAI, 2022.
  55. On the impact of spurious correlation for out-of-distribution detection. In AAAI, 2022.
  56. Imagenet-trained cnns are biased towards texture; increasing shape bias improves accuracy and robustness. In ICLR, 2019.
  57. Denoising diffusion probabilistic models. In NeurIPS, 2020.
  58. Pseudo numerical methods for diffusion models on manifolds. In ICLR, 2022.
  59. Genie: Higher-order denoising diffusion solvers. In NeurIPS, 2022.
  60. Progressive distillation for fast sampling of diffusion models. In ICLR, 2022.
  61. Denoising diffusion implicit models. In ICLR, 2021b.
  62. Novelty detection via blurring. In ICLR, 2020.
  63. Why normalizing flows fail to detect out-of-distribution data. In NeurIPS, 2020.
  64. Understanding failures in out-of-distribution detection with deep generative models. In ICML, 2021.
  65. Learning from failure: training debiased classifier from biased classifier. In NeurIPS, 2020.
  66. On feature learning in the presence of spurious correlations. In NeurIPS, 2022.
  67. Discover and cure: concept-aware mitigation of spurious correlations. In ICML, 2023.
  68. Grad-cam: visual explanations from deep networks via gradient-based localization. In ICCV, 2017.
  69. Dual diffusion implicit bridges for image-to-image translation. In ICLR, 2023.
  70. Gans trained by a two time-scale update rule converge to a local nash equilibrium. In NeurIPS, 2017.
Citations (7)
List To Do Tasks Checklist Streamline Icon: https://streamlinehq.com

Collections

Sign up for free to add this paper to one or more collections.

Summary

We haven't generated a summary for this paper yet.

Dice Question Streamline Icon: https://streamlinehq.com

Follow-Up Questions

We haven't generated follow-up questions for this paper yet.

Don't miss out on important new AI/ML research

See which papers are being discussed right now on X, Reddit, and more:

“Emergent Mind helps me see which AI papers have caught fire online.”

Philip

Philip

Creator, AI Explained on YouTube