Papers
Topics
Authors
Recent
Gemini 2.5 Flash
Gemini 2.5 Flash
125 tokens/sec
GPT-4o
53 tokens/sec
Gemini 2.5 Pro Pro
42 tokens/sec
o3 Pro
4 tokens/sec
GPT-4.1 Pro
47 tokens/sec
DeepSeek R1 via Azure Pro
28 tokens/sec
2000 character limit reached

Uncertainty Estimations by Softplus normalization in Bayesian Convolutional Neural Networks with Variational Inference (1806.05978v6)

Published 15 Jun 2018 in cs.LG, cs.CV, cs.NE, and stat.ML

Abstract: We introduce a novel uncertainty estimation for classification tasks for Bayesian convolutional neural networks with variational inference. By normalizing the output of a Softplus function in the final layer, we estimate aleatoric and epistemic uncertainty in a coherent manner. The intractable posterior probability distributions over weights are inferred by Bayes by Backprop. Firstly, we demonstrate how this reliable variational inference method can serve as a fundamental construct for various network architectures. On multiple datasets in supervised learning settings (MNIST, CIFAR-10, CIFAR-100), this variational inference method achieves performances equivalent to frequentist inference in identical architectures, while the two desiderata, a measure for uncertainty and regularization are incorporated naturally. Secondly, we examine how our proposed measure for aleatoric and epistemic uncertainties is derived and validate it on the aforementioned datasets.

User Edit Pencil Streamline Icon: https://streamlinehq.com
Authors (3)
  1. Kumar Shridhar (25 papers)
  2. Felix Laumann (9 papers)
  3. Marcus Liwicki (86 papers)
Citations (18)

Summary

We haven't generated a summary for this paper yet.