Emergent Mind

Analytical Bounds between Entropy and Error Probability in Binary Classifications

(1205.6602)
Published May 30, 2012 in cs.IT and math.IT

Abstract

The existing upper and lower bounds between entropy and error probability are mostly derived from the inequality of the entropy relations, which could introduce approximations into the analysis. We derive analytical bounds based on the closed-form solutions of conditional entropy without involving any approximation. Two basic types of classification errors are investigated in the context of binary classification problems, namely, Bayesian and non-Bayesian errors. We theoretically confirm that Fano's lower bound is an exact lower bound for any types of classifier in a relation diagram of "error probability vs. conditional entropy". The analytical upper bounds are achieved with respect to the minimum prior probability, which are tighter than Kovalevskij's upper bound.

We're not able to analyze this paper right now due to high demand.

Please check back later (sorry!).

Generate a summary of this paper on our Pro plan:

We ran into a problem analyzing this paper.

Newsletter

Get summaries of trending comp sci papers delivered straight to your inbox:

Unsubscribe anytime.