Computer Science – Information Theory
Scientific paper
2010-05-17
Computer Science
Information Theory
submitted to IEEE Trans. on Information Theory
Scientific paper
In this paper, two new classes of lower bounds on the probability of error for $m$-ary hypothesis testing are proposed. Computation of the minimum probability of error which is attained by the maximum a-posteriori probability (MAP) criterion is usually not tractable. The new classes are derived using Holder's inequality and reverse Holder's inequality. The bounds in these classes provide good prediction of the minimum probability of error in multiple hypothesis testing. The new classes generalize and extend existing bounds and their relation to some existing upper bounds is presented. It is shown that the tightest bounds in these classes asymptotically coincide with the optimum probability of error provided by the MAP criterion for binary or multiple hypothesis testing problem. These bounds are compared with other existing lower bounds in several typical detection and classification problems in terms of tightness and computational complexity.
Routtenberg Tirza
Tabrikian Joseph
No associations
LandOfFree
General Classes of Lower Bounds on the Probability of Error in Multiple Hypothesis Testing does not yet have a rating. At this time, there are no reviews or comments for this scientific paper.
If you have personal experience with General Classes of Lower Bounds on the Probability of Error in Multiple Hypothesis Testing, we encourage you to share that experience with our LandOfFree.com community. Your opinion is very important and General Classes of Lower Bounds on the Probability of Error in Multiple Hypothesis Testing will most certainly appreciate the feedback.
Profile ID: LFWR-SCP-O-298342