<< Chapter < Page Chapter >> Page >

Frequently, more than two viable models for data generation can be defined for a given situation. The classification problem is to determine which of several models best "fits" a set of measurements. For example,determining the type of airplane from its radar returns forms a classification problem. The model evaluation framework has theright structure if we can allow more than two models. We happily note that in deriving the likelihood ratio test we did not needto assume that only two possible descriptions exist. Go back and examine the expression for the maximum probability correct decision rule . If K models seem appropriate for a specific problem, the decision rule maximizingthe probability of making a correct choice is i i 1 K i p r i r To determine the largest of K quantities, exactly K 1 numeric comparisons need be made. When we have two possible models( K 2 ), this decision rule reduces to the computation of the likelihood ratio and its comparison to a threshold. In general, K 1 likelihood ratios need to be computed and compared to a threshold. Thus the likelihood ratio test can be viewed as aspecific method for determining the largest of the decision statistics i p r i r .

Since we need only the relative ordering of the K decision statistics to make a decision, we can apply any transformation T to them that does not affect ordering. In general, possible transformations must be positively monotonic to satisfythis condition. For example, the needless common additive components in the decision statistics can be eliminated, even ifthey depend on the observations. Mathematically, "common" means that the quantity does not depend on the model index i . The transformation in this case would be of the form T z i z i a , clearly a monotonic transformation. A positive multiplicative factor can also be "canceled"; if negative, the ordering would be reversed andthat cannot be allowed. The simplest resulting expression becomes the sufficient statistic i r for the model. Expressed in terms of the sufficient statistic, the maximum probability correct or the Bayesiandecision rule becomes i i 1 K C i i r where C i summarizes all additive terms that do not depend onthe observation vector r . The quantity i r is termed the sufficient statistic associated with model i . In many cases, the functional form of the sufficient statistic varies little from one model to another andexpresses the necessary operations that summarize the observations. The constants C i are usually lumped together to yield the threshold against which we compare the sufficient statistic. For example,in the binary model situation, the decision rule becomes 1 r C 1 0 1 0 r C 0 or 1 r 0 r 0 1 C 0 C 1 Thus, the sufficient statistic for the decision rule is 1 r 0 r and the threshold is C 0 C 1 .

In the Gaussian problem just discussed, the logarithm of the likelihood function is p r i r L 2 2 2 1 2 2 l 0 L 1 r l m ( i ) 2 where m ( i ) is the mean under model i . After appropriate simplification that retains the ordering, we have i r m ( i ) 2 l 0 L 1 r l C i 1 2 L m ( i ) 2 2 c i The term c i is a constant defined by the error criterion; for the maximum probability correct criterion, this constant is i .

When employing the Neyman-Pearson test, we need to specify the various error probabilities H j true say i . These specifications amount to determining the constants c i when the sufficient statistic is used. Since K 1 comparisons will be used to home in on the optimal decision, only K 1 error probabilities need be specified. Typically, the quantities 0 true say H i , i 1 K 1 are used, particularly when the model 0 represents the situation when no signal is present (see this problem ).

Get Jobilize Job Search Mobile App in your pocket Now!

Get it on Google Play Download on the App Store Now




Source:  OpenStax, Statistical signal processing. OpenStax CNX. Dec 05, 2011 Download for free at http://cnx.org/content/col11382/1.1
Google Play and the Google Play logo are trademarks of Google Inc.

Notification Switch

Would you like to follow the 'Statistical signal processing' conversation and receive update notifications?

Ask